statistical analysis of complex systems - santa fe institute
TRANSCRIPT
![Page 1: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/1.jpg)
Cosma ShaliziStatistics Department, Carnegie Mellon University
Statistical Analysis of Complex Systems
Complex Systems Summer School 2006, Beijing
![Page 2: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/2.jpg)
exploitation vs. explorationincreasing returns→skew distributions
stability through hierarchylocal inhibition + long-range activation
etc.
nonlinear dynamicstime-series analysiscellular automatastochastic spatial modelsinformation theoryevolutionary game theoryagent-based modelsrenormalization groupthermodynamic formalismheavy-tailed stochastics(data-mining/statistical learning)etc.
pattern formationspin glassesturbulence
social insectsmachine learning
optimizationneural networks (real, fake)
evolution (real, fake)immune system
networksgene regulation
evolutionary economicssocial dynamics
sandpilesetc.
measures of complexityorganization: represent, detect, use
intrinsic computation, semanticsphysics of information
does complexity increase?
Patterns
Foundations
Topics Tools
After a triangle drawnby Prof. Michael Cohen(School of Information,University of Michigan)
![Page 3: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/3.jpg)
1. Power laws (today)what they arewhat they mean; what they don’t meanhow to tell if you have one
2. Statistics for complex systems (Wednesday)Fitting modelsInductive complexityComparison of alternatives
3. Model discovery (Thursday)
![Page 4: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/4.jpg)
Power Laws
![Page 5: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/5.jpg)
What they areWhy they meanWhat they don’t meanHow to tell if you have one
![Page 6: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/6.jpg)
What Are They?
![Page 7: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/7.jpg)
Three Kinds of Things Called “Power Laws”
1. Physical laws, like Newton’s law of gravitation:
2. Scaling relations, like the “three-quarters” law of biology:
3. Statistical distributions, like Zipf ’s law of city sizes:
We are only going to look at distributions
F ! r−2
F ! r−2
P ! m−3/4
Pr(X ≥ x) ! x−1.3
![Page 8: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/8.jpg)
probability density
cumulative distribution
parametersthreshold or minimumslope or exponent
strongly heavy-tailed or right-skewedmedian is much higher than meanvery large fluctuations from mean“80/20” rule
Power Law (Pareto) Distributions
p(x) =!−1xmin
(x
xmin
)−!
Pr(X ≥ x) =1
xmin
(x
xmin
)−(!−1)
xmin
xmin
!
![Page 9: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/9.jpg)
Plots
0e+00 2e+05 4e+05 6e+05 8e+05 1e+06
0.0
e+
00
5.0
e-1
11
.0e
-10
1.5
e-1
0
x
pro
ba
bility d
en
sity
Density
1e+00 1e+02 1e+04 1e+06
1e
-15
1e
-11
1e
-07
1e
-03
x
pro
ba
bility d
en
sity
Cumulative Probability
0e+00 2e+05 4e+05 6e+05 8e+05 1e+06
0.0
0.2
0.4
0.6
0.8
1.0
x
cu
mu
lative
pro
ba
bility
1e+00 1e+02 1e+04 1e+06
1e
-09
1e
-07
1e
-05
1e
-03
1e
-01
x
cu
mu
lative
pro
ba
bility
linear
log-log
![Page 10: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/10.jpg)
What Do They Mean?
![Page 11: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/11.jpg)
Why they matter for reality
Many quantities have power law distributions...Income and wealth (Pareto; apparently oldest use)City sizes (Zipf)Word frequenciesPaper citationsEarthquake amplitudeSolar flaresFamily namesSizes of genera
... so Gaussian/normal assumptions are badly wrong
... and models need to match reality
![Page 12: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/12.jpg)
Why They Matter: TheoryDifferent mechanisms produce different distributions
So: use distribution to learn about mechanism
Averaging many independent variables ⇒ Gaussiancentral limit theoremalso true if only approximately independent (“mixing”)
Self-reinforcing growth + random seeding ⇒ power law
Simon, 1955 (Carnegie Mellon)Barabasi and Albert, 1999
Exponential growth for a random (exponential) timeReed and Hughes, 2002
Many other power-law mechanisms!
![Page 13: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/13.jpg)
What Don’t They Mean?
![Page 14: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/14.jpg)
Why do physicists care?In equilibrium statistical mechanics, Gaussian fluctuations
Observables average over many, many moleculesMolecules become independent very quickly (mixing)Einstein fluctuation formula based on entropy
Except: power-law fluctuations at critical pointsCorrelation length diverges, no mixingUsual CLT does not applyPower laws come out of lack of scale
So physicists think “power law ⇒ critical ⇒ cool”Origin of “self-organized criticality”
BUT THIS IS WRONG, WRONG, WRONG!
![Page 15: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/15.jpg)
Why is this wrong?
Because there are many other ways to produce power laws!Many of them do not involve long-range correlation, memory, or anything else we would call “complex”
Not all complex systems have power lawsNot all power law distributions indicate complexity
![Page 16: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/16.jpg)
How Can You Tell If You Have One?
![Page 17: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/17.jpg)
The bad way: draw a straight lineEasyCommon (outside of statistics and economics)Worthless
The right way: use maximum likelihoodA little harder (must learn some statistics)Not so common (outside of statistics and economics)Works
![Page 18: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/18.jpg)
The Bad Way
Plot the cumulative distributionFit a line by least squares; slope = α - 1Use error estimates in slope from fitCheck R2, the fraction of variance which comes from this line; accept if it is high
![Page 19: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/19.jpg)
Advantages of the Bad Way
It is easyExcel or Gnuplot will do it for you
It requires no knowledgeIt sounds reasonableMany other people do it
vast majority of papers in physics or econophysics on power laws do exactly this
![Page 20: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/20.jpg)
Why then is it bad?
It does not give a proper probability distribution!If you do have a power law: bad estimates
Asymptotic convergence to true values, but very slowAlso, the standard error for the slope is wrong
If you don’t: it can’t tell the differenceLow power against heavy-tailed alternativesA small part of any curve looks like a straight lineWe will see an example
Both of problems have been known to statisticians since the 1960s at least
![Page 21: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/21.jpg)
What then is the right way?
Estimate parameters by maximum likelihoodError bars by bootstrappingCheck fit by Kolmogorov-Smirnov test (good)Test against real alternatives (better)
![Page 22: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/22.jpg)
Parameter estimation
Likelihood of a parameter value = probability of data, under that parameter valueMaximum likelihood estimate (MLE) = what parameter value makes the data most probable?
Curve fitting says “what curve goes closest to my data, in Euclidean distance?”MLE says, “what distribution goes closest to my data, in Kullback divergence/relative entropy?”This idea leads to information geometry
![Page 23: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/23.jpg)
Use log-likelihood instead of likelihood
Explicit solutions
More on MLE
L(!,xmin)=n
"i=1log
!−1xmin
(xi
xmin
)−!
=n log!−1+n(!−1) logxmin−!n
"i=1logx
x̂min = x(1), smallest value
!̂= 1+n
[n
"i=1log
x
xmin
]−1
![Page 24: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/24.jpg)
For Pareto distribution, MLEs are consistenti.e., they converge in probability on the true valuesmuch more accurate than line-fitting
They are also sufficienti.e., they use all the information in the data
MLE does not give error estimatesMLE does not check the fitMLE does not test against alternatives
![Page 25: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/25.jpg)
Key technique of modern statisticsSimilar to “surrogate data” in nonlinear dynamics
Fit model to dataSimulate new data set from modelApply procedure to simulated dataRepeat many times to get typical results
if model is correct
this is parametric bootstrap, nonparametric is trickier
Bootstrapping
![Page 26: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/26.jpg)
Bootstrap Error Estimates
Fit Pareto to n data points with MLEGenerate n random values from fitted ParetoApply MLE to simulated data, calculate errorRepeat m times to get average error
Works for any distribution, not just Paretobut is not always the most efficient way(“local asymptotic normality”)
![Page 27: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/27.jpg)
Checking the Fit
Goodness-of-fit tests: if the model is right, what is the probability of results like the data?Or: probability of results as far from expectations as the dataModel here is a distribution, so we need a measure of distance between distributions
![Page 28: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/28.jpg)
Kolmogorov-Smirnov Test
If the model is right, D is usually small, shrinks as n grows
null distribution of D
K&S found null distribution as function of n for fixed model F(x)For model estimated from data, find distribution of D by bootstrapping
F(x)=cumulative distribution according to modelF̂n(x)=cumulative distribution according to data
D=maxx
∣∣∣F(x)− F̂n(x)∣∣∣
![Page 29: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/29.jpg)
More on the KS Test
Does not require binning the dataHigh probability of detecting bad models
has high power against most alternativesnon-parametric test
Not informative if the fit is badprice of being non-parametric
![Page 30: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/30.jpg)
Testing Alternatives
Pareto is only appropriate for heavy-tailed dataCompare it to other heavy-tailed distributions
i.e., don’t bother with Gaussian, exponential, etc.
There are many!Weibull distributionStretched exponentiallognormaletc.
Lognormal is usually the most important
![Page 31: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/31.jpg)
Lognormal Distribution
log(X) has a normal/Gaussian distributionArises from multiplicative CLT (multiplying many small independent factors)just like Gaussian arises from CLT (adding many small independent variables)
Extremely commonParameters (m and s) easily found by MLECan be hard to tell from power law
p(x) =1√2!sx
e−(logx−m)2
2s2
![Page 32: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/32.jpg)
Cumulative distribution of lognormal (log-log plot)Notice how straight the tail is
1e-02 1e+00 1e+02 1e+04 1e+06
1e
-44
1e
-34
1e
-24
1e
-14
1e
-04
x
cu
mu
lative
pro
ba
bility
~
1e+00 1e+02 1e+04 1e+06
1e
-44
1e
-34
1e
-24
1e
-14
1e
-04
x
cu
mu
lative
pro
ba
bility
![Page 33: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/33.jpg)
Likelihood Ratio TestNull hypothesis: data comes from ParetoAlternative hypothesis: some other heavy-tailed distribution, say lognormalFit Pareto by MLEFit alternative by MLE
Lnull= likelihood of fitted power law modelLalt= of fitted alternative model
T =Lalt
Lnull
T ≤ c⇒accept null hypothesisT > c⇒reject null hypothesis
![Page 34: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/34.jpg)
More on LRT
Why does this work?If Pareto is right, then eventually T → 0If alternative is right, then eventually T → ∞
How can it go wrong?False alarm or Type I error: reject null (Pareto) when it’s rightMiss or Type II error: accept null when it’s wrong
Usual way to pick c:Find null distribution of TFind c such that probability of false alarm is some desired small value, the “significance level”
Alternately: c = 1
![Page 35: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/35.jpg)
Severity and Evidence
Accept/reject is not enoughIf we accept power law (T < c), still need to know
power = probability, under alternative, that T > cseverity = probability, under alternative, that T > actual likelihood ratio from data
If we reject power law (T > c), need to knowsignificance = probability, under null, that T > cseverity = probability, under null, that T < actual likelihood ratio
Low significance and high power are both goodHigh severity: the test provides strong evidenceLow severity: the test provides weak evidence
![Page 36: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/36.jpg)
How not to draw a straight line
http://bactra.org/weblog/232.html
![Page 37: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/37.jpg)
Cumulative degree distribution of English-language political blogs, 2004
(data courtesy H. Farrell & D. Drenzer)3 orders of magnitude
m11M1M
1
55M5M
5
1010M10M
10
5050M50M
50
100500500M500M
500
1000m5e-045e-04M5e-04M
5e-0
4
1e-035e-035e-03M5e-03M
5e-0
3
1e-025e-025e-02M5e-02M
5e-0
2
1e-015e-015e-01M5e-01M
5e-0
1
1e+00DataMDataM
Data
Blog degree dMBlog degree dM
Blog degree d
Prob(D>=d)MProb(D>=d)M
Pro
b(D
>=
d)
![Page 38: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/38.jpg)
Least-squares fitα = -2.15, R2 = 0.898
m11M1M
1
55M5M
5
1010M10M
10
5050M50M
50
100500500M500M
500
1000m5e-045e-04M5e-04M
5e-0
4
1e-035e-035e-03M5e-03M
5e-0
3
1e-025e-025e-02M5e-02M
5e-0
2
1e-015e-015e-01M5e-01M
5e-0
1
1e+00Least-Squares Fit to DataMLeast-Squares Fit to DataM
Least-Squares Fit to Data
Blog degree dMBlog degree dM
Blog degree d
Prob(D>=d)MProb(D>=d)M
Pro
b(D
>=
d)
![Page 39: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/39.jpg)
10,000 numbers from a log-normal distributionm= 0, s= 3
m1e-051e-05M1e-05M
1e-05
1e-021e-02M1e-02M
1e-02
1e+011e+01M1e+01M
1e+01
1e+041e+04M1e+04M
1e+04
m1e-041e-04M1e-04M
1e-0
4
1e-031e-03M1e-03M
1e-0
3
1e-021e-02M1e-02M
1e-0
2
1e-011e-01M1e-01M
1e-0
1
1e+001e+00M1e+00M
1e+
00
Simulation of Log-normalMSimulation of Log-normalM
Simulation of Log-normal
xMxM
x
Prob(X >= x)MProb(X >= x)M
Pro
b(X
>=
x)
![Page 40: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/40.jpg)
Least-squares fit to random numbers ≥ 15112 data points, 4+ orders of magnitude
R2 = 0.962
m11M1M
1
100100M100M
100
1000010000M10000M
10000
m1e-041e-04M1e-04M
1e-0
4
5e-045e-04M5e-04M
5e-0
4
1e-035e-035e-03M5e-03M
5e-0
3
1e-025e-025e-02M5e-02M
5e-0
2
1e-015e-015e-01M5e-01M
5e-0
1
Least-Squares Fit to Tail of SimulationMLeast-Squares Fit to Tail of SimulationM
Least-Squares Fit to Tail of Simulation
xMxM
x
Prob(X >= x)MProb(X >= x)M
Pro
b(X
>=
x)
![Page 41: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/41.jpg)
Maximum likelihood fit of power lawα = -1.30, log L = -18481.51
m11M1M
1
55M5M
5
1010M10M
10
5050M50M
50
100500500M500M
500
1000m5e-045e-04M5e-04M
5e-0
4
1e-035e-035e-03M5e-03M
5e-0
3
1e-025e-025e-02M5e-02M
5e-0
2
1e-015e-015e-01M5e-01M
5e-0
1
1e+00Power-Law Fit to DataMPower-Law Fit to DataM
Power-Law Fit to Data
Blog degree dMBlog degree dM
Blog degree d
Prob(D>=d)MProb(D>=d)M
Pro
b(D
>=
d)
![Page 42: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/42.jpg)
Maximum likelihood fit of log-normalm = 2.60, s = 1.48, log L = -17218.22
m11M1M
1
55M5M
5
1010M10M
10
5050M50M
50
100500500M500M
500
1000m5e-045e-04M5e-04M
5e-0
4
1e-035e-035e-03M5e-03M
5e-0
3
1e-025e-025e-02M5e-02M
5e-0
2
1e-015e-015e-01M5e-01M
5e-0
1
1e+00Log-normal Fit to DataMLog-normal Fit to DataM
Log-normal Fit to Data
Blog degree dMBlog degree dM
Blog degree d
Prob(D>=d)MProb(D>=d)M
Pro
b(D
>=
d)
![Page 43: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/43.jpg)
data aree-17218.22+18481.51 = e1263.29 ≈ 13,000,000
times more likely under the log-normal
m11M1M
1
55M5M
5
1010M10M
10
5050M50M
50
100500500M500M
500
1000m5e-045e-04M5e-04M
5e-0
4
1e-035e-035e-03M5e-03M
5e-0
3
1e-025e-025e-02M5e-02M
5e-0
2
1e-015e-015e-01M5e-01M
5e-0
1
1e+00Combined Fits to DataMCombined Fits to DataM
Combined Fits to Data
Blog degree dMBlog degree dM
Blog degree d
Prob(D>=d)MProb(D>=d)M
Pro
b(D
>=
d)
![Page 44: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/44.jpg)
Morals1. Power laws are an important kind of heavy-tailed
distributions, often seen in complex systems2. Do not estimate their parameters by line-fitting3. There are other heavy-tailed distributions, so check
before you say something is a power law4. Many mechanisms can generate power laws5. Some, but not all, of these mechanisms are complex
![Page 45: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/45.jpg)
General References on Power LawsM. E. J. Newman (2004), "Power laws, Pareto distributions and Zipf's law", http://arxiv.org/abs/cond-mat/0412004
If you read only one thing on power laws, make it this.Manfred Schroeder, Fractals, Chaos, Power Laws: Minutes from an Infinite Paradise
Fun, recreational-mathematics-level survey of many of the topics related to this school, including power law distributions
Mechanisms for Power LawsHerbert Simon (1955), "On a Class of Skew Distribution Functions", Biometrika 42: 425--440
Classic paper explaining power laws through “rich get richer” growthDidier Sornette (1998), "Multiplicative Processes and Power Laws", Physical Review E 57: 4811--4813, http://arxiv.org/abs/cond-mat/9708231Michael Mitzenmacher (2003), "A Brief History of Generative Models for Power Law and Lognormal Distributions", Internet Mathematics 1: 226--251, http://www.internetmathematics.org/volumes/1/2/pp226_251.pdfWilliam J. Reed and Barry D. Hughes (2002), "From Gene Families and Genera to Incomes and Internet File Sizes: Why Power Laws are so Common in Nature", Physical Review E, 66: 067103
Critical FluctuationsJoel Keizer (1987), Statistical Thermodynamics of Nonequilibrium Processes (Berlin: Springer-Verlag)
Excellent discussion of critical fluctuations and how they relate to non-equilibrium phenomena.L. S. Landau and E. M. Lifshitz (1980), Statistical Physics (Oxford: Pergamon)
Essential reference on the Einstein fluctuation formula and critical phenomena.Julia M. Yeomans (1992), Statistical Mechanics of Phase Transitions (Oxford: Clarendon Press)
Easier than the above, but less detailed.
References
![Page 46: Statistical Analysis of Complex Systems - Santa Fe Institute](https://reader034.vdocuments.us/reader034/viewer/2022042421/626094434d978d00a10eeca9/html5/thumbnails/46.jpg)
Statistical IssuesMichel L. Goldstein, Steven A. Morris and Gary G. Yen (2004), "Fitting to the Power-Law Distribution", http://arxiv.org/abs/cond-mat/0402322
Pedestrian, but accurate, paper on goodness-of-fit testing for power laws.Norman L. Johnson and Samuel Kotz (1970), Continuous Univariate Distributions, Part 1 (New York: Wiley)
Standard reference work, discusses the properties of the Pareto distribution, and the pros and cons of various estimators. This volume also includes the log-normal distribution.
Deborah G. Mayo (1996), Error and the Growth of Experimental Knowledge (Chicago: University of Chicago Press)Excellent book on how to really use statistical methods in scientific work. Best general discussion of evidence and severe tests.
Cosma Rohilla Shalizi and M. E. J. Newman (in prep.), “Statistical Inference with Power Law Distributions: Parameter Estimation and Comparison to Alternatives”
Manuscript in preparation; will be accompanied by code for automated hypothesis testing.