Basic numerical and statistical methodsVladimir Volkov
Institute of Crystallography, Moscow, Russia
Plan of the talk:
1. Introduction: data intepretation
2. Statistics: errors and goodness-of-fit
3. Data fitting by simple functions, solution of system of linear equations.
4. Data fitting by nonlinear models, methods forminimization.
5. Fourier transform, basic ideas.
Main goal for end data interpretation
Y
X
rρρρρto build a mathematical description of a physical model,
calculate physical responses Y(x) from the model that could be measured in an experiment and
compare them with experimental data (what is goodness-of-fit?)
Then, correct the model parameters by using the comparison results (methods for automatic correction?)
Case study: structure refinement by molecular tectonics.
Case study: PDC in the crystal and in solution
3 nm
experiment crystal
solutionχ = 1.68χ = 0.94
Left: crystallographic model (interaction area, 13.44 nm2) Right: result of rigid body refinement (r.m.s., 0.58 nm)Svergun, D.I., Petoukhov, M.V., Koch, M.H.J. & König, S. (2000) J. Biol. Chem. 275, 297-303.
Good fit:Differences are less than above. In addition, they do not have systematic behaviour. This means, that the model may be adequate.
Merit function: analysis of residuals.
Goodness-of-fitPoor fit:Differences between experimental data and the model are large and systematic. The goal is to find model parameters that provide minimum difference.
0 1 2 3 4 5 6 7-0.2
0.0
0.2
0.4
0.6
X
Y Experiment Model Difference
0 1 2 3 4 5 6 7-0.1
0.0
0.1
0.2
0.3
0.4
0.5
0.6
X
Y Experiment Model, best fit Difference
Case of residuals: sequence of random numbers Y (errors)
0 200 400 600 800 1000-4-3-2-10123Sequence of normal random numbers
Value
N of value
-4
-3
-2
-1
0
1
2
3
0 500 1000 1500number of occurences
Middle box value
of Y
If sum over all occurences = 1 (normalization) then this plot relates to the probability density function: P{Y=y}.
y
Number of impulses counted by a detector: Poisson distribution.
0
20
40
60
80
0 1000 2000 3000 4000 5000 6000
Box value (number of occurences)
0 200 400 600 800 1000
0
10
20
30
40
50
60
70
80 Poisson random numbers with mean 50
N of value
0 200 400 600 800 10000
2
4
6
8
10
12
14
16 Poisson random numbers with mean 5Value
0
5
10
15
0 5000 10000 15000
Box value (number of occurences)
Middle box value
Asymmetric at low mean values
Almost symmetric, close to a normal distribution at large mean values
Statistical characterization of a data set
Most probable (expected) value of a random variable yi:
Mean value : , from
Mean value from density distribution histogram:
Measure of variancemean linear deviation:
variance (empirical):
standard deviation: σσσσ,
(measure of variability about mean value: )
�=
=N
iiy
Ny
1
1 yNyyN
i
N
ii ==��
== 11
��=j
bj
bj jj nnby
�=
−=N
jj yy
Nd
1
1
( )�=
−−
=N
jj yy
N 1
221
1σ
σ±y
-4
-3
-2
-1
0
1
2
3
0 500 1000 1500number of occurences
Middle box value
y
σσσσ
bj
jbn
-4 -2 0 2 4 6 8 10 12 140.0
0.2
0.4
0.6
0.8
1.0
1.2
1.4
x
a
eh
h
σσσσ σσσσ
Poisson random variable (discrete), which is the number of successes that will be observed in repeated intervals of time, each of which may result in the occurenceor nonoccurence of a success.
where y is the fixed value, t is the observation time interval, Lt is the mean number of successes per interval of measurement t.
mean value (expected value) E{Y} = Lt,variance = Lt (thus, we measure values )The larger is t, the less is the relative error.
ty
ey
LtyYP −==!
}{
LtLtY ±=
Normal random variable (continuous).De Moivre-Laplace's theorem:
Most of statistical estimates uses this fact.Common formula:
mean value (expected value) E{Y} = a,variance = σ σ σ σ 2.
dxedLtLtYc d
c
x
N �−
∞→=��
���
� <−< 22
21limπ
22
2)(
21),( 2 σ
πσσ ⋅
−−=
ax
eaN
(probability that randomly obtained Y equals to the prescribed value y)
Measure of degree of adjacency of thwo data setsa1, a2, …, aN and b1, b2, …, bN
Our goal: to get a single characteristic number.We can not use summmation over all deviations because they may have opposite signs and compensate each other giving a rezult near zero.But we may use the sum over absolute values of deviations:
It works, but the use of this formula leads to significant mathematical difficulties: this function is not differentiable.Squaring the deviations proves to be a much more attractive solution:
�=
−=N
jjj ba
Nd
1
1
( )�=
−−
=N
jjj ba
Nd
1
221
1
Weighting of the summation terms:equalization of influence of errors
An example: level of Poisson noise is proportional to square root of the intensity. To reduce the influence of data points with large associated errors, all the elements are divided by their corresponding deviates. These deviates must be determined independently.
A: Curve to be fitted to
B’’: curve minimizing
E: Errors in Arelatively small large
B’: curve minimizing ( )�
=−
−=
N
jjj BA
Nd
1
221
1
�=
��
�
�
��
�
� −−
=N
j j
jj BAN
d1
22
11
σ
AB’
B’’E
0
I
Tangential gradient g
gmin=0
2χ
Chi-square: least squares fitting of a set of N data points Aj with errors σσσσj.
Main idea of a fitting procedure: At the point of the minimum, the derivatives (gradients)
because we have here zero slope of the surface of the minimizing function.
We use here the term “chi-square”, because sum of squares of random deviates have chi square distribution. Expected minimum value of chi-square is about unity.
02
=∂∂
kpχ
pk
��
�
�
��
�
�
��
�
�
� −−−
= �=
N
j j
jj BAMN 1
22
11min
σχ
Calculated from a model described by a set of M
parameters pk.
σ±y
We consider the problem of fitting a set of N data points (xi; yi) to a straight-line model withparameters a and b :
y(x) = y(x; a; b) = a+bx
This problem is often called linear regression. We assume that the uncertaintyσσσσιιιι associated with each measurement yi is known, and that the xi’s (values ofthe dependent variable) are known exactly.
To measure how well the model agrees with the data, we use the chi-square merit function, which in this case is
At its minimum, derivatives with respect to a; b vanish.
�=
−−−=
∂∂=
N
j j
jij bxayxb 1
2
2 )(20
σχ
�=
−−−=
∂∂=
N
j j
ji bxaya 1
2
220
σχ
Fitting data by a straight line.
�=
��
�
�
��
�
� −−−
=N
j j
jj bxayN 1
22
11
σχ
x
xyxxx
yxSbSaSSbSaS
=+=+
�= 2j
jjxy
yxS
σ�= 2
2
j
jxx
xS
σ
These conditions can be rewritten in a convenient form if we define the following sums:
With these definitions, we have here the system of linear equations:
This system may be solved in any common way giving
Fitting data by a straight line.
�= 21
jS
σ�= 2
j
jx
xS
σ�= 2
j
jy
yS
σ
Now, we must estimate the probable uncertainties in the estimates of a and b, since obviously the measurement errors in the data must introduce some uncertainty in the determination of those parameters. If the data are independent, then each contributes its own bit of uncertainty to the parameters.
2)( xxx SSS −=∆
∆−
= xyxyxx SSSSa
∆−
= yxxy SSSSb
Sourceof possible
computationalunstability
Consideration of propagation of errors shows that the variance in the value of any function will be
For the straight line,
Thus, we have the solution for coefficients of a straight line: Because we minimize sum of squares, it is possible to obtain once more Important estimate - the goodness-of-fit of the data to the model. This estimate indicates whether the parameters a and b in the model have any meaning at all! The probability Q that a value of chi-square as poor as the value we use to derive the solution,
should occur by chance is
where is incomplete gamma function. Using known tables of
statistics one may find that if Q > 0.1 then the solution is reliable, if 0.001<Q<0.1 then a and b may be acceptable. If Q < 0.001 then the model and/or estimation procedure can rightly be called into question.
Fitting data by a straight line. Propagation of errors in solution.
2
1
22��
�
�
��
�
�
∂∂= �
= j
N
jji y
fσσ
∆= xx
aS2σ
∆= S
b2σ
ba ba σσ ±± ,
�=
��
�
�
��
�
� −−−
=N
j j
jj bxayN 1
22
11
σχ �
�
�
�
��
�
� −Γ=2
,2
2 2χNQ inc
( )qpQ inc ,Γ=
i: index of a model parameter,
j: index of an experimental point
Fitting data by polynomialsm
mxaxaxaaxF ++++= ...)( 2210
�=
��
�
�
��
�
� −−−
=N
j j
jj xFymN 1
22 )(
11
σχ
mja j
,...,0,02
==∂∂χ
Polynom :
What we minimize:
Condition of minimum:
Systemof mlinear equations:
�=
N
j ja
1201
σ jN
j jxa �
=1211
σmj
N
j jm xa �
=121
σ jN
j jy�
=121
σ
jN
j jxa �
=1201
σ jN
j jxa �
=1211
σ1
121 +
=�
mj
N
j jm xa
σjj
N
j jyx�
=121
σ
mj
N
j jm xa �
=121
σ1
1211 +
=�
mj
N
j jxa
σm
jN
j jm xa 2
121
�= σ j
mj
N
j jyx�
=121
σ
+…+ +
+…+ +
+…+ + =
=
=
…….. …….. …….. ……..
Systems of linear equations - matrix notation
=C BAm
mMatrix notation: C A = B,
solution: A = C-1B
�=
N
j ja
1201
σj
N
j jxa �
=1211
σmj
N
j jm xa �
=121
σj
N
j jy�
=121
σ
jN
j jxa �
=1201
σ jN
j jxa �
=1211
σ1
121 +
=�
mj
N
j jm xa
σjj
N
j jyx�
=121
σ
mj
N
j jm xa �
=121
σ1
1211 +
=�
mj
N
j jxa
σm
jN
j jm xa 2
121
�= σ j
mj
N
j jyx�
=121
σ
+…+ +
+…+ +
+…+ + =
=
=
…….. …….. …….. ……..
Errors in solution: 122 −= jja Cj
χσ
Solution of systems of linear equations
=b1a1
m
mc11
cm1 cmm
c1m
am bm
Method of Gauss-Jordan elimination (triangulation)
1) C11a1 + C12a2 + C13a3 = b1
2) C21a1 + C22a2 + C23a3 = b2
3) C31a1 + C32a2 + C33a3 = b3multiply 1) by C21/C11, subtract from 2);multiply 1) by C31/C11, subtract from 3);
1) C11a1 + C12a2 + C13a3 = b1
2) C*22a2 + C*
23a3 = b*2
3) C*32a2 + C*
33a3 = b*3
1) C11a1 + C12a2 + C13a3 = b1
2) C*22a2 + C*
23a3 = b*2
3) C**33a3 = b**
3
multiply 2) by C*32 /C*22, subtract from 3);
Direct move:
Backward substitution:a3 = b**
3 / C**33 from 3);
a2 = (b*2 – C*
23a3) / C*22 from 2);
a1 = (b1 - C13a3 - C12a2) / C11 from 1)
“Pivoting”:Avoiding deletion by small coefficients by rearrangement of rows.
Solution of systems of linear equationsOverestimated systems: least squares solution.
This system does not have an exact solution.
Case study: fitting data by polynomials(N equations):
imimii yxaxaxaa =++++ ...2
210
=
D
BA
M= m+1
N
1 x1 x12 …x1
m
1 x2 x22 …x2
m
. . . . . . . . .
1 xn xn2 …xn
m
If we look at the matrix C, we find that
C = DTD (matrix of normal equations).
=DT
DC
Solution by least squares:C A = DT B,A=(C)-1 DT B
yi
11
Solution of systems of linear equations
Singular value decomposition (SVD)
.=.D VTU S
VT V
left
right
UUT =
=
1 01
10 1
1 01
10 1
Identity matrix (E)
Orthonormalmatrices
Diagonal matrix of singular values
U-1 = UT;V-1 = VT
Solution of systems of linear equations
Solution by least squares using normal equations:
Solution by least squares using SVD:D = U S VT;UT U S VT A = UT B; because UTU = E, and E play
role of a unity in multiplications:S VT A = UT B; S-1S = E, then
VT A = S-1 UT B; critical operation - inversion of siiA = V S-1 UT B because V VT = E
DTD A = DT B,A=(DTD)-1 DT B
)()1(1
)(1
)( 11m
mi
M
i i
iSSS
VVVBU
A ±±±���
����
� ⋅=�
=�
2
1
2�=
���
����
�=
M
i i
jia Sj
Vσ
Errors
Minimization of functionsYou are given a single function f(x) that depends on one or more independent variables.
You want to find the value of those variables {x} where f(x) takes on a minimum value. You can then calculate what value of f(x) is achieved at the minimum.An extremum (minimum point) can be either global (truly the lowest function value) or local (the lowest in a finite neighborhood and not on the boundary of that neighborhood).
f(x)
x1
x2
Minimization of functions
One-dimensional search:Golden section approach
Successive bracketing of a minimum. The minimum is originally bracketed by points 1,3,2. The function is evaluated at 4, which replaces 2; then at 5, which replaces 1; then at 6, which replaces 4. The rule at each stage is to keep a centerpoint that is lower than the two outside points. After the steps shown, the minimum is bracketed by points 5,3,6.
Golden proprtion:
b / (a+b) = a / b ~ 0.618…
The main idea: at each step, calculate only one function value with reducing the interval of uncertainty.
2
1
3
4
5 6
a bc d
Minimization of functions
One-dimensional search:
Golden section with parabolic interpolation (Brent's Method)It is often much more effectivet than the successive golden section.
At any particular stage, it is keeping track of 5 function points (not necessarily all distinct), a, b, u, v, w and x, defined as follows:the minimum is bracketed between a and b;x is the point with the very least function value found so far (or the most recent one in case of a tie); w is the point with the second least function value;v is the previous value of w; u is the point at which the function was evaluated most recently.
a b
2
1
453
At the each particular step we first try to find the minimum of a
parabola passing throug the three best points and use golden section
if the minimum point does not reduce the interval better.
Minimization of functions
Multidimensional search.1. Direct methods (only function values are used): random search, downhill simplex method due to Nelder and Mead2. Gradient methods: conjugate gradients, variable metrix3. Nonlinear Least Squares: algorithm NL2SOL.3. Global minimization
Downhill simplex method due to Nelder and MeadPossible outcomes for a step in the downhill simplex method in a three-dimensional case f(x1,x2,x3): the simplex at the beginning of the step, here a tetrahedron, is shown, top. The simplex at the end of the step can be any one of (a) a reflection away from the high point, (b) a reflection and expansion away from the high point, (c) a contraction along one dimension from the high point, or (d) a contraction along all dimensions towards the low point. An appropriate sequence of such steps will always converge to a minimum of the function.
Sample page from NUMERICAL RECIPES IN FORTRAN 77: THE ART OF SCIENTIFIC COMPUTING (ISBN 0-521-43064-X) Cambridge University Press 1992.
Minimization of functions
Sample page from NUMERICAL RECIPES IN FORTRAN 77: THE ART OF SCIENTIFIC COMPUTING (ISBN 0-521-43064-X) Cambridge University Press 1992.
Minimization of functionsSteepest descent method
(most reasonable but with worst behaviour!):
Successive minimizations along coordinate directions in a long, narrow “valley” (shown as contour lines). Unless the valley is optimally oriented, this method is extremely inefficient, taking many tiny steps to get to the minimum, crossing and re-crossing the principal axis.
Gradient vector
Descent vector
Next gradient
Contour lines, view from the top to F(x)
Minimization of functionsConjugate gradient approach
xAxxgPx ⋅⋅+⋅−≈+∂∂
∂+∂∂+= �� 2
1...21)()(
,
2cxx
xxfx
xfff ji
ji jjii
i
-g :antigradient
vector
Function value at the current
point P
Taylor series approximation of any function:
The matrix A whose components are the second partial derivative matrix of the function is called theHessian matrix of the function at P.
Conjugate conditions for d (vectors of search steps)
jiwith ≠=⋅⋅ 0ji dAd
Quadratcapproximation
���
+==
1-kkkk
00dg-d
g-dβRule for generating directions
{ }��
���
+=+=
+
++
kk
kk1kdxdxx
k1k
1kαα
α
αminLocal search along dk
)g(gd)g(gg
AddAdg
1-kk1-k
1-kkk
1-k1-k
1-kk−
−==kβRule for conjugate gradient multiplier
Minimization of functionsConjuage gradient approach
Gradient vector g0 Descent
vector d0
Next gradient g1
01dβ
Antigradient direction(steepest descent) -g1
Conjugate directiond1 = -g1 + ββββ1d0
Minimization of functionsNewton method: A is known
01-
0min gAxx −=
xAxxgx ⋅⋅+⋅−=21)( cf
01- gA−
Gradient vector g0
No linear search is required for
quadratic functions.
initial step
Newton step
Accumulation of A
step to min along dk
difference gradient
search path
Minimization of functionsQuasi-Newton methods
01-
0min gAxx −=
xAxxgx ⋅⋅+⋅−=21)( cf
�=
×=M
1k kkkk1-
ypppA
=x
kkk1-kkk1-kk
k1-k
kk
kkk ωωpAp
pAppA
ppyyU ×⋅⋅+−×= kϕ
00 gd −=
kkk gAd −=
1-k1-kk UAA +=
kk1k dxx kα+=+
1-kkk ggy −=
1k1-kkk dxxp −−=−= 1kα
Common formula for quasi-Newton corrections:
Search paths for different minimization methodsNelder-Mead
(each 5-th step)
100 iterations/260 function evaluations
Steepest descent (no success)
Conjugate directions
30/217
Quasi-Newton20/140
Algorithm NL2SOL for nonlinear least squares problems
Consider vector function r(x) which values are elements of discepancy between two data sets: yj and F(xj), j=1,…,M. Here xi are the model parameters, i=1,…,N.
j
jji
sFyxr
σ)(
)(−
=
Chi-square is computed as rr ⋅=21)(xf
j
ixr
∂∂=JJacobian matrix
rJrr T ⋅=��
��
∂∂=��
��
⋅∂∂=
∂∂= �
=
M
iirxxx
xfg1
221
21)(
SJJT +=��
�
�
��
∂∂+
��
��
��
���
∂∂
��
��
��
���
∂∂=
∂
∂�=
2
2
12
2 )(
j
ii
T
j
i
j
iM
ij
j
xrr
xr
xr
x
xf
Gradient at the point x
Hessian matrix for f(x)
NL2SOL uses Quasi-Newton scheme where S is accumulated by quasi-Newton formulae and J is computed directly at each step. The approximation to A in this case is much better. This is the most powerful known algorithm for least-squares problems (TENSOLVE is better?).
Global minimization: Simulated annealing
Main idea of the simulated annealing: let the value of f(x) is the objective function. The system state is the point x. The control parameter T is something like a temperature, with an annealing schedule by which it is gradually reduced. And there must be a generator of random changes in the parameter values, a procedure for taking a random step from xi to xi+1 . Offered a succession of options, a simulating annealing method is assumed to jump to the new point with probability P=exp[(fi-fi+1)/kT]. Notice that if fi>fi+1,this probability is greater than unity; in such cases the change is arbitrarily assigned a probability P=1, i.e., the system always took such an option. This general scheme, of always taking a downhill step while sometimes taking an uphill step, has come to be known as the Metropolisalgorithm. In other words, the system sometimes goes uphill as well asdownhill; but the lower the temperature, the less likely is any significant uphill excursion. The schedule of the algorithm is in succesive decreasing the annealing temperature during the search. Finally, it takes only successful steps.Simulated annealing tries “random” steps; but in a long, narrow valley, almost all random steps are uphill and this algorithm is able to accept them thus continuing the minimization.
Simulated annealing approachbased on the downhill simplexmethod due to Nelder and MeadThe implementation of the Metropolis simulated annealing is here as follows: we ADD a [logarithmically] distributed random number, proportional to the temperature T, to the STORED vertex function value each time we look at it, and we SUBTRACT a random number from the function value calculated for every NEW point that is tried as a replacement candidate. This addition forces the simplex to randomly fluctuate keeping its size proportional to the current temperature. The subtraction forces the procedure always to accept a true downhill step, whereas the addition randomly leads to accepting an uphill one. Sometimes, it may help the program to fling out from a local minimum. The annealing itself is in decreasing the temperature during the search, T->0. At T=0 the method is exactly the downhill simplex scheme of Nelder and Mead.
Sample page from NUMERICAL RECIPES IN FORTRAN 77: THE ART OF SCIENTIFIC COMPUTING (ISBN 0-521-43064-X) Cambridge University Press 1992.
Minimization of functions:global search
Fourier transformA physical process can be described either in the time domain, by the values of some quantity h as a function of time t, e.g., h(t), or else in the frequency domain,where the process is specified by giving its amplitude H (generally a complex number indicating phase also) as a function of frequency f, that is H(f), with -1 < f <1. For many purposes it is useful to think of h(t) and H(f) as being two different representations of the same function. One goes back and forth between these two representations by means of the Fourier transform equations,
�∞
∞−= dtethfH iftπ2)()(
�∞
∞−= dfefHth iftπ2)()(
fπω 2=
�∞
∞−= dtethH tiωω )()(
�∞
∞−= ωω
πω deHth ti)(
21)(
If we use angular frequency, in radians:
This is a linear operation
�∞
∞−= dtethH tiωω )()(
h(t) may be considered as coefficients of orthogonal functions which sum over all values of t reproduces the function Therefore, we could find these coefficient by a system of linear equations writing the equation coefficients
for each value of t:
tie ω
)(ωH
kjtie ω
)()( jti
kk
Heth kj ωω =�∞
−∞=This is the basis for different indirect methods of Fourier calculations. The Fourier transform may be represented as decomposition of a function to a sum of other orthogonal functions, such as sin, cos, etc.
h(t1) x
= + h(t2) x
+ h(t3) x
+ …
Case study: scattering from monodisperse systems
Scattering intensity from a dilute solution of chaotically oriented identical particles is proportional to the spherically averaged scattering from a single particle
p(r) is distance distribution function, D is the maximum particle size.
drsr
srrpsID
�=0
sin)(4)( π
Case study: computation of the density profiles
♦ The electron density profile of a centrosymmetric bilayer is computed using one-dimensional Fourier transformation as
where N is the number of peaks and sl and S(l) are the position and the area of the l-th peak, respectively. The Lorentz factor sl reflects random orientation of the lamella microcrystals in PODS. The most plausible combination of signs of the amplitudes A(s,l) is selected by a visual inspection of the density profiles.
)cos(])([)cos(),()(11
srlSssrlsAsrN
ll
N
ll ππρ ±== ��
==
0 2 4 61
2
3
l=2 l=3 l=5 l=6
l=1
lg I, relative
s, nm-1
Internal structure of PODS
Peak fitting for the initial PODS sample
♦ PODS samples display up to five Bragg peaks at sl=2πl/d, where d=5.24±0.03 nm is the bilayer thickness.
♦ The long-range order dimension L=37±3 nmprovides the average size of the POSD lamella crystallite (about seven bilayers)
r, nm
0 1 2 3 4 5
ρ, relative
-6
-4
-2
0
2
4
6
8acetone, self-reductionethanol, self-reductionwater/CPC, NaBH4water/CPC, UV-irradiationPODS+CPCInitial PODS
Density profiles of PODS bilayers♦ Compound: AuCl3
♦ Small particles grow in the central (hydrophilic) part of the bilayer
LiteratureR.J.Larsen, M.L.Marx. An Introduction to Mathematical Statistics and its Applications. Prentice-Hall, Inc. Englewood Cliffs, New Jersey 07632, 1981.Gilbert Streng. Linear Algebra and its Applications.Massachusetts Institute of Technology, Acad.Press, New York-London, 1976.P.E.Gill, W.Murray, M.H.Wright. Practical Optimization. SystemsOptimization Laboratory, Dept. of Operations Research, Stanford University.Acad.Press, London-San Francisco, 1981.NUMERICAL RECIPES IN FORTRAN 77: THE ART OF SCIENTIFIC COMPUTING (ISBN 0-521-43064-X) Cambridge University Press, 1992 (now available via Internet at www.nrc.com).