+ All Categories
Home > Documents > A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation...

A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation...

Date post: 27-Apr-2020
Category:
Upload: others
View: 5 times
Download: 0 times
Share this document with a friend
227
Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions A Gentle Introduction to the Time Complexity Analysis of Evolutionary Algorithms Pietro S. Oliveto Department of Computer Science, University of Sheffield, UK Symposium Series in Computational Intelligence (SSCI 2017) Hilton Hawaiian Village Waikiki, Honolulu, Hawaii, 27 November 2017
Transcript
Page 1: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

A Gentle Introduction to the Time Complexity Analysis ofEvolutionary Algorithms

Pietro S. Oliveto

Department of Computer Science, University of Sheffield, UK

Symposium Series in Computational Intelligence (SSCI 2017)Hilton Hawaiian Village Waikiki, Honolulu, Hawaii,

27 November 2017

Page 2: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Bio-sketch - Dr Pietro S. Oliveto

Senior Lecturer in the Department of Computer Science, atthe University of Sheffield.

Leads the Rigorous Research groupwww.dcs.shef.ac.uk/people/P.Oliveto/rig@RigResearch

Laurea Degree in Computer Science from the University ifCatania, Italy (2005).

PhD in Computer Science (2006-2009), EPSRC PhD+Research Fellow (2009-2010), EPSRC Postdoctoral Fellowin Theoretical Computer Science (2010-2013),Vice-Chancellor Fellow (2013-2016), EPSRC Early CareerFellow (2015-2020).

Guest editor for special issues of Evolutionary Computation(MIT Press, 2015) and Computer Science and Technology(Springer, 2012).

Best paper awards at GECCO 2008, ICARIS 2011, GECCO2014 and best paper nominations at CEC 2009, ECTA2011.

Chair of IEEE CIS Task Force on Theoretical Foundationsof Bio-inspired Computation, Associate Editor for IEEETEVC, EPSRC Board Member.

Page 3: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Aims and Goals of this Tutorial

This tutorial will provide an overview ofthe goals of time complexity analysis of Evolutionary Algorithms (EAs)the most common and effective techniques

You should attend if you wish totheoretically understand the behaviour and performance of the searchalgorithms you designfamiliarise with the techniques used in the time complexity analysis of EAspursue research in the area

enable you or enhance your ability tounderstand theoretically the behaviour of EAs on different problemsperform time complexity analysis of simple EAs on common toy problemsread and understand research papers on the computational complexity ofEAshave the basic skills to start independent research in the area

Page 4: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Introduction to the theory of EAs

Evolutionary Algorithms and Computer Science

Goals of design and analysis of algorithms

1 correctness“does the algorithm always output the correct solution?”

2 computational complexity“how many computational resources are required?”

For Evolutionary Algorithms (General purpose)

1 convergence“Does the EA find the solution in finite time?”

2 time complexity“how long does it take to find the optimum?”(time = n. of fitness function evaluations)

Page 5: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Introduction to the theory of EAs

Brief history

Theoretical studies of Evolutionary Algorithms (EAs), albeit few, have alwaysexisted since the seventies [Goldberg, 1989];

Early studies were concerned with explaining the behaviour rather thananalysing their performance.

Schema Theory was considered fundamental;First proposed to understand the behaviour of the simple GA[Holland, 1992];It cannot explain the performance or limit behaviour of EAs;Building Block Hypothesis was controversial [Reeves and Rowe, 2002];

Convergence results appeared in the nineties [Rudolph, 1998];Related to the time limit behaviour of EAs.

Page 6: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Convergence analysis of EAs

Convergence

Definition

Ideally the EA should find the solution in finite steps with probability 1(visit the global optimum in finite time);

If the solution is held forever after, then the algorithm converges to theoptimum!

Conditions for Convergence ([Rudolph, 1998])

1 There is a positive probability to reach any point in the search space fromany other point

2 The best found solution is never removed from the population (elitism)

Canonical GAs using mutation, crossover and proportional selection DoNot converge!

Elitist variants Do converge!

In practice, is it interesting that an algorithm converges to the optimum?

Most EAs visit the global optimum in finite time (RLS does not!)

How much time?

Page 7: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Convergence analysis of EAs

Convergence

Definition

Ideally the EA should find the solution in finite steps with probability 1(visit the global optimum in finite time);

If the solution is held forever after, then the algorithm converges to theoptimum!

Conditions for Convergence ([Rudolph, 1998])

1 There is a positive probability to reach any point in the search space fromany other point

2 The best found solution is never removed from the population (elitism)

Canonical GAs using mutation, crossover and proportional selection DoNot converge!

Elitist variants Do converge!

In practice, is it interesting that an algorithm converges to the optimum?

Most EAs visit the global optimum in finite time (RLS does not!)

How much time?

Page 8: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Convergence analysis of EAs

Convergence

Definition

Ideally the EA should find the solution in finite steps with probability 1(visit the global optimum in finite time);

If the solution is held forever after, then the algorithm converges to theoptimum!

Conditions for Convergence ([Rudolph, 1998])

1 There is a positive probability to reach any point in the search space fromany other point

2 The best found solution is never removed from the population (elitism)

Canonical GAs using mutation, crossover and proportional selection DoNot converge!

Elitist variants Do converge!

In practice, is it interesting that an algorithm converges to the optimum?

Most EAs visit the global optimum in finite time (RLS does not!)

How much time?

Page 9: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Computational complexity of EAs

Computational Complexity of EAs

P. K. Lehre, 2011

Page 10: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Computational complexity of EAs

Computational Complexity of EAs

P. K. Lehre, 2011

Generally means predicting the resources the algorithm requires:

Usually the computational time: the number of primitive steps;

Usually grows with size of the input;

Usually expressed in asymptotic notation;

Exponential runtime: Inefficient algorithmPolynomial runtime: “Efficient” algorithm

Page 11: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Computational complexity of EAs

Computational Complexity of EAs

P. K. Lehre, 2011

However (EAs):

1 In practice the time for a fitness function evaluation is much higher thanthe rest;

2 EAs are randomised algorithmsThey do not perform the same operations even if the input is the same!They do not output the same result if run twice!

Hence, the runtime of an EA is a random variable Tf .We are interested in:

1 Estimating E(Tf ), the expected runtime of the EA for f ;2 Estimating p(Tf ≤ t), the success probability of the EA in t steps for f .

Page 12: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Computational complexity of EAs

Computational Complexity of EAs

P. K. Lehre, 2011

However (EAs):

1 In practice the time for a fitness function evaluation is much higher thanthe rest;

2 EAs are randomised algorithmsThey do not perform the same operations even if the input is the same!They do not output the same result if run twice!

Hence, the runtime of an EA is a random variable Tf .We are interested in:

1 Estimating E(Tf ), the expected runtime of the EA for f ;

2 Estimating p(Tf ≤ t), the success probability of the EA in t steps for f .

Page 13: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Computational complexity of EAs

Computational Complexity of EAs

P. K. Lehre, 2011

However (EAs):

1 In practice the time for a fitness function evaluation is much higher thanthe rest;

2 EAs are randomised algorithmsThey do not perform the same operations even if the input is the same!They do not output the same result if run twice!

Hence, the runtime of an EA is a random variable Tf .We are interested in:

1 Estimating E(Tf ), the expected runtime of the EA for f ;2 Estimating p(Tf ≤ t), the success probability of the EA in t steps for f .

Page 14: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Computational complexity of EAs

Asymptotic notation

f(n) ∈ O(g(n)) ⇐⇒ ∃ constants c, n0 > 0 st. 0 ≤ f(n)≤cg(n) ∀n ≥ n0

f(n) ∈ Ω(g(n)) ⇐⇒ ∃ constants c, n0 > 0 st. 0 ≤ cg(n)≤f(n) ∀n ≥ n0

f(n) ∈ Θ(g(n)) ⇐⇒ f(n) ∈ O(g(n)) and f(n) ∈ Ω(g(n))

f(n) ∈ o(g(n)) ⇐⇒ limn→∞

f(n)

g(n)= 0

Page 15: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Computational complexity of EAs

Exercise 1: Asymptotic Notation

[Lehre, Tutorial]

Page 16: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Computational complexity of EAs

Motivation Overview

Overview

Goal: Analyze the correctness and performance of EAs;

Difficulties: General purpose, randomised;

EAs find the solution in finite time; (convergence analysis)

How much time? → Derive the expected runtime and the successprobability;

Next

Basic Probability Theory: probability space, random variables,expectations (expected runtime)

Randomised Algorithm Tools: Tail inequalities (success probabilities)

Along the way

Understand that the analysis cannot be done over all functions

Understand why the success probability is important (expected runtimenot always sufficient)

Page 17: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Evolutionary Algorithms

Algorithm ((µ+λ)-EA)1 Let t = 0;

2 Initialize P0 with µ individuals chosen uniformly at random;Repeat

3 Create λ new individuals:1 choose x ∈ Pt uniformly at random;2 flip each bit in x with probability p;

4 Create the new population Pt+1 by choosing the best µ individuals out ofµ+ λ;

5 Let t = t+ 1.Until a stopping condition is fulfilled.

if µ = λ = 1 we get a (1+1)-EA;

p = 1/n is generally considered as best choice[Back, 1993, Droste et al., 1998];

By introducing stochastic selection and crossover we obtain a GeneticAlgorithm(GA)

Page 18: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Evolutionary Algorithms

Algorithm ((µ+λ)-EA)1 Let t = 0;

2 Initialize P0 with µ individuals chosen uniformly at random;Repeat

3 Create λ new individuals:1 choose x ∈ Pt uniformly at random;2 flip each bit in x with probability p;

4 Create the new population Pt+1 by choosing the best µ individuals out ofµ+ λ;

5 Let t = t+ 1.Until a stopping condition is fulfilled.

if µ = λ = 1 we get a (1+1)-EA;

p = 1/n is generally considered as best choice[Back, 1993, Droste et al., 1998];

By introducing stochastic selection and crossover we obtain a GeneticAlgorithm(GA)

Page 19: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Evolutionary Algorithms

Algorithm ((µ+λ)-EA)1 Let t = 0;

2 Initialize P0 with µ individuals chosen uniformly at random;Repeat

3 Create λ new individuals:1 choose x ∈ Pt uniformly at random;2 flip each bit in x with probability p;

4 Create the new population Pt+1 by choosing the best µ individuals out ofµ+ λ;

5 Let t = t+ 1.Until a stopping condition is fulfilled.

if µ = λ = 1 we get a (1+1)-EA;

p = 1/n is generally considered as best choice[Back, 1993, Droste et al., 1998];

By introducing stochastic selection and crossover we obtain a GeneticAlgorithm(GA)

Page 20: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Evolutionary Algorithms

Algorithm ((µ+λ)-EA)1 Let t = 0;

2 Initialize P0 with µ individuals chosen uniformly at random;Repeat

3 Create λ new individuals:1 choose x ∈ Pt uniformly at random;2 flip each bit in x with probability p;

4 Create the new population Pt+1 by choosing the best µ individuals out ofµ+ λ;

5 Let t = t+ 1.Until a stopping condition is fulfilled.

if µ = λ = 1 we get a (1+1)-EA;

p = 1/n is generally considered as best choice[Back, 1993, Droste et al., 1998];

By introducing stochastic selection and crossover we obtain a GeneticAlgorithm(GA)

Page 21: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

1+1-EA

Algorithm ((1+1)-EA)

Initialise P0 with x ∈ 1, 0n by flipping each bit with p = 1/2 ;Repeat

Create x′ by flipping each bit in x with p = 1/n;

If f(x′) ≥ f(x) Then x′ ∈ Pt+1 Else x ∈ Pt+1;

Let t = t+ 1; Until stopping condition.

If only one bit is flipped per iteration: Random Local Search (RLS).How does it work?

Given x, how many bits will flip in expectation?

E[X] = E[X1 +X2 + · · ·+Xn] = E[X1] + E[X2] + · · ·+ E[Xn] =

(E[Xi] = 1 · 1/n+ 0 · (1− 1/n) = 1 · 1/n = 1/n E(X) = np

)

=

n∑i=1

1 · 1/n = n/n = 1

Page 22: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

1+1-EA

Algorithm ((1+1)-EA)

Initialise P0 with x ∈ 1, 0n by flipping each bit with p = 1/2 ;Repeat

Create x′ by flipping each bit in x with p = 1/n;

If f(x′) ≥ f(x) Then x′ ∈ Pt+1 Else x ∈ Pt+1;

Let t = t+ 1; Until stopping condition.

If only one bit is flipped per iteration: Random Local Search (RLS).How does it work?

Given x, how many bits will flip in expectation?

E[X] = E[X1 +X2 + · · ·+Xn] = E[X1] + E[X2] + · · ·+ E[Xn] =

(E[Xi] = 1 · 1/n+ 0 · (1− 1/n) = 1 · 1/n = 1/n E(X) = np

)

=

n∑i=1

1 · 1/n = n/n = 1

Page 23: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

1+1-EA

Algorithm ((1+1)-EA)

Initialise P0 with x ∈ 1, 0n by flipping each bit with p = 1/2 ;Repeat

Create x′ by flipping each bit in x with p = 1/n;

If f(x′) ≥ f(x) Then x′ ∈ Pt+1 Else x ∈ Pt+1;

Let t = t+ 1; Until stopping condition.

If only one bit is flipped per iteration: Random Local Search (RLS).How does it work?

Given x, how many bits will flip in expectation?

E[X] = E[X1 +X2 + · · ·+Xn] = E[X1] + E[X2] + · · ·+ E[Xn] =

(E[Xi] = 1 · 1/n+ 0 · (1− 1/n) = 1 · 1/n = 1/n E(X) = np

)

=

n∑i=1

1 · 1/n = n/n = 1

Page 24: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

1+1-EA

Algorithm ((1+1)-EA)

Initialise P0 with x ∈ 1, 0n by flipping each bit with p = 1/2 ;Repeat

Create x′ by flipping each bit in x with p = 1/n;

If f(x′) ≥ f(x) Then x′ ∈ Pt+1 Else x ∈ Pt+1;

Let t = t+ 1; Until stopping condition.

If only one bit is flipped per iteration: Random Local Search (RLS).How does it work?

Given x, how many bits will flip in expectation?

E[X] = E[X1 +X2 + · · ·+Xn] = E[X1] + E[X2] + · · ·+ E[Xn] =

(E[Xi] = 1 · 1/n+ 0 · (1− 1/n) = 1 · 1/n = 1/n E(X) = np

)

=

n∑i=1

1 · 1/n = n/n = 1

Page 25: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

1+1-EA

Algorithm ((1+1)-EA)

Initialise P0 with x ∈ 1, 0n by flipping each bit with p = 1/2 ;Repeat

Create x′ by flipping each bit in x with p = 1/n;

If f(x′) ≥ f(x) Then x′ ∈ Pt+1 Else x ∈ Pt+1;

Let t = t+ 1; Until stopping condition.

If only one bit is flipped per iteration: Random Local Search (RLS).How does it work?

Given x, how many bits will flip in expectation?

E[X] = E[X1 +X2 + · · ·+Xn] = E[X1] + E[X2] + · · ·+ E[Xn] =

(E[Xi] = 1 · 1/n+ 0 · (1− 1/n) = 1 · 1/n = 1/n E(X) = np

)

=

n∑i=1

1 · 1/n = n/n = 1

Page 26: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General properties

1+1-EA: 2

How likely is it that exactly one bit flips?

(Pr(X = j) =

(nj

)pj(1− p)n−j

)

What is the probability of exactly one bit flipping?

Pr(X = 1) =

(n

1

)· 1/n · (1− 1/n)n−1 = (1− 1/n)n−1 ≥ 1/e ≈ 0.37

Is it more likely that 2 bits flip or none?

Pr(X = 2) =

(n

2

)· 1/n2 · (1− 1/n)n−2 =

=n · (n− 1)

21/n2 · (1− 1/n)n−2 =

= 1/2 · (1− 1/n)n−1 ≈ 1/(2e)

While

Pr(X = 0) =

(n

0

)(1/n)0 · (1− 1/n)n ≈ 1/e

Page 27: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General properties

1+1-EA: 2

How likely is it that exactly one bit flips?

(Pr(X = j) =

(nj

)pj(1− p)n−j

)What is the probability of exactly one bit flipping?

Pr(X = 1) =

(n

1

)· 1/n · (1− 1/n)n−1 = (1− 1/n)n−1 ≥ 1/e ≈ 0.37

Is it more likely that 2 bits flip or none?

Pr(X = 2) =

(n

2

)· 1/n2 · (1− 1/n)n−2 =

=n · (n− 1)

21/n2 · (1− 1/n)n−2 =

= 1/2 · (1− 1/n)n−1 ≈ 1/(2e)

While

Pr(X = 0) =

(n

0

)(1/n)0 · (1− 1/n)n ≈ 1/e

Page 28: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General properties

1+1-EA: 2

How likely is it that exactly one bit flips?

(Pr(X = j) =

(nj

)pj(1− p)n−j

)What is the probability of exactly one bit flipping?

Pr(X = 1) =

(n

1

)· 1/n · (1− 1/n)n−1 = (1− 1/n)n−1 ≥ 1/e ≈ 0.37

Is it more likely that 2 bits flip or none?

Pr(X = 2) =

(n

2

)· 1/n2 · (1− 1/n)n−2 =

=n · (n− 1)

21/n2 · (1− 1/n)n−2 =

= 1/2 · (1− 1/n)n−1 ≈ 1/(2e)

While

Pr(X = 0) =

(n

0

)(1/n)0 · (1− 1/n)n ≈ 1/e

Page 29: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General properties

1+1-EA: 2

How likely is it that exactly one bit flips?

(Pr(X = j) =

(nj

)pj(1− p)n−j

)What is the probability of exactly one bit flipping?

Pr(X = 1) =

(n

1

)· 1/n · (1− 1/n)n−1 = (1− 1/n)n−1 ≥ 1/e ≈ 0.37

Is it more likely that 2 bits flip or none?

Pr(X = 2) =

(n

2

)· 1/n2 · (1− 1/n)n−2 =

=n · (n− 1)

21/n2 · (1− 1/n)n−2 =

= 1/2 · (1− 1/n)n−1 ≈ 1/(2e)

While

Pr(X = 0) =

(n

0

)(1/n)0 · (1− 1/n)n ≈ 1/e

Page 30: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General properties

1+1-EA: 2

How likely is it that exactly one bit flips?

(Pr(X = j) =

(nj

)pj(1− p)n−j

)What is the probability of exactly one bit flipping?

Pr(X = 1) =

(n

1

)· 1/n · (1− 1/n)n−1 = (1− 1/n)n−1 ≥ 1/e ≈ 0.37

Is it more likely that 2 bits flip or none?

Pr(X = 2) =

(n

2

)· 1/n2 · (1− 1/n)n−2 =

=n · (n− 1)

21/n2 · (1− 1/n)n−2 =

= 1/2 · (1− 1/n)n−1 ≈ 1/(2e)

While

Pr(X = 0) =

(n

0

)(1/n)0 · (1− 1/n)n ≈ 1/e

Page 31: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General properties

1+1-EA: 2

How likely is it that exactly one bit flips?

(Pr(X = j) =

(nj

)pj(1− p)n−j

)What is the probability of exactly one bit flipping?

Pr(X = 1) =

(n

1

)· 1/n · (1− 1/n)n−1 = (1− 1/n)n−1 ≥ 1/e ≈ 0.37

Is it more likely that 2 bits flip or none?

Pr(X = 2) =

(n

2

)· 1/n2 · (1− 1/n)n−2 =

=n · (n− 1)

21/n2 · (1− 1/n)n−2 =

= 1/2 · (1− 1/n)n−1 ≈ 1/(2e)

While

Pr(X = 0) =

(n

0

)(1/n)0 · (1− 1/n)n ≈ 1/e

Page 32: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General upper bound

1+1-EA: General Upper bound

Theorem ([Droste et al., 2002])

The expected runtime of the (1+1)-EA for an arbitrary function defined in0, 1n is O(nn)

Proof

1 Let i be the number of bit positions in which the current solution x andthe global optimum x∗ differ;

2 Each bit flips with probability 1/n, hence does not flip with probability(1− 1/n);

3 In order to reach the global optimum the algorithm has to mutate the ibits and leave the n− i bits unchanged;

4 Then:

p(x∗|x) =

(1

n

)i(1− 1

n

)n−i≥(

1

n

)n= n−n

(p = n−n

)5 it implies an upper bound on the expected runtime of O(nn)

(E(X) = 1/p = nn) (waiting time argument).

Page 33: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General upper bound

1+1-EA: General Upper bound

Theorem ([Droste et al., 2002])

The expected runtime of the (1+1)-EA for an arbitrary function defined in0, 1n is O(nn)

Proof

1 Let i be the number of bit positions in which the current solution x andthe global optimum x∗ differ;

2 Each bit flips with probability 1/n, hence does not flip with probability(1− 1/n);

3 In order to reach the global optimum the algorithm has to mutate the ibits and leave the n− i bits unchanged;

4 Then:

p(x∗|x) =

(1

n

)i(1− 1

n

)n−i≥(

1

n

)n= n−n

(p = n−n

)5 it implies an upper bound on the expected runtime of O(nn)

(E(X) = 1/p = nn) (waiting time argument).

Page 34: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General upper bound

1+1-EA: General Upper bound

Theorem ([Droste et al., 2002])

The expected runtime of the (1+1)-EA for an arbitrary function defined in0, 1n is O(nn)

Proof

1 Let i be the number of bit positions in which the current solution x andthe global optimum x∗ differ;

2 Each bit flips with probability 1/n, hence does not flip with probability(1− 1/n);

3 In order to reach the global optimum the algorithm has to mutate the ibits and leave the n− i bits unchanged;

4 Then:

p(x∗|x) =

(1

n

)i(1− 1

n

)n−i≥(

1

n

)n= n−n

(p = n−n

)5 it implies an upper bound on the expected runtime of O(nn)

(E(X) = 1/p = nn) (waiting time argument).

Page 35: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General upper bound

1+1-EA: General Upper bound

Theorem ([Droste et al., 2002])

The expected runtime of the (1+1)-EA for an arbitrary function defined in0, 1n is O(nn)

Proof

1 Let i be the number of bit positions in which the current solution x andthe global optimum x∗ differ;

2 Each bit flips with probability 1/n, hence does not flip with probability(1− 1/n);

3 In order to reach the global optimum the algorithm has to mutate the ibits and leave the n− i bits unchanged;

4 Then:

p(x∗|x) =

(1

n

)i(1− 1

n

)n−i≥(

1

n

)n= n−n

(p = n−n

)5 it implies an upper bound on the expected runtime of O(nn)

(E(X) = 1/p = nn) (waiting time argument).

Page 36: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General upper bound

1+1-EA: General Upper bound

Theorem ([Droste et al., 2002])

The expected runtime of the (1+1)-EA for an arbitrary function defined in0, 1n is O(nn)

Proof

1 Let i be the number of bit positions in which the current solution x andthe global optimum x∗ differ;

2 Each bit flips with probability 1/n, hence does not flip with probability(1− 1/n);

3 In order to reach the global optimum the algorithm has to mutate the ibits and leave the n− i bits unchanged;

4 Then:

p(x∗|x) =

(1

n

)i(1− 1

n

)n−i≥(

1

n

)n= n−n

(p = n−n

)5 it implies an upper bound on the expected runtime of O(nn)

(E(X) = 1/p = nn) (waiting time argument).

Page 37: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General upper bound

1+1-EA: General Upper bound

Theorem ([Droste et al., 2002])

The expected runtime of the (1+1)-EA for an arbitrary function defined in0, 1n is O(nn)

Proof

1 Let i be the number of bit positions in which the current solution x andthe global optimum x∗ differ;

2 Each bit flips with probability 1/n, hence does not flip with probability(1− 1/n);

3 In order to reach the global optimum the algorithm has to mutate the ibits and leave the n− i bits unchanged;

4 Then:

p(x∗|x) =

(1

n

)i(1− 1

n

)n−i≥(

1

n

)n= n−n

(p = n−n

)

5 it implies an upper bound on the expected runtime of O(nn)(E(X) = 1/p = nn) (waiting time argument).

Page 38: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General upper bound

1+1-EA: General Upper bound

Theorem ([Droste et al., 2002])

The expected runtime of the (1+1)-EA for an arbitrary function defined in0, 1n is O(nn)

Proof

1 Let i be the number of bit positions in which the current solution x andthe global optimum x∗ differ;

2 Each bit flips with probability 1/n, hence does not flip with probability(1− 1/n);

3 In order to reach the global optimum the algorithm has to mutate the ibits and leave the n− i bits unchanged;

4 Then:

p(x∗|x) =

(1

n

)i(1− 1

n

)n−i≥(

1

n

)n= n−n

(p = n−n

)5 it implies an upper bound on the expected runtime of O(nn)

(E(X) = 1/p = nn) (waiting time argument).

Page 39: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General upper bound

General Upper bound Exercises

Theorem

The expected runtime of the (1+1)-EA with mutation probability p = 1/2 foran arbitrary function defined in 0, 1n is O(2n)

Proof Left as Exercise.

Theorem

The expected runtime of the (1+1)-EA with mutation probability p = χ/n foran arbitrary function defined in 0, 1n is O((n/χ)n)

Proof Left as Exercise.

Theorem

The expected runtime of RLS for an arbitrary function defined in 0, 1n isinfinite.

Proof Left as Exercise.

Page 40: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General upper bound

General Upper bound Exercises

Theorem

The expected runtime of the (1+1)-EA with mutation probability p = 1/2 foran arbitrary function defined in 0, 1n is O(2n)

Proof Left as Exercise.

Theorem

The expected runtime of the (1+1)-EA with mutation probability p = χ/n foran arbitrary function defined in 0, 1n is O((n/χ)n)

Proof Left as Exercise.

Theorem

The expected runtime of RLS for an arbitrary function defined in 0, 1n isinfinite.

Proof Left as Exercise.

Page 41: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General upper bound

General Upper bound Exercises

Theorem

The expected runtime of the (1+1)-EA with mutation probability p = 1/2 foran arbitrary function defined in 0, 1n is O(2n)

Proof Left as Exercise.

Theorem

The expected runtime of the (1+1)-EA with mutation probability p = χ/n foran arbitrary function defined in 0, 1n is O((n/χ)n)

Proof Left as Exercise.

Theorem

The expected runtime of RLS for an arbitrary function defined in 0, 1n isinfinite.

Proof Left as Exercise.

Page 42: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General upper bound

General Upper bound Exercises

Theorem

The expected runtime of the (1+1)-EA with mutation probability p = 1/2 foran arbitrary function defined in 0, 1n is O(2n)

Proof Left as Exercise.

Theorem

The expected runtime of the (1+1)-EA with mutation probability p = χ/n foran arbitrary function defined in 0, 1n is O((n/χ)n)

Proof Left as Exercise.

Theorem

The expected runtime of RLS for an arbitrary function defined in 0, 1n isinfinite.

Proof Left as Exercise.

Page 43: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General upper bound

1+1-EA: Conclusions & Exercises

In general:

P (i− bitflip) =

(n

i

)1

ni

(1− 1

n

)n−i≤ 1

i!

(1− 1

n

)n−i≈ 1

i!e

What about RLS?

Expectation: E[X] = 1

P(1-bitflip) = 1

What about initialisation?

How many one-bits in expectation after initialisation?

E[X] = n · 1/2 = n/2

How likely is it that we get exactly n/2 one-bits?

Pr(X = n/2) =

(n

n/2

)1

nn/2

(1− 1

n

)n/2(n = 100, P r(X = 50) ≈ 0.0796

)

Tail Inequalities help us!

Page 44: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General upper bound

1+1-EA: Conclusions & Exercises

In general:

P (i− bitflip) =

(n

i

)1

ni

(1− 1

n

)n−i≤ 1

i!

(1− 1

n

)n−i≈ 1

i!e

What about RLS?

Expectation: E[X] = 1

P(1-bitflip) = 1

What about initialisation?

How many one-bits in expectation after initialisation?

E[X] = n · 1/2 = n/2

How likely is it that we get exactly n/2 one-bits?

Pr(X = n/2) =

(n

n/2

)1

nn/2

(1− 1

n

)n/2(n = 100, P r(X = 50) ≈ 0.0796

)

Tail Inequalities help us!

Page 45: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General upper bound

1+1-EA: Conclusions & Exercises

In general:

P (i− bitflip) =

(n

i

)1

ni

(1− 1

n

)n−i≤ 1

i!

(1− 1

n

)n−i≈ 1

i!e

What about RLS?

Expectation: E[X] = 1

P(1-bitflip) = 1

What about initialisation?

How many one-bits in expectation after initialisation?

E[X] = n · 1/2 = n/2

How likely is it that we get exactly n/2 one-bits?

Pr(X = n/2) =

(n

n/2

)1

nn/2

(1− 1

n

)n/2(n = 100, P r(X = 50) ≈ 0.0796

)

Tail Inequalities help us!

Page 46: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General upper bound

1+1-EA: Conclusions & Exercises

In general:

P (i− bitflip) =

(n

i

)1

ni

(1− 1

n

)n−i≤ 1

i!

(1− 1

n

)n−i≈ 1

i!e

What about RLS?

Expectation: E[X] = 1

P(1-bitflip) = 1

What about initialisation?

How many one-bits in expectation after initialisation?

E[X] = n · 1/2 = n/2

How likely is it that we get exactly n/2 one-bits?

Pr(X = n/2) =

(n

n/2

)1

nn/2

(1− 1

n

)n/2(n = 100, P r(X = 50) ≈ 0.0796

)

Tail Inequalities help us!

Page 47: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General upper bound

1+1-EA: Conclusions & Exercises

In general:

P (i− bitflip) =

(n

i

)1

ni

(1− 1

n

)n−i≤ 1

i!

(1− 1

n

)n−i≈ 1

i!e

What about RLS?

Expectation: E[X] = 1

P(1-bitflip) = 1

What about initialisation?

How many one-bits in expectation after initialisation?

E[X] = n · 1/2 = n/2

How likely is it that we get exactly n/2 one-bits?

Pr(X = n/2) =

(n

n/2

)1

nn/2

(1− 1

n

)n/2(n = 100, P r(X = 50) ≈ 0.0796

)

Tail Inequalities help us!

Page 48: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General upper bound

1+1-EA: Conclusions & Exercises

In general:

P (i− bitflip) =

(n

i

)1

ni

(1− 1

n

)n−i≤ 1

i!

(1− 1

n

)n−i≈ 1

i!e

What about RLS?

Expectation: E[X] = 1

P(1-bitflip) = 1

What about initialisation?

How many one-bits in expectation after initialisation?

E[X] = n · 1/2 = n/2

How likely is it that we get exactly n/2 one-bits?

Pr(X = n/2) =

(n

n/2

)1

nn/2

(1− 1

n

)n/2(n = 100, P r(X = 50) ≈ 0.0796

)

Tail Inequalities help us!

Page 49: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General upper bound

1+1-EA: Conclusions & Exercises

In general:

P (i− bitflip) =

(n

i

)1

ni

(1− 1

n

)n−i≤ 1

i!

(1− 1

n

)n−i≈ 1

i!e

What about RLS?

Expectation: E[X] = 1

P(1-bitflip) = 1

What about initialisation?

How many one-bits in expectation after initialisation?

E[X] = n · 1/2 = n/2

How likely is it that we get exactly n/2 one-bits?

Pr(X = n/2) =

(n

n/2

)1

nn/2

(1− 1

n

)n/2(n = 100, P r(X = 50) ≈ 0.0796

)

Tail Inequalities help us!

Page 50: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

General upper bound

1+1-EA: Conclusions & Exercises

In general:

P (i− bitflip) =

(n

i

)1

ni

(1− 1

n

)n−i≤ 1

i!

(1− 1

n

)n−i≈ 1

i!e

What about RLS?

Expectation: E[X] = 1

P(1-bitflip) = 1

What about initialisation?

How many one-bits in expectation after initialisation?

E[X] = n · 1/2 = n/2

How likely is it that we get exactly n/2 one-bits?

Pr(X = n/2) =

(n

n/2

)1

nn/2

(1− 1

n

)n/2(n = 100, P r(X = 50) ≈ 0.0796

)

Tail Inequalities help us!

Page 51: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Tail Inequalities

Given a random variable X it may assume values that are considerably larger orlower than its expectation;

Tail inequalities:

The expectation can often be estimate easily;

We would like to know the probability of deviating far from theexpectation i.e., the “tails” of the distribution

Tail inequalities give bounds on the tails given the expectation.

Page 52: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Markov’s inequality

Markov Inequality

The fundamental inequality from which many others are derived.

Definition (Markov’s Inequality)

Let X be a random variable assuming only non-negative values, and E[X] itsexpectation. Then for all t ∈ R+,

Pr[X ≥ t] ≤ E[X]

t.

E[X] = 1; then: Pr[X ≥ 2] ≤ E[X]2≤ 1

2(Number of bits that flip)

E[X] = n/2; then Pr[X ≥ (2/3)n] ≤ E[X](2/3)n

= n/2(2/3)n

= 34

(Number of one-bits after initialisation)

Markov’s inequality is often used iteratively in repeated phases to obtainstronger bounds!

Page 53: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Markov’s inequality

Markov Inequality

The fundamental inequality from which many others are derived.

Definition (Markov’s Inequality)

Let X be a random variable assuming only non-negative values, and E[X] itsexpectation. Then for all t ∈ R+,

Pr[X ≥ t] ≤ E[X]

t.

E[X] = 1; then: Pr[X ≥ 2] ≤ E[X]2≤ 1

2(Number of bits that flip)

E[X] = n/2; then Pr[X ≥ (2/3)n] ≤ E[X](2/3)n

= n/2(2/3)n

= 34

(Number of one-bits after initialisation)

Markov’s inequality is often used iteratively in repeated phases to obtainstronger bounds!

Page 54: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Markov’s inequality

Markov Inequality

The fundamental inequality from which many others are derived.

Definition (Markov’s Inequality)

Let X be a random variable assuming only non-negative values, and E[X] itsexpectation. Then for all t ∈ R+,

Pr[X ≥ t] ≤ E[X]

t.

E[X] = 1; then: Pr[X ≥ 2] ≤ E[X]2≤ 1

2(Number of bits that flip)

E[X] = n/2; then Pr[X ≥ (2/3)n] ≤ E[X](2/3)n

= n/2(2/3)n

= 34

(Number of one-bits after initialisation)

Markov’s inequality is often used iteratively in repeated phases to obtainstronger bounds!

Page 55: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Markov’s inequality

Markov Inequality

The fundamental inequality from which many others are derived.

Definition (Markov’s Inequality)

Let X be a random variable assuming only non-negative values, and E[X] itsexpectation. Then for all t ∈ R+,

Pr[X ≥ t] ≤ E[X]

t.

E[X] = 1; then: Pr[X ≥ 2] ≤ E[X]2≤ 1

2(Number of bits that flip)

E[X] = n/2; then Pr[X ≥ (2/3)n] ≤ E[X](2/3)n

= n/2(2/3)n

= 34

(Number of one-bits after initialisation)

Markov’s inequality is often used iteratively in repeated phases to obtainstronger bounds!

Page 56: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Markov’s inequality

Markov Inequality

The fundamental inequality from which many others are derived.

Definition (Markov’s Inequality)

Let X be a random variable assuming only non-negative values, and E[X] itsexpectation. Then for all t ∈ R+,

Pr[X ≥ t] ≤ E[X]

t.

E[X] = 1; then: Pr[X ≥ 2] ≤ E[X]2≤ 1

2(Number of bits that flip)

E[X] = n/2; then Pr[X ≥ (2/3)n] ≤ E[X](2/3)n

= n/2(2/3)n

= 34

(Number of one-bits after initialisation)

Markov’s inequality is often used iteratively in repeated phases to obtainstronger bounds!

Page 57: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Chernoff bounds

Chernoff Bounds

Let X1, X2, . . . Xn be independent Poisson trials each with probability pi;For X =

∑ni=1 Xi the expectation is E(X) =

∑ni=1 pi.

Definition (Chernoff Bounds)

1 for 0 ≤ δ ≤ 1, Pr(X ≤ (1− δ)E[X]) ≤ e−E[X]δ2

2 .

2 for δ > 0, Pr(X > (1 + δ)E[X]) ≤[

(1+δ)1+δ

]E[X]

.

What is the probability that we have more than (2/3)n one-bits atinitialisation?

pi = 1/2, E[X] = n · 1/2 = n/2,(we fix δ = 1/3→ (1 + δ)E[X] = (2/3)n); then:

Pr[X > (2/3)n] ≤(

e1/3

(4/3)4/3

)n/2= c−n/2

Page 58: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Chernoff bounds

Chernoff Bounds

Let X1, X2, . . . Xn be independent Poisson trials each with probability pi;For X =

∑ni=1 Xi the expectation is E(X) =

∑ni=1 pi.

Definition (Chernoff Bounds)

1 for 0 ≤ δ ≤ 1, Pr(X ≤ (1− δ)E[X]) ≤ e−E[X]δ2

2 .

2 for δ > 0, Pr(X > (1 + δ)E[X]) ≤[

(1+δ)1+δ

]E[X]

.

What is the probability that we have more than (2/3)n one-bits atinitialisation?

pi = 1/2, E[X] = n · 1/2 = n/2,(we fix δ = 1/3→ (1 + δ)E[X] = (2/3)n); then:

Pr[X > (2/3)n] ≤(

e1/3

(4/3)4/3

)n/2= c−n/2

Page 59: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Chernoff bounds

Chernoff Bounds

Let X1, X2, . . . Xn be independent Poisson trials each with probability pi;For X =

∑ni=1 Xi the expectation is E(X) =

∑ni=1 pi.

Definition (Chernoff Bounds)

1 for 0 ≤ δ ≤ 1, Pr(X ≤ (1− δ)E[X]) ≤ e−E[X]δ2

2 .

2 for δ > 0, Pr(X > (1 + δ)E[X]) ≤[

(1+δ)1+δ

]E[X]

.

What is the probability that we have more than (2/3)n one-bits atinitialisation?

pi = 1/2, E[X] = n · 1/2 = n/2,

(we fix δ = 1/3→ (1 + δ)E[X] = (2/3)n); then:

Pr[X > (2/3)n] ≤(

e1/3

(4/3)4/3

)n/2= c−n/2

Page 60: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Chernoff bounds

Chernoff Bounds

Let X1, X2, . . . Xn be independent Poisson trials each with probability pi;For X =

∑ni=1 Xi the expectation is E(X) =

∑ni=1 pi.

Definition (Chernoff Bounds)

1 for 0 ≤ δ ≤ 1, Pr(X ≤ (1− δ)E[X]) ≤ e−E[X]δ2

2 .

2 for δ > 0, Pr(X > (1 + δ)E[X]) ≤[

(1+δ)1+δ

]E[X]

.

What is the probability that we have more than (2/3)n one-bits atinitialisation?

pi = 1/2, E[X] = n · 1/2 = n/2,(we fix δ = 1/3→ (1 + δ)E[X] = (2/3)n); then:

Pr[X > (2/3)n] ≤(

e1/3

(4/3)4/3

)n/2= c−n/2

Page 61: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Chernoff bounds

Chernoff Bound Simple Application

Bitstring of length n = 100

Pr(Xi) = 1/2 and E(X) = np = 100/2 = 50.

What is the probability to have at least 75 1-bits?

Markov: Pr(X ≥ 75) ≤ 5075

= 23

Chernoff: Pr(X ≥ (1 + 1/2)50) ≤(

√e

(3/2)3/2

)50

< 0.0045

Truth: Pr(X ≥ 75) =∑100i=75

(100i

)2−100 < 0.000000282

Page 62: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Chernoff bounds

Chernoff Bound Simple Application

Bitstring of length n = 100

Pr(Xi) = 1/2 and E(X) = np = 100/2 = 50.What is the probability to have at least 75 1-bits?

Markov: Pr(X ≥ 75) ≤ 5075

= 23

Chernoff: Pr(X ≥ (1 + 1/2)50) ≤(

√e

(3/2)3/2

)50

< 0.0045

Truth: Pr(X ≥ 75) =∑100i=75

(100i

)2−100 < 0.000000282

Page 63: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Chernoff bounds

Chernoff Bound Simple Application

Bitstring of length n = 100

Pr(Xi) = 1/2 and E(X) = np = 100/2 = 50.What is the probability to have at least 75 1-bits?

Markov: Pr(X ≥ 75) ≤ 5075

= 23

Chernoff: Pr(X ≥ (1 + 1/2)50) ≤(

√e

(3/2)3/2

)50

< 0.0045

Truth: Pr(X ≥ 75) =∑100i=75

(100i

)2−100 < 0.000000282

Page 64: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Chernoff bounds

OneMax

OneMax(x)=∑ni=1 x[i])

ones(x)

f(x)

1

1

2

2

n

n

Page 65: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

0 1 2 3 4 5

0 0 0 0 0 0 p0 = 66

E(T0) = 66

Page 66: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

50 1 2 3 4 5

0 0 0 0 0 0 p0 = 66

E(T0) = 66

Page 67: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

0 1 2 3 4 5

0 0 0 0 0 1 p0 = 66

E(T0) = 66

Page 68: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

0 1 2 3 4 5

0 0 0 0 0 1 p0 = 66

E(T0) = 66

0 1 2 3 4 5

0 0 0 0 0 1 p0 = 66

E(T0) = 66

Page 69: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

0 1 2 3 4 5

0 0 0 0 0 1 p0 = 66

E(T0) = 66

20 1 2 3 4 5

0 0 0 0 0 1 p1 = 56

E(T1) = 65

Page 70: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

0 1 2 3 4 5

0 0 0 0 0 1 p0 = 66

E(T0) = 66

0 1 2 3 4 5

0 0 1 0 0 1 p1 = 56

E(T1) = 65

Page 71: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

0 1 2 3 4 5

0 0 0 0 0 1 p0 = 66

E(T0) = 66

0 1 2 3 4 5

0 0 1 0 0 1 p1 = 56

E(T1) = 65

0 1 2 3 4 5

0 0 1 0 0 1 p1 = 56

E(T1) = 65

Page 72: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

0 1 2 3 4 5

0 0 0 0 0 1 p0 = 66

E(T0) = 66

0 1 2 3 4 5

0 0 1 0 0 1 p1 = 56

E(T1) = 65

00 1 2 3 4 5

0 0 1 0 0 1 p2 = 46

E(T2) = 64

Page 73: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

0 1 2 3 4 5

0 0 0 0 0 1 p0 = 66

E(T0) = 66

0 1 2 3 4 5

0 0 1 0 0 1 p1 = 56

E(T1) = 65

0 1 2 3 4 5

1 0 1 0 0 1 p2 = 46

E(T2) = 64

Page 74: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

0 1 2 3 4 5

0 0 0 0 0 1 p0 = 66

E(T0) = 66

0 1 2 3 4 5

0 0 1 0 0 1 p1 = 56

E(T1) = 65

0 1 2 3 4 5

1 0 1 0 0 1 p2 = 46

E(T2) = 64

0 1 2 3 4 5

1 0 1 0 0 1 p2 = 46

E(T2) = 64

Page 75: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

0 1 2 3 4 5

0 0 0 0 0 1 p0 = 66

E(T0) = 66

0 1 2 3 4 5

0 0 1 0 0 1 p1 = 56

E(T1) = 65

0 1 2 3 4 5

1 0 1 0 0 1 p2 = 46

E(T2) = 64

50 1 2 3 4 5

1 0 1 0 0 1 p3 = 36

E(T0) = 63

Page 76: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

0 1 2 3 4 5

0 0 0 0 0 1 p0 = 66

E(T0) = 66

0 1 2 3 4 5

0 0 1 0 0 1 p1 = 56

E(T1) = 65

0 1 2 3 4 5

1 0 1 0 0 1 p2 = 46

E(T2) = 64

0 1 2 3 4 5

1 0 1 0 0 0 p3 = 36

E(T3) = 63

Page 77: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

0 1 2 3 4 5

0 0 0 0 0 1 p0 = 66

E(T0) = 66

0 1 2 3 4 5

0 0 1 0 0 1 p1 = 56

E(T1) = 65

0 1 2 3 4 5

1 0 1 0 0 1 p2 = 46

E(T2) = 64

0 1 2 3 4 5

1 0 1 0 0 1 p2 = 46

E(T2) = 64

Page 78: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

0 1 2 3 4 5

0 0 0 0 0 1 p0 = 66

E(T0) = 66

0 1 2 3 4 5

0 0 1 0 0 1 p1 = 56

E(T1) = 65

0 1 2 3 4 5

1 0 1 0 0 1 p2 = 46

E(T2) = 64

40 1 2 3 4 5

1 0 1 0 0 1 p3 = 36

E(T3) = 63

Page 79: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

0 1 2 3 4 5

0 0 0 0 0 1 p0 = 66

E(T0) = 66

0 1 2 3 4 5

0 0 1 0 0 1 p1 = 56

E(T1) = 65

0 1 2 3 4 5

1 0 1 0 0 1 p2 = 46

E(T2) = 64

0 1 2 3 4 5

1 0 1 0 1 1 p3 = 36

E(T3) = 63

Page 80: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

0 1 2 3 4 5

0 0 0 0 0 1 p0 = 66

E(T0) = 66

0 1 2 3 4 5

0 0 1 0 0 1 p1 = 56

E(T1) = 65

0 1 2 3 4 5

1 0 1 0 0 1 p2 = 46

E(T2) = 64

0 1 2 3 4 5

1 0 1 0 1 1 p3 = 36

E(T3) = 63

0 1 2 3 4 5

1 0 1 0 1 1 p3 = 36

E(T3) = 63

Page 81: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

0 1 2 3 4 5

0 0 0 0 0 1 p0 = 66

E(T0) = 66

0 1 2 3 4 5

0 0 1 0 0 1 p1 = 56

E(T1) = 65

0 1 2 3 4 5

1 0 1 0 0 1 p2 = 46

E(T2) = 64

0 1 2 3 4 5

1 0 1 0 1 1 p3 = 36

E(T3) = 63

10 1 2 3 4 5

1 0 1 0 1 1 p4 = 26

E(T4) = 62

Page 82: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

0 1 2 3 4 5

0 0 0 0 0 1 p0 = 66

E(T0) = 66

0 1 2 3 4 5

0 0 1 0 0 1 p1 = 56

E(T1) = 65

0 1 2 3 4 5

1 0 1 0 0 1 p2 = 46

E(T2) = 64

0 1 2 3 4 5

1 0 1 0 1 1 p3 = 36

E(T3) = 63

0 1 2 3 4 5

1 1 1 0 1 1 p4 = 26

E(T4) = 62

Page 83: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

0 1 2 3 4 5

0 0 0 0 0 1 p0 = 66

E(T0) = 66

0 1 2 3 4 5

0 0 1 0 0 1 p1 = 56

E(T1) = 65

0 1 2 3 4 5

1 0 1 0 0 1 p2 = 46

E(T2) = 64

0 1 2 3 4 5

1 0 1 0 1 1 p3 = 36

E(T3) = 63

0 1 2 3 4 5

1 1 1 0 1 1 p4 = 26

E(T4) = 62

0 1 2 3 4 5

1 1 1 0 1 1 p4 = 26

E(T4) = 62

Page 84: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

0 1 2 3 4 5

0 0 0 0 0 1 p0 = 66

E(T0) = 66

0 1 2 3 4 5

0 0 1 0 0 1 p1 = 56

E(T1) = 65

0 1 2 3 4 5

1 0 1 0 0 1 p2 = 46

E(T2) = 64

0 1 2 3 4 5

1 0 1 0 1 1 p3 = 36

E(T3) = 63

0 1 2 3 4 5

1 1 1 0 1 1 p4 = 26

E(T4) = 62

30 1 2 3 4 5

1 1 1 0 1 1 p5 = 16

E(T5) = 61

Page 85: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

0 1 2 3 4 5

0 0 0 0 0 1 p0 = 66

E(T0) = 66

0 1 2 3 4 5

0 0 1 0 0 1 p1 = 56

E(T1) = 65

0 1 2 3 4 5

1 0 1 0 0 1 p2 = 46

E(T2) = 64

0 1 2 3 4 5

1 0 1 0 1 1 p3 = 36

E(T3) = 63

0 1 2 3 4 5

1 1 1 0 1 1 p4 = 26

E(T4) = 62

0 1 2 3 4 5

1 1 1 1 1 1 p5 = 16

E(T5) = 61

Page 86: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i])

0 1 2 3 4 5

0 0 0 0 0 1 p0 = 66

E(T0) = 66

0 1 2 3 4 5

0 0 1 0 0 1 p1 = 56

E(T1) = 65

0 1 2 3 4 5

1 0 1 0 0 1 p2 = 46

E(T2) = 64

0 1 2 3 4 5

1 0 1 0 1 1 p3 = 36

E(T3) = 63

0 1 2 3 4 5

1 1 1 0 1 1 p4 = 26

E(T4) = 62

0 1 2 3 4 5

1 1 1 1 1 1 p5 = 16

E(T5) = 61

E(T ) = E(T0) + E(T1) + · · ·+ E(T5) = 1/p0 + 1/p1 + · · ·+ 1/p5 =

=

5∑i=0

1

pi=

5∑i=0

6

i= 6

6∑i=1

1

i= 6 · 2.45 = 14.7

Page 87: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i]) : Generalisation

0 1 2 3 n0 0 0 0 0 0 0 0 0 0 p0 = n

nE(T0) = n

n

Page 88: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i]) : Generalisation

0 1 2 3 5 n0 0 0 0 0 0 0 0 0 0 p0 = n

nE(T0) = n

n

Page 89: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i]) : Generalisation

0 1 2 3 n0 0 0 0 0 1 0 0 0 0 p0 = n

nE(T0) = n

n

Page 90: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i]) : Generalisation

0 1 2 3 n0 0 0 0 0 1 0 0 0 0 p0 = n

nE(T0) = n

n

0 1 2 3 n0 0 0 0 0 1 0 0 0 0 p0 = n

nE(T0) = n

n

Page 91: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i]) : Generalisation

0 1 2 3 n0 0 0 0 0 1 0 0 0 0 p0 = n

nE(T0) = n

n

20 1 2 3 n0 0 0 0 0 1 0 0 0 0 p1 = n−1

nE(T1) = n

n−1

Page 92: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i]) : Generalisation

0 1 2 3 n0 0 0 0 0 1 0 0 0 0 p0 = n

nE(T0) = n

n

0 1 2 3 n0 0 1 0 0 1 0 0 0 0 p1 = n−1

nE(T1) = n

n−1

Page 93: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i]) : Generalisation

0 1 2 3 n0 0 0 0 0 1 0 0 0 0 p0 = n

nE(T0) = n

n

0 1 2 3 n0 0 1 0 0 1 0 0 0 0 p1 = n−1

nE(T1) = n

n−1

0 1 2 3 n0 0 1 0 0 1 0 0 0 0 p1 = n−1

nE(T1) = n

n−1

Page 94: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i]) : Generalisation

0 1 2 3 n0 0 0 0 0 1 0 0 0 0 p0 = n

nE(T0) = n

n

0 1 2 3 n0 0 1 0 0 1 0 0 0 0 p1 = n−1

nE(T1) = n

n−1

n0 1 2 3

0 0 1 0 0 1 0 0 0 0 p2 = n−2n

E(T2) = nn−2

Page 95: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i]) : Generalisation

0 1 2 3 n0 0 0 0 0 1 0 0 0 0 p0 = n

nE(T0) = n

n

0 1 2 3 n0 0 1 0 0 1 0 0 0 0 p1 = n−1

nE(T1) = n

n−1

0 1 2 3 n0 0 1 0 0 1 0 0 0 1 p2 = n−2

nE(T2) = n

n−2

Page 96: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i]) : Generalisation

0 1 2 3 n0 0 0 0 0 1 0 0 0 0 p0 = n

nE(T0) = n

n

0 1 2 3 n0 0 1 0 0 1 0 0 0 0 p1 = n−1

nE(T1) = n

n−1

0 1 2 3 n0 0 1 0 0 1 0 0 0 1 p2 = n−2

nE(T2) = n

n−2

0 1 2 3 n1 1 1 1 0 1 1 1 1 1 pn−1 = 1

nE(Tn−1) = n

1

Page 97: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i]) : Generalisation

0 1 2 3 n0 0 0 0 0 1 0 0 0 0 p0 = n

nE(T0) = n

n

0 1 2 3 n0 0 1 0 0 1 0 0 0 0 p1 = n−1

nE(T1) = n

n−1

0 1 2 3 n0 0 1 0 0 1 0 0 0 1 p2 = n−2

nE(T2) = n

n−2

40 1 2 3 n1 1 1 1 0 1 1 1 1 1 pn−1 = 1

nE(Tn−1) = n

1

Page 98: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i]) : Generalisation

0 1 2 3 n0 0 0 0 0 1 0 0 0 0 p0 = n

nE(T0) = n

n

0 1 2 3 n0 0 1 0 0 1 0 0 0 0 p1 = n−1

nE(T1) = n

n−1

0 1 2 3 n0 0 1 0 0 1 0 0 0 1 p2 = n−2

nE(T2) = n

n−2

0 1 2 3 n1 1 1 1 1 1 1 1 1 1 pn−1 = 1

nE(Tn−1) = n

1

Page 99: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

RLS for OneMax( OneMax(x)=∑n

i=1 x[i]) : Generalisation

0 1 2 3 n0 0 0 0 0 1 0 0 0 0 p0 = n

nE(T0) = n

n

0 1 2 3 n0 0 1 0 0 1 0 0 0 0 p1 = n−1

nE(T1) = n

n−1

0 1 2 3 n0 0 1 0 0 1 0 0 0 1 p2 = n−2

nE(T2) = n

n−2

0 1 2 3 n1 1 1 1 1 1 1 1 1 1 pn−1 = 1

nE(Tn−1) = n

1

E(T ) = E(T0) + E(T1) + · · ·+ E(Tn−1) = 1/p1 + 1/p2 + · · ·+ 1/pn−1 =

=

n−1∑i=0

1

pi=

n∑i=1

n

i= n

n∑i=1

1

i= n ·H(n) = n logn+ Θ(n) = O(n logn)

Page 100: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Coupon collector’s problem

Coupon collector’s problem

The Coupon collector’s problemThere are n types of coupons and at each trial one coupon is chosen at random.Each coupon has the same probability of being extracted. The goal is to findthe exact number of trials before the collector has obtained all the n coupons.

Theorem (The coupon collector’s Theorem)

Let T be the time for all the n coupons to be collected. Then

E(T ) =

n−1∑i=0

1

pi+1=

n−1∑i=0

n

n− i = n

n−1∑i=0

1

i=

= n(logn+ Θ(1)) = n logn+O(n).

Page 101: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Coupon collector’s problem

Coupon collector’s problem: Upper bound on time

What is the probability that the time to collect n coupons is greater thann lnn+O(n)?

Theorem (Coupon collector upper bound on time)

Let T be the time for all the n coupons to be collected. Then

Pr(T ≥ (1 + ε)n lnn) ≤ n−ε

Proof

1n

Probability of choosing a given coupon1− 1

nProbability of not choosing a given coupon(

1− 1n

)tProbability of not choosing a given coupon for t rounds

The probability that one of the n coupons is not chosen in t rounds is less than

n ·(

1− 1n

)t(Union Bound)

Hence, for t = cn lnn

Pr(T ≥ cn lnn) ≤ n(1− 1/n)cn lnn ≤ n · e−c lnn = n · n−c = n−c+1

Page 102: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Coupon collector’s problem

Coupon collector’s problem: Upper bound on time

What is the probability that the time to collect n coupons is greater thann lnn+O(n)?

Theorem (Coupon collector upper bound on time)

Let T be the time for all the n coupons to be collected. Then

Pr(T ≥ (1 + ε)n lnn) ≤ n−ε

Proof

1n

Probability of choosing a given coupon1− 1

nProbability of not choosing a given coupon(

1− 1n

)tProbability of not choosing a given coupon for t rounds

The probability that one of the n coupons is not chosen in t rounds is less than

n ·(

1− 1n

)t(Union Bound)

Hence, for t = cn lnn

Pr(T ≥ cn lnn) ≤ n(1− 1/n)cn lnn ≤ n · e−c lnn = n · n−c = n−c+1

Page 103: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Coupon collector’s problem

Coupon collector’s problem: lower bound on time

What is the probability that the time to collect n coupons is less thann lnn+O(n)?

Theorem (Coupon collector lower bound on time (Doerr, 2011))

Let T be the time for all the n coupons to be collected. Then for all ε > 0

Pr(T < (1− ε)(n− 1) lnn) ≤ exp(−nε)

Corollary

The expected time for RLS to optimise OneMaxis Θ(n lnn). Furthermore,

Pr(T ≥ (1 + ε)n lnn) ≤ n−ε

andPr(T < (1− ε)(n− 1) lnn) ≤ exp(−nε)

What about the (1+1)-EA?

Page 104: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Coupon collector’s problem

Coupon collector’s problem: lower bound on time

What is the probability that the time to collect n coupons is less thann lnn+O(n)?

Theorem (Coupon collector lower bound on time (Doerr, 2011))

Let T be the time for all the n coupons to be collected. Then for all ε > 0

Pr(T < (1− ε)(n− 1) lnn) ≤ exp(−nε)

Corollary

The expected time for RLS to optimise OneMaxis Θ(n lnn). Furthermore,

Pr(T ≥ (1 + ε)n lnn) ≤ n−ε

andPr(T < (1− ε)(n− 1) lnn) ≤ exp(−nε)

What about the (1+1)-EA?

Page 105: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Artificial Fitness Levels

Observation Due to elitism, fitness is monotone increasing

D. Sudholt, Tutorial 2011

Idea Divide the search space |S| = 2n into m < 2n sets A1, . . . Am suchthat:

1 ∀i 6= j : Ai ∩Aj = ∅2⋃mi=0 Ai = 0, 1n

3 for all points a ∈ Ai and b ∈ Aj it happens that f(a) < f(b) if i < j.

requirement Am contains only optimal search points.

Page 106: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Artificial Fitness Levels

Observation Due to elitism, fitness is monotone increasing

D. Sudholt, Tutorial 2011

Idea Divide the search space |S| = 2n into m < 2n sets A1, . . . Am suchthat:

1 ∀i 6= j : Ai ∩Aj = ∅2⋃mi=0 Ai = 0, 1n

3 for all points a ∈ Ai and b ∈ Aj it happens that f(a) < f(b) if i < j.

requirement Am contains only optimal search points.

Page 107: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Artificial Fitness Levels [Droste et al., 2002]

Idea Divide the search space |S| = 2n into m < 2n sets A1, . . . Am suchthat:

1 ∀i 6= j : Ai ∩Aj = ∅2⋃mi=0 Ai = 0, 1n

3 for all points a ∈ Ai and b ∈ Aj it happens that f(a) < f(b) if i < j.

requirement Am contains only optimal search points.

Then:

si probability that point in Ai is mutated to a point in Aj with j > i.

Expected time: E(T ) ≤∑i

1si

Very simple, yet often powerful method for upper bounds

Page 108: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Artificial Fitness Levels [Droste et al., 2002]

Idea Divide the search space |S| = 2n into m < 2n sets A1, . . . Am suchthat:

1 ∀i 6= j : Ai ∩Aj = ∅2⋃mi=0 Ai = 0, 1n

3 for all points a ∈ Ai and b ∈ Aj it happens that f(a) < f(b) if i < j.

requirement Am contains only optimal search points.

Then:

si probability that point in Ai is mutated to a point in Aj with j > i.

Expected time: E(T ) ≤∑i

1si

Very simple, yet often powerful method for upper bounds

Page 109: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Artificial Fitness Levels

D. Sudholt, Tutorial 2011

Let:

p(Ai) be the probability that a random initial point belongs to level Ai

si be the probability to leave level Ai for Aj with j > i

Then:

E(T ) ≤∑

1≤i≤m−1

p(Ai) ·(

1

si+ · · ·+ 1

sm−1

)≤(

1

s1+ · · ·+ 1

sm−1

)=

m−1∑i=1

1

si

Inequality 1: Law of total probability(E(T ) =

∑i Pr(F ) · E(T |F )

)Inequality 2: Trivial!

Page 110: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

(1+1)-EA for OneMax

Theorem

The expected runtime of the (1+1)-EA for OneMaxis O(n lnn).

Proof

The current solution is in level Ai if it has i zeroes (hence n− i ones)

To reach a higher fitness level it is sufficient to flip a zero into a one andleave the other bits unchanged, which occurs with probability

si ≥ i ·1

n

(1− 1

n

)n−1

≥ i

en

Then(

Artificial Fitness Levels):

E(T ) ≤m−1∑i=1

s−1i ≤

n∑i=1

en

i≤ e · n

m−1∑i=1

1

i≤ e · n · (lnn+ 1) = O(n lnn)

Is the (1+1)-EA quicker than n lnn?

Page 111: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

(1+1)-EA for OneMax

Theorem

The expected runtime of the (1+1)-EA for OneMaxis O(n lnn).

Proof

The current solution is in level Ai if it has i zeroes (hence n− i ones)

To reach a higher fitness level it is sufficient to flip a zero into a one andleave the other bits unchanged, which occurs with probability

si ≥ i ·1

n

(1− 1

n

)n−1

≥ i

en

Then(

Artificial Fitness Levels):

E(T ) ≤m−1∑i=1

s−1i ≤

n∑i=1

en

i≤ e · n

m−1∑i=1

1

i≤ e · n · (lnn+ 1) = O(n lnn)

Is the (1+1)-EA quicker than n lnn?

Page 112: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

(1+1)-EA for OneMax

Theorem

The expected runtime of the (1+1)-EA for OneMaxis O(n lnn).

Proof

The current solution is in level Ai if it has i zeroes (hence n− i ones)

To reach a higher fitness level it is sufficient to flip a zero into a one andleave the other bits unchanged, which occurs with probability

si ≥ i ·1

n

(1− 1

n

)n−1

≥ i

en

Then(

Artificial Fitness Levels):

E(T ) ≤m−1∑i=1

s−1i ≤

n∑i=1

en

i≤ e · n

m−1∑i=1

1

i≤ e · n · (lnn+ 1) = O(n lnn)

Is the (1+1)-EA quicker than n lnn?

Page 113: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

(1+1)-EA for OneMax

Theorem

The expected runtime of the (1+1)-EA for OneMaxis O(n lnn).

Proof

The current solution is in level Ai if it has i zeroes (hence n− i ones)

To reach a higher fitness level it is sufficient to flip a zero into a one andleave the other bits unchanged, which occurs with probability

si ≥ i ·1

n

(1− 1

n

)n−1

≥ i

en

Then(

Artificial Fitness Levels):

E(T ) ≤m−1∑i=1

s−1i ≤

n∑i=1

en

i≤ e · n

m−1∑i=1

1

i≤ e · n · (lnn+ 1) = O(n lnn)

Is the (1+1)-EA quicker than n lnn?

Page 114: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

(1+1)-EA lower bound for OneMax

Theorem (Droste,Jansen,Wegener, 2002)

The expected runtime of the (1+1)-EA for OneMaxis Ω(n lnn).

Proof Idea

1 At most n/2 one-bits are created during initialisation with probability atleast 1/2 (By symmetry of the binomial distribution).

2 There is a constant probability that in cn lnn steps one of the n/2remaining zero-bits does not flip.

Page 115: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

(1+1)-EA lower bound for OneMax

Theorem (Droste,Jansen,Wegener, 2002)

The expected runtime of the (1+1)-EA for OneMaxis Ω(n lnn).

Proof Idea

1 At most n/2 one-bits are created during initialisation with probability atleast 1/2 (By symmetry of the binomial distribution).

2 There is a constant probability that in cn lnn steps one of the n/2remaining zero-bits does not flip.

Page 116: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

(1+1)-EA lower bound for OneMax

Theorem (Droste,Jansen,Wegener, 2002)

The expected runtime of the (1+1)-EA for OneMaxis Ω(n lnn).

Proof Idea

1 At most n/2 one-bits are created during initialisation with probability atleast 1/2 (By symmetry of the binomial distribution).

2 There is a constant probability that in cn lnn steps one of the n/2remaining zero-bits does not flip.

Page 117: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Lower bound for OneMax

Theorem (Droste,Jansen,Wegener, 2002)

The expected runtime of the (1+1)-EA for OneMaxis Ω(n logn).

Proof of 2.1− 1/n a given bit does not flip

(1− 1/n)t a given bit does not flip in t steps1− (1− 1/n)t it flips at least once in t steps

(1− (1− 1/n)t)n/2 n/2 bits flip at least once in t steps

1− [1− (1− 1/n)t]n/2 at least one of the n/2 bits does not flip in t steps

Set t = (n− 1) logn. Then:

1− [1− (1− 1/n)t]n/2 = 1− [1− (1− 1/n)(n−1) logn]n/2 ≥

≥ 1− [1− (1/e)logn]n/2 = 1− [1− 1/n]n/2 =

= 1− [1− 1/n]n·1/2 ≥ 1− (2e)−1/2 = c

Page 118: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Lower bound for OneMax

Theorem (Droste,Jansen,Wegener, 2002)

The expected runtime of the (1+1)-EA for OneMaxis Ω(n logn).

Proof of 2.1− 1/n a given bit does not flip(1− 1/n)t a given bit does not flip in t steps

1− (1− 1/n)t it flips at least once in t steps

(1− (1− 1/n)t)n/2 n/2 bits flip at least once in t steps

1− [1− (1− 1/n)t]n/2 at least one of the n/2 bits does not flip in t steps

Set t = (n− 1) logn. Then:

1− [1− (1− 1/n)t]n/2 = 1− [1− (1− 1/n)(n−1) logn]n/2 ≥

≥ 1− [1− (1/e)logn]n/2 = 1− [1− 1/n]n/2 =

= 1− [1− 1/n]n·1/2 ≥ 1− (2e)−1/2 = c

Page 119: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Lower bound for OneMax

Theorem (Droste,Jansen,Wegener, 2002)

The expected runtime of the (1+1)-EA for OneMaxis Ω(n logn).

Proof of 2.1− 1/n a given bit does not flip(1− 1/n)t a given bit does not flip in t steps1− (1− 1/n)t it flips at least once in t steps

(1− (1− 1/n)t)n/2 n/2 bits flip at least once in t steps

1− [1− (1− 1/n)t]n/2 at least one of the n/2 bits does not flip in t steps

Set t = (n− 1) logn. Then:

1− [1− (1− 1/n)t]n/2 = 1− [1− (1− 1/n)(n−1) logn]n/2 ≥

≥ 1− [1− (1/e)logn]n/2 = 1− [1− 1/n]n/2 =

= 1− [1− 1/n]n·1/2 ≥ 1− (2e)−1/2 = c

Page 120: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Lower bound for OneMax

Theorem (Droste,Jansen,Wegener, 2002)

The expected runtime of the (1+1)-EA for OneMaxis Ω(n logn).

Proof of 2.1− 1/n a given bit does not flip(1− 1/n)t a given bit does not flip in t steps1− (1− 1/n)t it flips at least once in t steps

(1− (1− 1/n)t)n/2 n/2 bits flip at least once in t steps

1− [1− (1− 1/n)t]n/2 at least one of the n/2 bits does not flip in t steps

Set t = (n− 1) logn. Then:

1− [1− (1− 1/n)t]n/2 = 1− [1− (1− 1/n)(n−1) logn]n/2 ≥

≥ 1− [1− (1/e)logn]n/2 = 1− [1− 1/n]n/2 =

= 1− [1− 1/n]n·1/2 ≥ 1− (2e)−1/2 = c

Page 121: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Lower bound for OneMax

Theorem (Droste,Jansen,Wegener, 2002)

The expected runtime of the (1+1)-EA for OneMaxis Ω(n logn).

Proof of 2.1− 1/n a given bit does not flip(1− 1/n)t a given bit does not flip in t steps1− (1− 1/n)t it flips at least once in t steps

(1− (1− 1/n)t)n/2 n/2 bits flip at least once in t steps

1− [1− (1− 1/n)t]n/2 at least one of the n/2 bits does not flip in t steps

Set t = (n− 1) logn. Then:

1− [1− (1− 1/n)t]n/2 = 1− [1− (1− 1/n)(n−1) logn]n/2 ≥

≥ 1− [1− (1/e)logn]n/2 = 1− [1− 1/n]n/2 =

= 1− [1− 1/n]n·1/2 ≥ 1− (2e)−1/2 = c

Page 122: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Lower bound for OneMax

Theorem (Droste,Jansen,Wegener, 2002)

The expected runtime of the (1+1)-EA for OneMaxis Ω(n logn).

Proof of 2.1− 1/n a given bit does not flip(1− 1/n)t a given bit does not flip in t steps1− (1− 1/n)t it flips at least once in t steps

(1− (1− 1/n)t)n/2 n/2 bits flip at least once in t steps

1− [1− (1− 1/n)t]n/2 at least one of the n/2 bits does not flip in t steps

Set t = (n− 1) logn. Then:

1− [1− (1− 1/n)t]n/2 = 1− [1− (1− 1/n)(n−1) logn]n/2 ≥

≥ 1− [1− (1/e)logn]n/2 = 1− [1− 1/n]n/2 =

= 1− [1− 1/n]n·1/2 ≥ 1− (2e)−1/2 = c

Page 123: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Lower bound for OneMax(2)

Theorem (Droste, Jansen, Wegener, 2002)

The expected runtime of the (1+1)-EA for OneMaxis Ω(n logn).

Proof

1 At most n/2 one-bits are created during initialisation with probability atleast 1/2 (By symmetry of the binomial distribution).

2 There is a constant probability that in cn logn steps one of the n/2remaining zero-bits does not flip.

The Expected runtime is:

E[T ] =

∞∑t=1

t · p(t) ≥ [(n− 1) logn] · p[t = (n− 1) logn] ≥

≥ [(n− 1) logn] · [(1/2) · (1− (2e)−1/2) = Ω(n logn)

First inequality: law of total probability

The upper bound given by artificial fitness levels is indeed tight!

Page 124: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Lower bound for OneMax(2)

Theorem (Droste, Jansen, Wegener, 2002)

The expected runtime of the (1+1)-EA for OneMaxis Ω(n logn).

Proof

1 At most n/2 one-bits are created during initialisation with probability atleast 1/2 (By symmetry of the binomial distribution).

2 There is a constant probability that in cn logn steps one of the n/2remaining zero-bits does not flip.

The Expected runtime is:

E[T ] =∞∑t=1

t · p(t) ≥ [(n− 1) logn] · p[t = (n− 1) logn] ≥

≥ [(n− 1) logn] · [(1/2) · (1− (2e)−1/2) = Ω(n logn)

First inequality: law of total probability

The upper bound given by artificial fitness levels is indeed tight!

Page 125: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Artificial Fitness Levels Exercises:

(LeadingOnes(x) =

∑ni=1

∏ij=1 x[j]

)

Theorem

The expected runtime of RLS for LeadingOnes is O(n2).

Proof

Let partition Ai contain search points with exactly i leading ones

To leave level Ai it suffices to flip the zero at position i+ 1

si = 1n

and s−1i = n

E(T ) ≤∑n−1i=1 s

−1i =

∑ni=1 n = O(n2)

Theorem

The expected runtime of the (1+1)-EA for LeadingOnes is O(n2).

Proof Left as Exercise.

Page 126: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Artificial Fitness Levels Exercises:

(LeadingOnes(x) =

∑ni=1

∏ij=1 x[j]

)

Theorem

The expected runtime of RLS for LeadingOnes is O(n2).

Proof

Let partition Ai contain search points with exactly i leading ones

To leave level Ai it suffices to flip the zero at position i+ 1

si = 1n

and s−1i = n

E(T ) ≤∑n−1i=1 s

−1i =

∑ni=1 n = O(n2)

Theorem

The expected runtime of the (1+1)-EA for LeadingOnes is O(n2).

Proof Left as Exercise.

Page 127: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Artificial Fitness Levels Exercises:

(LeadingOnes(x) =

∑ni=1

∏ij=1 x[j]

)

Theorem

The expected runtime of RLS for LeadingOnes is O(n2).

Proof

Let partition Ai contain search points with exactly i leading ones

To leave level Ai it suffices to flip the zero at position i+ 1

si = 1n

and s−1i = n

E(T ) ≤∑n−1i=1 s

−1i =

∑ni=1 n = O(n2)

Theorem

The expected runtime of the (1+1)-EA for LeadingOnes is O(n2).

Proof Left as Exercise.

Page 128: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Artificial Fitness Levels Exercises:

(LeadingOnes(x) =

∑ni=1

∏ij=1 x[j]

)

Theorem

The expected runtime of RLS for LeadingOnes is O(n2).

Proof

Let partition Ai contain search points with exactly i leading ones

To leave level Ai it suffices to flip the zero at position i+ 1

si = 1n

and s−1i = n

E(T ) ≤∑n−1i=1 s

−1i =

∑ni=1 n = O(n2)

Theorem

The expected runtime of the (1+1)-EA for LeadingOnes is O(n2).

Proof Left as Exercise.

Page 129: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Fitness Levels Advanced Exercises (Populations)

Theorem

The expected runtime of (1+λ)-EA for LeadingOnes is O(λn+ n2)[Jansen et al., 2005].

Proof

Let partition Ai contain search points with exactly i leading ones

To leave level Ai it suffices to flip the zero at position i+ 1

si = 1−(

1− 1en

)λ≥ 1− e−λ/(en)

1 si ≥ 1− 1e

Case 1: λ ≥ en2 si ≥ λ

2enCase 2: λ < en

E(T ) ≤ λ ·∑n−1i=1 s

−1i ≤ λ

((∑ni=1

1c

)+

(∑ni=1

2enλ

))=

O

(λ ·(n+ n2

λ

))= O(λ · n+ n2)

Page 130: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Fitness Levels Advanced Exercises (Populations)

Theorem

The expected runtime of (1+λ)-EA for LeadingOnes is O(λn+ n2)[Jansen et al., 2005].

Proof

Let partition Ai contain search points with exactly i leading ones

To leave level Ai it suffices to flip the zero at position i+ 1

si = 1−(

1− 1en

)λ≥ 1− e−λ/(en)

1 si ≥ 1− 1e

Case 1: λ ≥ en2 si ≥ λ

2enCase 2: λ < en

E(T ) ≤ λ ·∑n−1i=1 s

−1i ≤ λ

((∑ni=1

1c

)+

(∑ni=1

2enλ

))=

O

(λ ·(n+ n2

λ

))= O(λ · n+ n2)

Page 131: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Fitness Levels Advanced Exercises (Populations)

Theorem

The expected runtime of the (µ+1)-EA for LeadingOnes is O(µ · n2).

Proof Left as Exercise.

Theorem

The expected runtime of the (µ+1)-EA for OneMaxis O(µ · n logn).

Proof Left as Exercise.

Page 132: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Fitness Levels Advanced Exercises (Populations)

Theorem

The expected runtime of the (µ+1)-EA for LeadingOnes is O(µ · n2).

Proof Left as Exercise.

Theorem

The expected runtime of the (µ+1)-EA for OneMaxis O(µ · n logn).

Proof Left as Exercise.

Page 133: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for upper bounds

Fitness Levels Advanced Exercises (Populations)

Theorem

The expected runtime of the (µ+1)-EA for LeadingOnes is O(µ · n2).

Proof Left as Exercise.

Theorem

The expected runtime of the (µ+1)-EA for OneMaxis O(µ · n logn).

Proof Left as Exercise.

Page 134: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for parent populations

Artificial Fitness Levels for Populations

D. Sudholt, Tutorial 2011

Let:

To be the expected time for a fraction χ(i) of the population to be in levelAi

si be the probability to leave level Ai for Aj with j > i given χ(i) in levelAi

Then:

E(T ) ≤m−1∑i=1

(1

si+ To

)

Page 135: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for parent populations

Applications to (µ+1)-EA

Theorem

The expected runtime of (µ+1)-EA for LeadingOnes is O(µn logn+ n2)[Witt, 2006].

ProofLet partition Ai contain search points with exactly i leading onesTo leave level Ai it suffices to flip the zero at position i+ 1 of the bestindividualWe set χ(i) = n/ lnnGiven j copies of the best individual another replica is created with

probability jµ

(1− 1

n

)n≥ j

2eµ

To ≤∑n/ lnnj=1

2eµj≤ 2eµ lnn

1 si ≥ n/ lnnµ· 1en

= 1eµ lnn

Case 1: µ > nlnn

2 si ≥ n/ lnnµ· 1en≥ 1

enCase 2: µ ≤ n

lnn

E(T ) ≤∑n−1i=1 (To + s−1

i ) ≤∑ni=1

(2eµ lnn+

(en+ eµ lnn

))=

n ·(

2eµ lnn+(en+ eµ lnn

))= O(nµ lnn+ n2)

Page 136: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for parent populations

Applications to (µ+1)-EA

Theorem

The expected runtime of (µ+1)-EA for LeadingOnes is O(µn logn+ n2)[Witt, 2006].

ProofLet partition Ai contain search points with exactly i leading ones

To leave level Ai it suffices to flip the zero at position i+ 1 of the bestindividualWe set χ(i) = n/ lnnGiven j copies of the best individual another replica is created with

probability jµ

(1− 1

n

)n≥ j

2eµ

To ≤∑n/ lnnj=1

2eµj≤ 2eµ lnn

1 si ≥ n/ lnnµ· 1en

= 1eµ lnn

Case 1: µ > nlnn

2 si ≥ n/ lnnµ· 1en≥ 1

enCase 2: µ ≤ n

lnn

E(T ) ≤∑n−1i=1 (To + s−1

i ) ≤∑ni=1

(2eµ lnn+

(en+ eµ lnn

))=

n ·(

2eµ lnn+(en+ eµ lnn

))= O(nµ lnn+ n2)

Page 137: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for parent populations

Applications to (µ+1)-EA

Theorem

The expected runtime of (µ+1)-EA for LeadingOnes is O(µn logn+ n2)[Witt, 2006].

ProofLet partition Ai contain search points with exactly i leading onesTo leave level Ai it suffices to flip the zero at position i+ 1 of the bestindividual

We set χ(i) = n/ lnnGiven j copies of the best individual another replica is created with

probability jµ

(1− 1

n

)n≥ j

2eµ

To ≤∑n/ lnnj=1

2eµj≤ 2eµ lnn

1 si ≥ n/ lnnµ· 1en

= 1eµ lnn

Case 1: µ > nlnn

2 si ≥ n/ lnnµ· 1en≥ 1

enCase 2: µ ≤ n

lnn

E(T ) ≤∑n−1i=1 (To + s−1

i ) ≤∑ni=1

(2eµ lnn+

(en+ eµ lnn

))=

n ·(

2eµ lnn+(en+ eµ lnn

))= O(nµ lnn+ n2)

Page 138: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for parent populations

Applications to (µ+1)-EA

Theorem

The expected runtime of (µ+1)-EA for LeadingOnes is O(µn logn+ n2)[Witt, 2006].

ProofLet partition Ai contain search points with exactly i leading onesTo leave level Ai it suffices to flip the zero at position i+ 1 of the bestindividualWe set χ(i) = n/ lnn

Given j copies of the best individual another replica is created with

probability jµ

(1− 1

n

)n≥ j

2eµ

To ≤∑n/ lnnj=1

2eµj≤ 2eµ lnn

1 si ≥ n/ lnnµ· 1en

= 1eµ lnn

Case 1: µ > nlnn

2 si ≥ n/ lnnµ· 1en≥ 1

enCase 2: µ ≤ n

lnn

E(T ) ≤∑n−1i=1 (To + s−1

i ) ≤∑ni=1

(2eµ lnn+

(en+ eµ lnn

))=

n ·(

2eµ lnn+(en+ eµ lnn

))= O(nµ lnn+ n2)

Page 139: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for parent populations

Applications to (µ+1)-EA

Theorem

The expected runtime of (µ+1)-EA for LeadingOnes is O(µn logn+ n2)[Witt, 2006].

ProofLet partition Ai contain search points with exactly i leading onesTo leave level Ai it suffices to flip the zero at position i+ 1 of the bestindividualWe set χ(i) = n/ lnnGiven j copies of the best individual another replica is created with

probability jµ

(1− 1

n

)n≥ j

2eµ

To ≤∑n/ lnnj=1

2eµj≤ 2eµ lnn

1 si ≥ n/ lnnµ· 1en

= 1eµ lnn

Case 1: µ > nlnn

2 si ≥ n/ lnnµ· 1en≥ 1

enCase 2: µ ≤ n

lnn

E(T ) ≤∑n−1i=1 (To + s−1

i ) ≤∑ni=1

(2eµ lnn+

(en+ eµ lnn

))=

n ·(

2eµ lnn+(en+ eµ lnn

))= O(nµ lnn+ n2)

Page 140: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for parent populations

Applications to (µ+1)-EA

Theorem

The expected runtime of (µ+1)-EA for LeadingOnes is O(µn logn+ n2)[Witt, 2006].

ProofLet partition Ai contain search points with exactly i leading onesTo leave level Ai it suffices to flip the zero at position i+ 1 of the bestindividualWe set χ(i) = n/ lnnGiven j copies of the best individual another replica is created with

probability jµ

(1− 1

n

)n≥ j

2eµ

To ≤∑n/ lnnj=1

2eµj≤ 2eµ lnn

1 si ≥ n/ lnnµ· 1en

= 1eµ lnn

Case 1: µ > nlnn

2 si ≥ n/ lnnµ· 1en≥ 1

enCase 2: µ ≤ n

lnn

E(T ) ≤∑n−1i=1 (To + s−1

i ) ≤∑ni=1

(2eµ lnn+

(en+ eµ lnn

))=

n ·(

2eµ lnn+(en+ eµ lnn

))= O(nµ lnn+ n2)

Page 141: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for parent populations

Applications to (µ+1)-EA

Theorem

The expected runtime of (µ+1)-EA for LeadingOnes is O(µn logn+ n2)[Witt, 2006].

ProofLet partition Ai contain search points with exactly i leading onesTo leave level Ai it suffices to flip the zero at position i+ 1 of the bestindividualWe set χ(i) = n/ lnnGiven j copies of the best individual another replica is created with

probability jµ

(1− 1

n

)n≥ j

2eµ

To ≤∑n/ lnnj=1

2eµj≤ 2eµ lnn

1 si ≥ n/ lnnµ· 1en

= 1eµ lnn

Case 1: µ > nlnn

2 si ≥ n/ lnnµ· 1en≥ 1

enCase 2: µ ≤ n

lnn

E(T ) ≤∑n−1i=1 (To + s−1

i ) ≤∑ni=1

(2eµ lnn+

(en+ eµ lnn

))=

n ·(

2eµ lnn+(en+ eµ lnn

))= O(nµ lnn+ n2)

Page 142: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for parent populations

Applications to (µ+1)-EA

Theorem

The expected runtime of (µ+1)-EA for LeadingOnes is O(µn logn+ n2)[Witt, 2006].

ProofLet partition Ai contain search points with exactly i leading onesTo leave level Ai it suffices to flip the zero at position i+ 1 of the bestindividualWe set χ(i) = n/ lnnGiven j copies of the best individual another replica is created with

probability jµ

(1− 1

n

)n≥ j

2eµ

To ≤∑n/ lnnj=1

2eµj≤ 2eµ lnn

1 si ≥ n/ lnnµ· 1en

= 1eµ lnn

Case 1: µ > nlnn

2 si ≥ n/ lnnµ· 1en≥ 1

enCase 2: µ ≤ n

lnn

E(T ) ≤∑n−1i=1 (To + s−1

i ) ≤∑ni=1

(2eµ lnn+

(en+ eµ lnn

))=

n ·(

2eµ lnn+(en+ eµ lnn

))= O(nµ lnn+ n2)

Page 143: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for parent populations

Populations Fitness Levels: Exercise

Theorem

The expected runtime of the (µ+1)-EA for OneMaxis O(µn+ n logn).

Proof Left as Exercise.

Page 144: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL method for parent populations

Populations Fitness Levels: Exercise

Theorem

The expected runtime of the (µ+1)-EA for OneMaxis O(µn+ n logn).

Proof Left as Exercise.

Page 145: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL for lower bounds

Advanced: Fitness Levels for Lower Bounds [Sudholt, 2010]

ui := probability to leave level Ai;γi,j := probability of jumping from Ai to Aj .

By showing that with high probability the (1+1)-EA does not skip too manylevels, then it can be proven that the runtime for OneMax is at leasten lnn+O(n).

Page 146: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL for lower bounds

Advanced: Fitness Levels for Lower Bounds [Sudholt, 2010]

ui := probability to leave level Ai;γi,j := probability of jumping from Ai to Aj .

By showing that with high probability the (1+1)-EA does not skip too manylevels, then it can be proven that the runtime for OneMax is at leasten lnn+O(n).

Page 147: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL for lower bounds

Advanced: Fitness Levels for Populations and Crossover[Corus and Oliveto, 2017]

Theorem

The expected runtime of the (µ+1)-GA with µ ≥ 3 and mutation rate c/n forany constant c on OneMax is:

E[T ] ≤ 3ecn logn

c(3 + c)+O(nµ logµ).

For µ = o(logn/ log logn), the bound reduces to:

E[T ] ≤ 3

c(3 + c)ecn logn

(1 + o(1)

).

1 Mutation rate 1/n: E[T ] ≤ 3/4 · en logn(1 + o(1)) (GA) versusE[T ] ≥ e · n logn(1 + o(1)) (no crossover);

2 Higher mutation rates may be beneficial:E[T ] ≤ 0.72 · e · n logn(1 + o(1)) ≈ 1.3/n;

3 Populations may be beneficial: for µ = 2 the bound isE[T ] ≤ 4/5 · e · n logn(1 + o(1)).

Page 148: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL for lower bounds

Advanced: Fitness Levels for Populations and Crossover[Corus and Oliveto, 2017]

Theorem

The expected runtime of the (µ+1)-GA with µ ≥ 3 and mutation rate c/n forany constant c on OneMax is:

E[T ] ≤ 3ecn logn

c(3 + c)+O(nµ logµ).

For µ = o(logn/ log logn), the bound reduces to:

E[T ] ≤ 3

c(3 + c)ecn logn

(1 + o(1)

).

1 Mutation rate 1/n: E[T ] ≤ 3/4 · en logn(1 + o(1)) (GA) versusE[T ] ≥ e · n logn(1 + o(1)) (no crossover);

2 Higher mutation rates may be beneficial:E[T ] ≤ 0.72 · e · n logn(1 + o(1)) ≈ 1.3/n;

3 Populations may be beneficial: for µ = 2 the bound isE[T ] ≤ 4/5 · e · n logn(1 + o(1)).

Page 149: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL for lower bounds

Advanced: Fitness Levels for Populations and Crossover[Corus and Oliveto, 2017]

Theorem

The expected runtime of the (µ+1)-GA with µ ≥ 3 and mutation rate c/n forany constant c on OneMax is:

E[T ] ≤ 3ecn logn

c(3 + c)+O(nµ logµ).

For µ = o(logn/ log logn), the bound reduces to:

E[T ] ≤ 3

c(3 + c)ecn logn

(1 + o(1)

).

1 Mutation rate 1/n: E[T ] ≤ 3/4 · en logn(1 + o(1)) (GA) versusE[T ] ≥ e · n logn(1 + o(1)) (no crossover);

2 Higher mutation rates may be beneficial:E[T ] ≤ 0.72 · e · n logn(1 + o(1)) ≈ 1.3/n;

3 Populations may be beneficial: for µ = 2 the bound isE[T ] ≤ 4/5 · e · n logn(1 + o(1)).

Page 150: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL for lower bounds

Advanced: Fitness Levels for Populations and Crossover[Corus and Oliveto, 2017]

Theorem

The expected runtime of the (µ+1)-GA with µ ≥ 3 and mutation rate c/n forany constant c on OneMax is:

E[T ] ≤ 3ecn logn

c(3 + c)+O(nµ logµ).

For µ = o(logn/ log logn), the bound reduces to:

E[T ] ≤ 3

c(3 + c)ecn logn

(1 + o(1)

).

1 Mutation rate 1/n: E[T ] ≤ 3/4 · en logn(1 + o(1)) (GA) versusE[T ] ≥ e · n logn(1 + o(1)) (no crossover);

2 Higher mutation rates may be beneficial:E[T ] ≤ 0.72 · e · n logn(1 + o(1)) ≈ 1.3/n;

3 Populations may be beneficial: for µ = 2 the bound isE[T ] ≤ 4/5 · e · n logn(1 + o(1)).

Page 151: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL for lower bounds

Proof Idea [Corus and Oliveto, 2017]

We divide the search space in canonical fitness levelsLi = x ∈ 0, 1n|OneMax(x) = i;

Each level i is represented by a Markov Chain (and all individuals are atleast in Li);

The runtime is upper bounded by the time it takes to discover the nextlevel E[Li]+ the time it takes for the entire population to take over thelevel (E[Ttakeover] = O(µ logµ)).

E[T ] ≤n−1∑i=0

(E[Li] + E[Ttakeover]

)

Page 152: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL for lower bounds

Proof Idea [Corus and Oliveto, 2017]

We divide the search space in canonical fitness levelsLi = x ∈ 0, 1n|OneMax(x) = i;Each level i is represented by a Markov Chain (and all individuals are atleast in Li);

The runtime is upper bounded by the time it takes to discover the nextlevel E[Li]+ the time it takes for the entire population to take over thelevel (E[Ttakeover] = O(µ logµ)).

E[T ] ≤n−1∑i=0

(E[Li] + E[Ttakeover]

)

Page 153: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL for lower bounds

Proof Idea [Corus and Oliveto, 2017]

We divide the search space in canonical fitness levelsLi = x ∈ 0, 1n|OneMax(x) = i;Each level i is represented by a Markov Chain (and all individuals are atleast in Li);

The runtime is upper bounded by the time it takes to discover the nextlevel E[Li]+ the time it takes for the entire population to take over thelevel (E[Ttakeover] = O(µ logµ)).

E[T ] ≤n−1∑i=0

(E[Li] + E[Ttakeover]

)

Page 154: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL for lower bounds

Proof Idea [Corus and Oliveto, 2017]

We divide the search space in canonical fitness levelsLi = x ∈ 0, 1n|OneMax(x) = i;Each level i is represented by a Markov Chain (and all individuals are atleast in Li);

The runtime is upper bounded by the time it takes to discover the nextlevel E[Li]+ the time it takes for the entire population to take over thelevel (E[Ttakeover] = O(µ logµ)).

E[T ] ≤n−1∑i=0

(E[Li] + E[Ttakeover]

)

Page 155: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL for lower bounds

Proof Idea [Corus and Oliveto, 2017]

We divide the search space in canonical fitness levelsLi = x ∈ 0, 1n|OneMax(x) = i;Each level i is represented by a Markov Chain (and all individuals are atleast in Li);

The runtime is upper bounded by the time it takes to discover the nextlevel E[Li]+ the time it takes for the entire population to take over thelevel (E[Ttakeover] = O(µ logµ)).

E[T ] ≤n−1∑i=0

(E[Li] + E[Ttakeover]

)

Page 156: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL for lower bounds

Advanced: Fitness Levels for non-Elitist Populations [Lehre, 2011]

See the Oliveto, Lehre Tutorial at GECCO 2017 in Berlin!

Page 157: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

AFL for lower bounds

Artificial Fitness Levels: Conclusions

It’s a powerful general method to obtain (often) tight upper bounds onthe runtime of simple EAs;

For offspring populations tight bounds can often be achieved with thegeneral method;

For parent populations takeover times have to be introduced;

For detailed analysis of crossover Markov chains have to be introduced;

Recent methods have been presented to deal with non-elitism and forlower bounds.

Page 158: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

What is Drift1 Analysis?

Prediction of the long term behaviour of a process Xhitting time, stability, occupancy time etc.

from properties of ∆.

1NB! (Stochastic) drift is a different concept than genetic drift in population genetics.

Page 159: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

What is Drift1 Analysis?

Prediction of the long term behaviour of a process Xhitting time, stability, occupancy time etc.

from properties of ∆.

1NB! (Stochastic) drift is a different concept than genetic drift in population genetics.

Page 160: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Drift Analysis: Example 1

Friday night dinner at the restaurant.Peter walks back from the restaurant to the hotel.

The restaurant is n meters away from the hotel;

Peter moves towards the hotel of 1 meter in each step

QuestionHow many steps does Peter need to reach his hotel?

n steps

Page 161: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Drift Analysis: Example 1

Friday night dinner at the restaurant.Peter walks back from the restaurant to the hotel.

The restaurant is n meters away from the hotel;

Peter moves towards the hotel of 1 meter in each step

QuestionHow many steps does Peter need to reach his hotel?n steps

Page 162: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Drift Analysis: Formalisation

Define a distance function d(x) to measure the distance from the hotel;

d(x) = x, x ∈ 0, . . . , n

(In our case the distance is simply the number of metres from the hotel).

Estimate the expected “speed” (drift), the expected decrease in distancein one step from the goal;

d(Xt)− d(Xt+1) =

0, if Xt = 0,

1, if Xt ∈ 1, . . . , n

TimeThen the expected time to reach the hotel (goal) is:

E(T ) =maximum distance

drift=n

1= n

Page 163: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Drift Analysis: Example 2

Friday night dinner at the restaurant.Peter walks back from the restaurant to the hotel but had a few drinks.

The restaurant is n meters away from the hotel;

Peter moves towards the hotel of 1 meter in each step with probability 0.6.

Peter moves away from the hotel of 1 meter in each step with probability0.4.

QuestionHow many steps does Peter need to reach his hotel?

5n stepsLet us calculate this through drift analysis.

Page 164: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Drift Analysis: Example 2

Friday night dinner at the restaurant.Peter walks back from the restaurant to the hotel but had a few drinks.

The restaurant is n meters away from the hotel;

Peter moves towards the hotel of 1 meter in each step with probability 0.6.

Peter moves away from the hotel of 1 meter in each step with probability0.4.

QuestionHow many steps does Peter need to reach his hotel?5n stepsLet us calculate this through drift analysis.

Page 165: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Drift Analysis (2): Formalisation

Define the same distance function d(x) as before to measure the distancefrom the hotel;

d(x) = x, x ∈ 0, . . . , n

(simply the number of metres from the hotel).

Estimate the expected “speed” (drift), the expected decrease in distancein one step from the goal;

d(Xt)− d(Xt+1) =

0, if Xt = 0,

1, if Xt ∈ 1, . . . , nwith probability 0.6

−1, if Xt ∈ 1, . . . , nwith probability 0.4

The expected dicrease in distance (drift) is:

E[d(Xt)− d(Xt+1)] = 0.6 · 1 + 0.4 · (−1) = 0.6− 0.4 = 0.2

TimeThen the expected time to reach the hotel (goal) is:

E(T ) =maximum distance

drift=

n

0.2= 5n

Page 166: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Additive Drift Theorem

Theorem (Additive Drift Theorem for Upper Bounds [He and Yao, 2001])

Let Xtt≥0 be a Markov process over a set of states S, and d : S → R+0 a

function that assigns a non-negative real number to every state. Let the timeto reach the optimum be T := mint ≥ 0 : d(Xt) = 0. If there exists δ > 0such that at any time step t ≥ 0 and at any state Xt > 0 the followingcondition holds:

E(∆(t)|d(Xt) > 0) = E(d(Xt)− d(Xt+1) | d(Xt) > 0) ≥ δ (1)

then

E(T | d(X0) > 0) ≤ d(X0)

δ(2)

and

E(T ) ≤ E(d(X0))

δ. (3)

Page 167: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Drift Analysis for Leading Ones

Theorem

The expected time for the (1+1)-EA to optimise LeadingOnes is O(n2)

Proof

1 Let d(Xt) = i where i is the number of missing leading ones;

2 The negative drift is 0 since if a leading one is removed from the currentsolution the new point will not be accepted;

3 A positive drift (i.e. of length 1) is achieved as long as the first 0 is flippedand the leading ones are remained unchanged:

E(∆+(t)) =

n−i∑k=1

k · (p(∆+(t)) = k) ≥ 1 · 1/n ·(1− 1/n)n−1 ≥ 1/(en)

4 Hence, E[∆(t)|d(Xt)] ≥ 1/(en) = δ

5 The expected runtime is (i.e. Eq. (6)):

E(T | d(X0) > 0) ≤ d(X0)

δ≤ n

1/(en)= e · n2 = O(n2)

Page 168: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Drift Analysis for Leading Ones

Theorem

The expected time for the (1+1)-EA to optimise LeadingOnes is O(n2)

Proof

1 Let d(Xt) = i where i is the number of missing leading ones;

2 The negative drift is 0 since if a leading one is removed from the currentsolution the new point will not be accepted;

3 A positive drift (i.e. of length 1) is achieved as long as the first 0 is flippedand the leading ones are remained unchanged:

E(∆+(t)) =

n−i∑k=1

k · (p(∆+(t)) = k) ≥ 1 · 1/n ·(1− 1/n)n−1 ≥ 1/(en)

4 Hence, E[∆(t)|d(Xt)] ≥ 1/(en) = δ

5 The expected runtime is (i.e. Eq. (6)):

E(T | d(X0) > 0) ≤ d(X0)

δ≤ n

1/(en)= e · n2 = O(n2)

Page 169: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Drift Analysis for Leading Ones

Theorem

The expected time for the (1+1)-EA to optimise LeadingOnes is O(n2)

Proof

1 Let d(Xt) = i where i is the number of missing leading ones;

2 The negative drift is 0 since if a leading one is removed from the currentsolution the new point will not be accepted;

3 A positive drift (i.e. of length 1) is achieved as long as the first 0 is flippedand the leading ones are remained unchanged:

E(∆+(t)) =

n−i∑k=1

k · (p(∆+(t)) = k) ≥ 1 · 1/n ·(1− 1/n)n−1 ≥ 1/(en)

4 Hence, E[∆(t)|d(Xt)] ≥ 1/(en) = δ

5 The expected runtime is (i.e. Eq. (6)):

E(T | d(X0) > 0) ≤ d(X0)

δ≤ n

1/(en)= e · n2 = O(n2)

Page 170: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Drift Analysis for Leading Ones

Theorem

The expected time for the (1+1)-EA to optimise LeadingOnes is O(n2)

Proof

1 Let d(Xt) = i where i is the number of missing leading ones;

2 The negative drift is 0 since if a leading one is removed from the currentsolution the new point will not be accepted;

3 A positive drift (i.e. of length 1) is achieved as long as the first 0 is flippedand the leading ones are remained unchanged:

E(∆+(t)) =

n−i∑k=1

k · (p(∆+(t)) = k) ≥ 1 · 1/n ·(1− 1/n)n−1 ≥ 1/(en)

4 Hence, E[∆(t)|d(Xt)] ≥ 1/(en) = δ

5 The expected runtime is (i.e. Eq. (6)):

E(T | d(X0) > 0) ≤ d(X0)

δ≤ n

1/(en)= e · n2 = O(n2)

Page 171: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Drift Analysis for Leading Ones

Theorem

The expected time for the (1+1)-EA to optimise LeadingOnes is O(n2)

Proof

1 Let d(Xt) = i where i is the number of missing leading ones;

2 The negative drift is 0 since if a leading one is removed from the currentsolution the new point will not be accepted;

3 A positive drift (i.e. of length 1) is achieved as long as the first 0 is flippedand the leading ones are remained unchanged:

E(∆+(t)) =

n−i∑k=1

k · (p(∆+(t)) = k) ≥ 1 · 1/n ·(1− 1/n)n−1 ≥ 1/(en)

4 Hence, E[∆(t)|d(Xt)] ≥ 1/(en) = δ

5 The expected runtime is (i.e. Eq. (6)):

E(T | d(X0) > 0) ≤ d(X0)

δ≤ n

1/(en)= e · n2 = O(n2)

Page 172: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Drift Analysis for Leading Ones

Theorem

The expected time for the (1+1)-EA to optimise LeadingOnes is O(n2)

Proof

1 Let d(Xt) = i where i is the number of missing leading ones;

2 The negative drift is 0 since if a leading one is removed from the currentsolution the new point will not be accepted;

3 A positive drift (i.e. of length 1) is achieved as long as the first 0 is flippedand the leading ones are remained unchanged:

E(∆+(t)) =

n−i∑k=1

k · (p(∆+(t)) = k) ≥ 1 · 1/n ·(1− 1/n)n−1 ≥ 1/(en)

4 Hence, E[∆(t)|d(Xt)] ≥ 1/(en) = δ

5 The expected runtime is (i.e. Eq. (6)):

E(T | d(X0) > 0) ≤ d(X0)

δ≤ n

1/(en)= e · n2 = O(n2)

Page 173: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Exercises

Theorem

The expected time for RLS to optimise LeadingOnes is O(n2)

Proof

Left as exercise.

Theorem

Let λ ≥ en. Then the expected time for the (1+λ)-EA to optimiseLeadingOnes is O(λn)

Proof Left as exercise.

Theorem

Let λ < en. Then the expected time for the (1+λ)-EA to optimiseLeadingOnes is O(n2)

Proof Left as exercise.

Page 174: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Exercises

Theorem

The expected time for RLS to optimise LeadingOnes is O(n2)

Proof Left as exercise.

Theorem

Let λ ≥ en. Then the expected time for the (1+λ)-EA to optimiseLeadingOnes is O(λn)

Proof

Left as exercise.

Theorem

Let λ < en. Then the expected time for the (1+λ)-EA to optimiseLeadingOnes is O(n2)

Proof Left as exercise.

Page 175: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Exercises

Theorem

The expected time for RLS to optimise LeadingOnes is O(n2)

Proof Left as exercise.

Theorem

Let λ ≥ en. Then the expected time for the (1+λ)-EA to optimiseLeadingOnes is O(λn)

Proof Left as exercise.

Theorem

Let λ < en. Then the expected time for the (1+λ)-EA to optimiseLeadingOnes is O(n2)

Proof

Left as exercise.

Page 176: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Exercises

Theorem

The expected time for RLS to optimise LeadingOnes is O(n2)

Proof Left as exercise.

Theorem

Let λ ≥ en. Then the expected time for the (1+λ)-EA to optimiseLeadingOnes is O(λn)

Proof Left as exercise.

Theorem

Let λ < en. Then the expected time for the (1+λ)-EA to optimiseLeadingOnes is O(n2)

Proof Left as exercise.

Page 177: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

(1,λ)-EA Analysis for LeadingOnes

Theorem

Let λ = n. Then the expected time for the (1,λ)-EA to optimiseLeadingOnes is O(n2)

Proof

Distance: let d(x) = n− i where i is the number of leading ones;

Drift:

E[d(Xt)− d(Xt+1)|d(Xt) = n− i]

≥ 1 ·(

1−(

1− 1

en

)λ)− n ·

(1−

(1− 1

n

)n)λ= c1 − n · cn2 = Ω(1)

Hence,

E(generations) ≤ max distance

drift=

n

Ω(1)= O(n)

and,E(T ) ≤ n · E(generations) = O(n2)

Page 178: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

(1,λ)-EA Analysis for LeadingOnes

Theorem

Let λ = n. Then the expected time for the (1,λ)-EA to optimiseLeadingOnes is O(n2)

Proof

Distance: let d(x) = n− i where i is the number of leading ones;

Drift:

E[d(Xt)− d(Xt+1)|d(Xt) = n− i]

≥ 1 ·(

1−(

1− 1

en

)λ)− n ·

(1−

(1− 1

n

)n)λ= c1 − n · cn2 = Ω(1)

Hence,

E(generations) ≤ max distance

drift=

n

Ω(1)= O(n)

and,E(T ) ≤ n · E(generations) = O(n2)

Page 179: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Additive Drift Theorem

Theorem (Additive Drift Theorem for Lower Bounds [He and Yao, 2004])

Let Xtt≥0 be a Markov process over a set of states S, and d : S → R+0 a

function that assigns a non-negative real number to every state. Let the timeto reach the optimum be T := mint ≥ 0 : d(Xt) = 0. If there exists δ > 0such that at any time step t ≥ 0 and at any state Xt > 0 the followingcondition holds:

E(∆(t)|d(Xt) > 0) = E(d(Xt)− d(Xt+1) | d(Xt) > 0) ≤ δ (4)

then

E(T | X0 > 0) ≥ d(X0)

δ(5)

and

E(T ) ≥ E(d(X0))

δ. (6)

Page 180: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Theorem

The expected time for the (1+1)-EA to optimise LeadingOnes is Ω(n2).

Sources of progress

1 Flipping the leftmost zero-bit;

2 Bits to right of the leftmost zero-bit that are one-bits (free riders).

Proof

1 Let the current solution have n− i leading ones (i.e. 1n−i0∗).

2 We define the distance function as the number of missing leading ones, i.e.d(X) = i.

3 The n− i+ 1 bit is a zero;

4 let E[Y ] be the expected number of one-bits after the first zero (i.e. thefree riders).

5 Such i− 1 bits are uniformely distributed at initialisation and still are!

Page 181: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Theorem

The expected time for the (1+1)-EA to optimise LeadingOnes is Ω(n2).

Sources of progress

1 Flipping the leftmost zero-bit;

2 Bits to right of the leftmost zero-bit that are one-bits (free riders).

Proof

1 Let the current solution have n− i leading ones (i.e. 1n−i0∗).

2 We define the distance function as the number of missing leading ones, i.e.d(X) = i.

3 The n− i+ 1 bit is a zero;

4 let E[Y ] be the expected number of one-bits after the first zero (i.e. thefree riders).

5 Such i− 1 bits are uniformely distributed at initialisation and still are!

Page 182: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Theorem

The expected time for the (1+1)-EA to optimise LeadingOnes is Ω(n2).

Sources of progress

1 Flipping the leftmost zero-bit;

2 Bits to right of the leftmost zero-bit that are one-bits (free riders).

Proof

1 Let the current solution have n− i leading ones (i.e. 1n−i0∗).

2 We define the distance function as the number of missing leading ones, i.e.d(X) = i.

3 The n− i+ 1 bit is a zero;

4 let E[Y ] be the expected number of one-bits after the first zero (i.e. thefree riders).

5 Such i− 1 bits are uniformely distributed at initialisation and still are!

Page 183: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Theorem

The expected time for the (1+1)-EA to optimise LeadingOnes is Ω(n2).

Sources of progress

1 Flipping the leftmost zero-bit;

2 Bits to right of the leftmost zero-bit that are one-bits (free riders).

Proof

1 Let the current solution have n− i leading ones (i.e. 1n−i0∗).

2 We define the distance function as the number of missing leading ones, i.e.d(X) = i.

3 The n− i+ 1 bit is a zero;

4 let E[Y ] be the expected number of one-bits after the first zero (i.e. thefree riders).

5 Such i− 1 bits are uniformely distributed at initialisation and still are!

Page 184: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Theorem

The expected time for the (1+1)-EA to optimise LeadingOnes is Ω(n2).

Sources of progress

1 Flipping the leftmost zero-bit;

2 Bits to right of the leftmost zero-bit that are one-bits (free riders).

Proof

1 Let the current solution have n− i leading ones (i.e. 1n−i0∗).

2 We define the distance function as the number of missing leading ones, i.e.d(X) = i.

3 The n− i+ 1 bit is a zero;

4 let E[Y ] be the expected number of one-bits after the first zero (i.e. thefree riders).

5 Such i− 1 bits are uniformely distributed at initialisation and still are!

Page 185: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Theorem

The expected time for the (1+1)-EA to optimise LeadingOnes is Ω(n2).

Sources of progress

1 Flipping the leftmost zero-bit;

2 Bits to right of the leftmost zero-bit that are one-bits (free riders).

Proof

1 Let the current solution have n− i leading ones (i.e. 1n−i0∗).

2 We define the distance function as the number of missing leading ones, i.e.d(X) = i.

3 The n− i+ 1 bit is a zero;

4 let E[Y ] be the expected number of one-bits after the first zero (i.e. thefree riders).

5 Such i− 1 bits are uniformely distributed at initialisation and still are!

Page 186: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Theorem

The expected time for the (1+1)-EA to optimise LeadingOnes is Ω(n2).

Sources of progress

1 Flipping the leftmost zero-bit;

2 Bits to right of the leftmost zero-bit that are one-bits (free riders).

Proof

1 Let the current solution have n− i leading ones (i.e. 1n−i0∗).

2 We define the distance function as the number of missing leading ones, i.e.d(X) = i.

3 The n− i+ 1 bit is a zero;

4 let E[Y ] be the expected number of one-bits after the first zero (i.e. thefree riders).

5 Such i− 1 bits are uniformely distributed at initialisation and still are!

Page 187: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Theorem

The expected time for the (1+1)-EA to optimise LeadingOnes is Ω(n2).

Sources of progress

1 Flipping the leftmost zero-bit;

2 Bits to right of the leftmost zero-bit that are one-bits (free riders).

Proof

1 Let the current solution have n− i leading ones (i.e. 1n−i0∗).

2 We define the distance function as the number of missing leading ones, i.e.d(X) = i.

3 The n− i+ 1 bit is a zero;

4 let E[Y ] be the expected number of one-bits after the first zero (i.e. thefree riders).

5 Such i− 1 bits are uniformely distributed at initialisation and still are!

Page 188: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Drift Theorem for LeadingOnes (lower bound)

Theorem

The expected time for the (1+1)-EA to optimise LeadingOnes is Ω(n2).

The expected number of free riders is:

E[Y ] =

i−1∑k=1

k · Pr(Y = k) =

i−1∑k=1

Pr(Y ≥ k) =

i−1∑k=1

(1/2)k ≤ 1

The negative drift is 0;

Let p(A) be the probability that the first zero-bit flips into a one-bit.

The positive drift (i.e. the decrease in distance) is bounded as follows:

E(∆+(t)) ≤ p(A) · E[∆+(t)|A] = 1/n · (1 + E[Y ]) ≤ 2/n = δ

Since, also at initialisation the expected number of free riders is less than1, it follows that E[d(X0)] ≥ n− 1,

By applying the Drift Theorem we get

E(T ) ≥ E(d(X0)

δ=n− 1

2/n= Ω(n2)

Page 189: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Drift Theorem for LeadingOnes (lower bound)

Theorem

The expected time for the (1+1)-EA to optimise LeadingOnes is Ω(n2).

The expected number of free riders is:

E[Y ] =

i−1∑k=1

k · Pr(Y = k) =

i−1∑k=1

Pr(Y ≥ k) =

i−1∑k=1

(1/2)k ≤ 1

The negative drift is 0;

Let p(A) be the probability that the first zero-bit flips into a one-bit.

The positive drift (i.e. the decrease in distance) is bounded as follows:

E(∆+(t)) ≤ p(A) · E[∆+(t)|A] = 1/n · (1 + E[Y ]) ≤ 2/n = δ

Since, also at initialisation the expected number of free riders is less than1, it follows that E[d(X0)] ≥ n− 1,

By applying the Drift Theorem we get

E(T ) ≥ E(d(X0)

δ=n− 1

2/n= Ω(n2)

Page 190: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Drift Theorem for LeadingOnes (lower bound)

Theorem

The expected time for the (1+1)-EA to optimise LeadingOnes is Ω(n2).

The expected number of free riders is:

E[Y ] =

i−1∑k=1

k · Pr(Y = k) =

i−1∑k=1

Pr(Y ≥ k) =

i−1∑k=1

(1/2)k ≤ 1

The negative drift is 0;

Let p(A) be the probability that the first zero-bit flips into a one-bit.

The positive drift (i.e. the decrease in distance) is bounded as follows:

E(∆+(t)) ≤ p(A) · E[∆+(t)|A] = 1/n · (1 + E[Y ]) ≤ 2/n = δ

Since, also at initialisation the expected number of free riders is less than1, it follows that E[d(X0)] ≥ n− 1,

By applying the Drift Theorem we get

E(T ) ≥ E(d(X0)

δ=n− 1

2/n= Ω(n2)

Page 191: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Drift Theorem for LeadingOnes (lower bound)

Theorem

The expected time for the (1+1)-EA to optimise LeadingOnes is Ω(n2).

The expected number of free riders is:

E[Y ] =

i−1∑k=1

k · Pr(Y = k) =

i−1∑k=1

Pr(Y ≥ k) =

i−1∑k=1

(1/2)k ≤ 1

The negative drift is 0;

Let p(A) be the probability that the first zero-bit flips into a one-bit.

The positive drift (i.e. the decrease in distance) is bounded as follows:

E(∆+(t)) ≤ p(A) · E[∆+(t)|A] = 1/n · (1 + E[Y ]) ≤ 2/n = δ

Since, also at initialisation the expected number of free riders is less than1, it follows that E[d(X0)] ≥ n− 1,

By applying the Drift Theorem we get

E(T ) ≥ E(d(X0)

δ=n− 1

2/n= Ω(n2)

Page 192: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Additive Drift Theorem

Drift Theorem for LeadingOnes (lower bound)

Theorem

The expected time for the (1+1)-EA to optimise LeadingOnes is Ω(n2).

The expected number of free riders is:

E[Y ] =

i−1∑k=1

k · Pr(Y = k) =

i−1∑k=1

Pr(Y ≥ k) =

i−1∑k=1

(1/2)k ≤ 1

The negative drift is 0;

Let p(A) be the probability that the first zero-bit flips into a one-bit.

The positive drift (i.e. the decrease in distance) is bounded as follows:

E(∆+(t)) ≤ p(A) · E[∆+(t)|A] = 1/n · (1 + E[Y ]) ≤ 2/n = δ

Since, also at initialisation the expected number of free riders is less than1, it follows that E[d(X0)] ≥ n− 1,

By applying the Drift Theorem we get

E(T ) ≥ E(d(X0)

δ=n− 1

2/n= Ω(n2)

Page 193: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Multiplicative Drift Theorem

Drift Analysis for OneMax

Lets calculate the runtime of the (1+1)-EA using the additive Drift Theorem.

1 Let d(Xt) = i where i is the number of zeroes in the bitstring;

2 The negative drift is 0 since solution with less one-bits will not beaccepted;

3 A positive drift is achieved as long as a 0 is flipped and the ones remainunchanged:

E(∆(t)) = E[d(Xt)−d(Xt+1)|d(Xt) = i] ≥ 1· in

(1− 1

n

)n−1

≥ i

en≥ 1

en:= δ

4 The expected initial distance is E(d(X0)) = n/2

The expected runtime is (i.e. Eq. (6)):

E(T | d(X0) > 0) ≤ E[(d(X0)]

δ≤ n/2

1/(en)= e/2 · n2 = O(n2)

We need a different distance function!

Page 194: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Multiplicative Drift Theorem

Drift Analysis for OneMax

Lets calculate the runtime of the (1+1)-EA using the additive Drift Theorem.

1 Let d(Xt) = i where i is the number of zeroes in the bitstring;

2 The negative drift is 0 since solution with less one-bits will not beaccepted;

3 A positive drift is achieved as long as a 0 is flipped and the ones remainunchanged:

E(∆(t)) = E[d(Xt)−d(Xt+1)|d(Xt) = i] ≥ 1· in

(1− 1

n

)n−1

≥ i

en≥ 1

en:= δ

4 The expected initial distance is E(d(X0)) = n/2

The expected runtime is (i.e. Eq. (6)):

E(T | d(X0) > 0) ≤ E[(d(X0)]

δ≤ n/2

1/(en)= e/2 · n2 = O(n2)

We need a different distance function!

Page 195: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Multiplicative Drift Theorem

Drift Analysis for OneMax

Lets calculate the runtime of the (1+1)-EA using the additive Drift Theorem.

1 Let d(Xt) = i where i is the number of zeroes in the bitstring;

2 The negative drift is 0 since solution with less one-bits will not beaccepted;

3 A positive drift is achieved as long as a 0 is flipped and the ones remainunchanged:

E(∆(t)) = E[d(Xt)−d(Xt+1)|d(Xt) = i] ≥ 1· in

(1− 1

n

)n−1

≥ i

en≥ 1

en:= δ

4 The expected initial distance is E(d(X0)) = n/2

The expected runtime is (i.e. Eq. (6)):

E(T | d(X0) > 0) ≤ E[(d(X0)]

δ≤ n/2

1/(en)= e/2 · n2 = O(n2)

We need a different distance function!

Page 196: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Multiplicative Drift Theorem

Drift Analysis for OneMax

Lets calculate the runtime of the (1+1)-EA using the additive Drift Theorem.

1 Let d(Xt) = i where i is the number of zeroes in the bitstring;

2 The negative drift is 0 since solution with less one-bits will not beaccepted;

3 A positive drift is achieved as long as a 0 is flipped and the ones remainunchanged:

E(∆(t)) = E[d(Xt)−d(Xt+1)|d(Xt) = i] ≥ 1· in

(1− 1

n

)n−1

≥ i

en≥ 1

en:= δ

4 The expected initial distance is E(d(X0)) = n/2

The expected runtime is (i.e. Eq. (6)):

E(T | d(X0) > 0) ≤ E[(d(X0)]

δ≤ n/2

1/(en)= e/2 · n2 = O(n2)

We need a different distance function!

Page 197: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Multiplicative Drift Theorem

Drift Analysis for OneMax

1 Let g(Xt) = ln(i+ 1) where i is the number of zeroes in the bitstring;

2 For x ≥ 1, it holds that ln(1 + 1/x) ≥ 1/x− 1/(2x2) ≥ 1/(2x);

3 The distance decreases as long as a 0 is flipped and the ones remainunchanged:

E(∆(t)) = E[d(Xt)− d(Xt+1)|d(Xt) = i ≥ 1]

≥ i

en

(ln(i+ 1)− ln(i)

)=

i

enln

(1 +

1

i

)≥ i

en

1

2i=

1

2en:= δ

4 The initial distance is d(X0) ≤ ln(n+ 1)

The expected runtime is (i.e. Eq. (6)):

E(T | d(X0) > 0) ≤ d(X0)

δ≤ ln(n+ 1)

1/(2en)= O(n lnn)

If the amount of progress depends on the distance from the optimum we needto use a logarithmic distance!

Page 198: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Multiplicative Drift Theorem

Drift Analysis for OneMax

1 Let g(Xt) = ln(i+ 1) where i is the number of zeroes in the bitstring;

2 For x ≥ 1, it holds that ln(1 + 1/x) ≥ 1/x− 1/(2x2) ≥ 1/(2x);

3 The distance decreases as long as a 0 is flipped and the ones remainunchanged:

E(∆(t)) = E[d(Xt)− d(Xt+1)|d(Xt) = i ≥ 1]

≥ i

en

(ln(i+ 1)− ln(i)

)=

i

enln

(1 +

1

i

)≥ i

en

1

2i=

1

2en:= δ

4 The initial distance is d(X0) ≤ ln(n+ 1)

The expected runtime is (i.e. Eq. (6)):

E(T | d(X0) > 0) ≤ d(X0)

δ≤ ln(n+ 1)

1/(2en)= O(n lnn)

If the amount of progress depends on the distance from the optimum we needto use a logarithmic distance!

Page 199: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Multiplicative Drift Theorem

Drift Analysis for OneMax

1 Let g(Xt) = ln(i+ 1) where i is the number of zeroes in the bitstring;

2 For x ≥ 1, it holds that ln(1 + 1/x) ≥ 1/x− 1/(2x2) ≥ 1/(2x);

3 The distance decreases as long as a 0 is flipped and the ones remainunchanged:

E(∆(t)) = E[d(Xt)− d(Xt+1)|d(Xt) = i ≥ 1]

≥ i

en

(ln(i+ 1)− ln(i)

)=

i

enln

(1 +

1

i

)≥ i

en

1

2i=

1

2en:= δ

4 The initial distance is d(X0) ≤ ln(n+ 1)

The expected runtime is (i.e. Eq. (6)):

E(T | d(X0) > 0) ≤ d(X0)

δ≤ ln(n+ 1)

1/(2en)= O(n lnn)

If the amount of progress depends on the distance from the optimum we needto use a logarithmic distance!

Page 200: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Multiplicative Drift Theorem

Drift Analysis for OneMax

1 Let g(Xt) = ln(i+ 1) where i is the number of zeroes in the bitstring;

2 For x ≥ 1, it holds that ln(1 + 1/x) ≥ 1/x− 1/(2x2) ≥ 1/(2x);

3 The distance decreases as long as a 0 is flipped and the ones remainunchanged:

E(∆(t)) = E[d(Xt)− d(Xt+1)|d(Xt) = i ≥ 1]

≥ i

en

(ln(i+ 1)− ln(i)

)=

i

enln

(1 +

1

i

)≥ i

en

1

2i=

1

2en:= δ

4 The initial distance is d(X0) ≤ ln(n+ 1)

The expected runtime is (i.e. Eq. (6)):

E(T | d(X0) > 0) ≤ d(X0)

δ≤ ln(n+ 1)

1/(2en)= O(n lnn)

If the amount of progress depends on the distance from the optimum we needto use a logarithmic distance!

Page 201: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Multiplicative Drift Theorem

Multiplicative Drift Theorem

Theorem (Multiplicative Drift, [Doerr et al., 2010])

Let Xtt∈N0 be random variables describing a Markov process over a finitestate space S ⊆ R. Let T be the random variable that denotes the earliestpoint in time t ∈ N0 such that Xt = 0.If there exist δ, cmin, cmax > 0 such that

1 E[Xt −Xt+1 | Xt] ≥ δXt and

2 cmin ≤ Xt ≤ cmax,

for all t < T , then

E[T ] ≤ 2

δ· ln(

1 +cmax

cmin

)

Page 202: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Multiplicative Drift Theorem

(1+1)-EA Analysis for OneMax

Theorem

The expected time for the (1+1)-EA to optimise OneMaxis O(n lnn)

Proof

Distance: let Xt be the number of zeroes at time step t;

E[Xt+1|Xt] ≤ Xt − 1 · Xten

= Xt ·(1− 1

en

)E[Xt −Xt+1|Xt] ≤ Xt −Xt ·

(1− 1

en

)= Xt

en(δ = 1

en)

1 = cmin ≤ Xt ≤ cmax = n

Hence,

E[T ] ≤ 2

δ· ln(

1 +cmax

cmin

)= 2en ln(1 + n) = O(n lnn)

Page 203: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Multiplicative Drift Theorem

(1+1)-EA Analysis for OneMax

Theorem

The expected time for the (1+1)-EA to optimise OneMaxis O(n lnn)

Proof

Distance: let Xt be the number of zeroes at time step t;

E[Xt+1|Xt] ≤ Xt − 1 · Xten

= Xt ·(1− 1

en

)E[Xt −Xt+1|Xt] ≤ Xt −Xt ·

(1− 1

en

)= Xt

en(δ = 1

en)

1 = cmin ≤ Xt ≤ cmax = n

Hence,

E[T ] ≤ 2

δ· ln(

1 +cmax

cmin

)= 2en ln(1 + n) = O(n lnn)

Page 204: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Multiplicative Drift Theorem

Exercises

Theorem

The expected time for RLS to optimise OneMaxis O(n logn)

Proof

Left as exercise.

Theorem

Let λ ≥ en. Then the expected time for the (1+λ)-EA to optimise OneMaxisO(λn)

Proof Left as exercise.

Theorem

Let λ < en. Then the expected time for the (1+λ)-EA to optimise OneMaxisO(n logn)

Proof Left as exercise.

Page 205: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Multiplicative Drift Theorem

Exercises

Theorem

The expected time for RLS to optimise OneMaxis O(n logn)

Proof Left as exercise.

Theorem

Let λ ≥ en. Then the expected time for the (1+λ)-EA to optimise OneMaxisO(λn)

Proof

Left as exercise.

Theorem

Let λ < en. Then the expected time for the (1+λ)-EA to optimise OneMaxisO(n logn)

Proof Left as exercise.

Page 206: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Multiplicative Drift Theorem

Exercises

Theorem

The expected time for RLS to optimise OneMaxis O(n logn)

Proof Left as exercise.

Theorem

Let λ ≥ en. Then the expected time for the (1+λ)-EA to optimise OneMaxisO(λn)

Proof Left as exercise.

Theorem

Let λ < en. Then the expected time for the (1+λ)-EA to optimise OneMaxisO(n logn)

Proof

Left as exercise.

Page 207: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Multiplicative Drift Theorem

Exercises

Theorem

The expected time for RLS to optimise OneMaxis O(n logn)

Proof Left as exercise.

Theorem

Let λ ≥ en. Then the expected time for the (1+λ)-EA to optimise OneMaxisO(λn)

Proof Left as exercise.

Theorem

Let λ < en. Then the expected time for the (1+λ)-EA to optimise OneMaxisO(n logn)

Proof Left as exercise.

Page 208: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Simplified Negative Drift Theorem

Drift Analysis: Example 3

Friday night dinner at the restaurant.Peter walks back from the restaurant to the hotel but had too many drinks.

The restaurant is n meters away from the hotel;

Peter moves towards the hotel of 1 meter in each step with probability 0.4.

Peter moves away from the hotel of 1 meter in each step with probability0.6.

QuestionHow many steps does Peter need to reach his hotel?

at least 2cn steps with overwhelming probability (exponential time)We need Negative-Drift Analysis.

Page 209: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Simplified Negative Drift Theorem

Drift Analysis: Example 3

Friday night dinner at the restaurant.Peter walks back from the restaurant to the hotel but had too many drinks.

The restaurant is n meters away from the hotel;

Peter moves towards the hotel of 1 meter in each step with probability 0.4.

Peter moves away from the hotel of 1 meter in each step with probability0.6.

QuestionHow many steps does Peter need to reach his hotel?at least 2cn steps with overwhelming probability (exponential time)We need Negative-Drift Analysis.

Page 210: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Simplified Negative Drift Theorem

Simplified Drift Theorem

targeta b

drift away from target

no large jumpstowards target

start

Theorem (Simplified Negative-Drift Theorem, [Oliveto and Witt, 2011])

Suppose there exist three constants δ,ε,r such that for all t ≥ 0:

1 E(∆t(i)) ≥ ε for a < i < b,

2 Prob(|∆t(i)| = −j) ≤ 1(1+δ)j−r for i > a and j ≥ 1.

ThenProb(T ∗ ≤ 2c

∗(b−a)) = 2−Ω(b−a)

.

Page 211: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Simplified Negative Drift Theorem

Negative-Drift Analysis: Example (3)

Define the same distance function d(x) = x, x ∈ 0, . . . , n (metres fromthe hotel) (b=n-1, a=1).

Estimate the increase in distance from the goal (negative drift);

d(Xt+1)− d(Xt) =

0, if Xt = 0,

1, if Xt ∈ 1, . . . , nwith probability 0.6

−1, if Xt ∈ 1, . . . , nwith probability 0.4

The expected increase in distance (negative drift) is: (Condition 1)

E[d(Xt+1)− d(Xt)] = 0.6 · 1 + 0.4 · (−1) = 0.6− 0.4 = 0.2

Probability of jumps (i.e. Prob(∆t(i) = −j) ≤ 1(1+δ)j−r ) (set

δ = r = 1) (Condition 2):

Pr(∆t(i) = −j) =

0 < (1/2)j−1, if j > 1,

0.6 < (1/2)0 = 1, if j = 1

Then the expected time to reach the hotel (goal) is:

Pr(T ≤ 2c(b−a)) = Pr(T ≤ 2c(n−2)) = 2−Ω(n)

Page 212: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Simplified Negative Drift Theorem

Negative-Drift Analysis: Example (3)

Define the same distance function d(x) = x, x ∈ 0, . . . , n (metres fromthe hotel) (b=n-1, a=1).Estimate the increase in distance from the goal (negative drift);

d(Xt+1)− d(Xt) =

0, if Xt = 0,

1, if Xt ∈ 1, . . . , nwith probability 0.6

−1, if Xt ∈ 1, . . . , nwith probability 0.4

The expected increase in distance (negative drift) is: (Condition 1)

E[d(Xt+1)− d(Xt)] = 0.6 · 1 + 0.4 · (−1) = 0.6− 0.4 = 0.2

Probability of jumps (i.e. Prob(∆t(i) = −j) ≤ 1(1+δ)j−r ) (set

δ = r = 1) (Condition 2):

Pr(∆t(i) = −j) =

0 < (1/2)j−1, if j > 1,

0.6 < (1/2)0 = 1, if j = 1

Then the expected time to reach the hotel (goal) is:

Pr(T ≤ 2c(b−a)) = Pr(T ≤ 2c(n−2)) = 2−Ω(n)

Page 213: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Simplified Negative Drift Theorem

Negative-Drift Analysis: Example (3)

Define the same distance function d(x) = x, x ∈ 0, . . . , n (metres fromthe hotel) (b=n-1, a=1).Estimate the increase in distance from the goal (negative drift);

d(Xt+1)− d(Xt) =

0, if Xt = 0,

1, if Xt ∈ 1, . . . , nwith probability 0.6

−1, if Xt ∈ 1, . . . , nwith probability 0.4

The expected increase in distance (negative drift) is: (Condition 1)

E[d(Xt+1)− d(Xt)] = 0.6 · 1 + 0.4 · (−1) = 0.6− 0.4 = 0.2

Probability of jumps (i.e. Prob(∆t(i) = −j) ≤ 1(1+δ)j−r ) (set

δ = r = 1) (Condition 2):

Pr(∆t(i) = −j) =

0 < (1/2)j−1, if j > 1,

0.6 < (1/2)0 = 1, if j = 1

Then the expected time to reach the hotel (goal) is:

Pr(T ≤ 2c(b−a)) = Pr(T ≤ 2c(n−2)) = 2−Ω(n)

Page 214: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Simplified Negative Drift Theorem

Negative-Drift Analysis: Example (3)

Define the same distance function d(x) = x, x ∈ 0, . . . , n (metres fromthe hotel) (b=n-1, a=1).Estimate the increase in distance from the goal (negative drift);

d(Xt+1)− d(Xt) =

0, if Xt = 0,

1, if Xt ∈ 1, . . . , nwith probability 0.6

−1, if Xt ∈ 1, . . . , nwith probability 0.4

The expected increase in distance (negative drift) is: (Condition 1)

E[d(Xt+1)− d(Xt)] = 0.6 · 1 + 0.4 · (−1) = 0.6− 0.4 = 0.2

Probability of jumps (i.e. Prob(∆t(i) = −j) ≤ 1(1+δ)j−r ) (set

δ = r = 1) (Condition 2):

Pr(∆t(i) = −j) =

0 < (1/2)j−1, if j > 1,

0.6 < (1/2)0 = 1, if j = 1

Then the expected time to reach the hotel (goal) is:

Pr(T ≤ 2c(b−a)) = Pr(T ≤ 2c(n−2)) = 2−Ω(n)

Page 215: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Simplified Negative Drift Theorem

Negative-Drift Analysis: Example (3)

Define the same distance function d(x) = x, x ∈ 0, . . . , n (metres fromthe hotel) (b=n-1, a=1).Estimate the increase in distance from the goal (negative drift);

d(Xt+1)− d(Xt) =

0, if Xt = 0,

1, if Xt ∈ 1, . . . , nwith probability 0.6

−1, if Xt ∈ 1, . . . , nwith probability 0.4

The expected increase in distance (negative drift) is: (Condition 1)

E[d(Xt+1)− d(Xt)] = 0.6 · 1 + 0.4 · (−1) = 0.6− 0.4 = 0.2

Probability of jumps (i.e. Prob(∆t(i) = −j) ≤ 1(1+δ)j−r ) (set

δ = r = 1) (Condition 2):

Pr(∆t(i) = −j) =

0 < (1/2)j−1, if j > 1,

0.6 < (1/2)0 = 1, if j = 1

Then the expected time to reach the hotel (goal) is:

Pr(T ≤ 2c(b−a)) = Pr(T ≤ 2c(n−2)) = 2−Ω(n)

Page 216: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Simplified Negative Drift Theorem

Needle in a Haystack

Theorem (Oliveto,Witt, Algorithmica 2011)

Let η > 0 be constant. Then there is a constant c > 0 such that withprobability 1− 2−Ω(n) the (1+1)-EA on Needle creates only search pointswith at most n/2 + ηn ones in 2cn steps.

Proof Idea

By Chernoff bounds the probability that the initial bit string has less thann/2− γn zeroes is e−Ω(n).

we set b := n/2− γn and a := n/2− 2γn where γ := η/2 (we have izero-bits and n− i one-bits);

Proof of Condition 1

E(∆(i)) =n− in− i

n=

n− 2i

n≥ 2γ = ε

Proof of Condition 2

Prob(|∆(i)| ≤ −j) ≤

(n

j

)(1

n

)j≤ nj

j!

(1

n

)j=

1

j!≤(

1

2

)j−1

This proves Condition 2 by setting δ = r = 1.

Page 217: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Simplified Negative Drift Theorem

Needle in a Haystack

Theorem (Oliveto,Witt, Algorithmica 2011)

Let η > 0 be constant. Then there is a constant c > 0 such that withprobability 1− 2−Ω(n) the (1+1)-EA on Needle creates only search pointswith at most n/2 + ηn ones in 2cn steps.

Proof Idea

By Chernoff bounds the probability that the initial bit string has less thann/2− γn zeroes is e−Ω(n).

we set b := n/2− γn and a := n/2− 2γn where γ := η/2 (we have izero-bits and n− i one-bits);

Proof of Condition 1

E(∆(i)) =n− in− i

n=

n− 2i

n≥ 2γ = ε

Proof of Condition 2

Prob(|∆(i)| ≤ −j) ≤

(n

j

)(1

n

)j≤ nj

j!

(1

n

)j=

1

j!≤(

1

2

)j−1

This proves Condition 2 by setting δ = r = 1.

Page 218: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Simplified Negative Drift Theorem

Exercise: Trap Functions

Trap(x) =

n+ 1 if x = 0n

OneMax(x) otherwise.

0 ones(x)

f(x)

3

3

2

2

n

nn+ 1

1

1

Theorem

With overwhelming probability at least 1− 2−Ω(n) the (1+1)-EA requires 2Ω(n)

steps to optimise Trap.

Proof Left as exercise.

Page 219: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Simplified Negative Drift Theorem

Exercise: Trap Functions

Trap(x) =

n+ 1 if x = 0n

OneMax(x) otherwise.

0 ones(x)

f(x)

3

3

2

2

n

nn+ 1

1

1

Theorem

With overwhelming probability at least 1− 2−Ω(n) the (1+1)-EA requires 2Ω(n)

steps to optimise Trap.

Proof Left as exercise.

Page 220: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Simplified Negative Drift Theorem

Drift Analysis Conclusion

Overview

Additive Drift Analysis (upper and lower bounds);

Multiplicative Drift Analysis;

Simplified Negative-Drift Theorem;

Advanced Lower bound Drift Techniques

Drift Analysis for Stochastic Populations (mutation) [Lehre, 2010];

Simplified Drift Theorem combined with bandwidth analysis (mutation +crossover stochastic populations = GAs) [Oliveto and Witt, 2012];(See Oliveto,Lehre Tutorial at GECCO 2017 in Berlin)

Page 221: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Overview

Final Overview

Overview

Basic Probability Theory

Tail Inequalities

Artificial Fitness Levels

Drift Analysis

Other Techniques (Not covered)

Family Trees [Witt, 2006]

Gambler’s Ruin & Martingales [Jansen and Wegener, 2001]

Page 222: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

State-of-the-art

Not only toy problems...

MST (1+1) EA Θ(m2 log(nwmax))(1+λ) EA O(n log(nwmax))1-ANT O(mn log(nwmax))

Max. Clique (1+1) EA Θ(n5)

(rand. planar) (16n+1) RLS Θ(n5/3)

Eulerian Cycle (1+1) EA Θ(m2 logm)Partition (1+1) EA 4/3 approx., competitive avg.

Vertex Cover (1+1) EA eΩ(n) , arb. bad approx.

Set Cover (1+1) EA eΩ(n) , arb. bad approx.SEMO Pol. O(logn)-approx.

Intersection of (1+1) EA 1/p-approximation in

p ≥ 3 matroids O(|E|p+2 log(|E|wmax))

UIO/FSM conf. (1+1) EA eΩ(n)

See [Oliveto et al., 2007] for an overview.

Page 223: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Further reading

Further Reading

[Neumann and Witt, 2010, Auger and Doerr, 2011, Jansen, 2013]

Page 224: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Further reading

References I

Auger, A. and Doerr, B. (2011).

Theory of Randomized Search Heuristics: Foundations and Recent Developments.World Scientific Publishing Co., Inc., River Edge, NJ, USA.

Back, T. (1993).

Optimal mutation rates in genetic search.In In Proceedings of the Fifth International Conference on Genetic Algorithms (ICGA), pages 2–8.

Corus, D. and Oliveto, P. S. (2017).

Standard steady state genetic algorithms can hillclimb faster than mutation-only evolutionary algorithms.IEEE Transactions on Evolutionary Computation.

Doerr, B., Johannsen, D., and Winzen, C. (2010).

Multiplicative drift analysis.In Proceedings of the 12th annual conference on Genetic and evolutionary computation, GECCO ’10, pages1449–1456. ACM.

Droste, S., Jansen, T., and Wegener, I. (1998).

A rigorous complexity analysis of the (1 + 1) evolutionary algorithm for separable functions with booleaninputs.Evolutionary Computation, 6(2):185–196.

Droste, S., Jansen, T., and Wegener, I. (2002).

On the analysis of the (1+1) evolutionary algorithm.Theoretical Computer Science, 276(1-2):51–81.

Goldberg, D. E. (1989).

Genetic Algorithms for Search, Optimization, and Machine Learning.Addison-Wesley.

Page 225: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Further reading

References II

He, J. and Yao, X. (2001).

Drift analysis and average time complexity of evolutionary algorithms.Artificial Intelligence, 127(1):57–85.

He, J. and Yao, X. (2004).

A study of drift analysis for estimating computation time of evolutionary algorithms.Natural Computing: an international journal, 3(1):21–35.

Holland, J. H. (1992).

Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology,Control, and Artificial Intelligence.The MIT Press.

Jansen, T. (2013).

Analyzing Evolutionary Algorithms.Springer-Verlag New York, Inc., New York, NY, USA, 1st edition.

Jansen, T., Jong, K. A. D., and Wegener, I. A. (2005).

On the choice of the offspring population size in evolutionary algorithms.Evolutionary Computation, 13(4):413–440.

Jansen, T. and Wegener, I. (2001).

Evolutionary algorithms - how to cope with plateaus of constant fitness and when to reject strings of thesame fitness.IEEE Trans. Evolutionary Computation, 5(6):589–599.

Lehre, P. K. (2010).

Negative drift in populations.In PPSN (1), pages 244–253.

Page 226: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Further reading

References III

Lehre, P. K. (2011).

Fitness-levels for non-elitist populations.In Proceedings of the 13th annual conference on Genetic and evolutionary computation, GECCO ’11, pages2075–2082. ACM.

Neumann, F. and Witt, C. (2010).

Bioinspired Computation in Combinatorial Optimization: Algorithms and Their Computational Complexity.Springer-Verlag New York, Inc., New York, NY, USA, 1st edition.

Oliveto, P. and Witt, C. (2012).

On the analysis of the simple genetic algorithm (to appear).In Proceedings of the 12th annual conference on Genetic and evolutionary computation, GECCO ’12,pages –. ACM.

Oliveto, P. S., He, J., and Yao, X. (2007).

Time complexity of evolutionary algorithms for combinatorial optimization: A decade of results.International Journal of Automation and Computing, 4(3):281–293.

Oliveto, P. S. and Witt, C. (2011).

Simplified drift analysis for proving lower bounds inevolutionary computation.Algorithmica, 59(3):369–386.

Reeves, C. R. and Rowe, J. E. (2002).

Genetic Algorithms: Principles and Perspectives: A Guide to GA Theory.Kluwer Academic Publishers, Norwell, MA, USA.

Rudolph, G. (1998).

Finite Markov chain results in evolutionary computation: A tour d’horizon.Fundamenta Informaticae, 35(1–4):67–89.

Page 227: A Gentle Introduction to the Time Complexity Analysis of … · 2017-12-08 · Motivation Evolutionary Algorithms Tail Inequalities Arti cial Fitness Levels Drift Analysis Conclusions

Motivation Evolutionary Algorithms Tail Inequalities Artificial Fitness Levels Drift Analysis Conclusions

Further reading

References IV

Sudholt, D. (2010).

General lower bounds for the running time of evolutionary algorithms.In PPSN (1), pages 124–133.

Witt, C. (2006).

Runtime analysis of the (µ+1) ea on simple pseudo-boolean functions evolutionary computation.In GECCO ’06: Proceedings of the 8th annual conference on Genetic and evolutionary computation, pages651–658, New York, NY, USA. ACM Press.


Recommended