As seen in the Covariance Matrix Adaptation Evolution Strategy example, the eaGenerateUpdate() algorithm is suitable for algorithms learning the problem distribution from the population. Here we’ll cover how to implement a strategy that generates individuals based on an updated sampling function learnt from the sampled population.
The basic concept concept behind EDA is to sample  individuals
with a certain distribution and estimate the problem distribution from the
 individuals
with a certain distribution and estimate the problem distribution from the
 best individuals. This really simple concept adhere to the
generate-update logic. The strategy contains a random number generator which
is adapted from the population. The following EDA class do just that.
 best individuals. This really simple concept adhere to the
generate-update logic. The strategy contains a random number generator which
is adapted from the population. The following EDA class do just that.
class EDA(object):
    def __init__(self, centroid, sigma, mu, lambda_):
        self.dim = len(centroid)
        self.loc = numpy.array(centroid)
        self.sigma = numpy.array(sigma)
        self.lambda_ = lambda_
        self.mu = mu
    
    def generate(self, ind_init):
        # Generate lambda_ individuals and put them into the provided class
        arz = self.sigma * numpy.random.randn(self.lambda_, self.dim) + self.loc
        return list(map(ind_init, arz))
    
    def update(self, population):
        # Sort individuals so the best is first
        sorted_pop = sorted(population, key=attrgetter("fitness"), reverse=True)
        
        # Compute the average of the mu best individuals
        z = sorted_pop[:self.mu] - self.loc
        avg = numpy.mean(z, axis=0)
        
        # Adjust variances of the distribution
        self.sigma = numpy.sqrt(numpy.sum((z - avg)**2, axis=0) / (self.mu - 1.0))
        self.loc = self.loc + avg
A normal random number generator is initialized with a certain mean (centroid) and standard deviation (sigma) for each dimension. The generate() method uses numpy to generate lambda_ sequences in dim dimensions, then the sequences are used to initialize individuals of class given in the ind_init argument. Finally, the update() computes the average (centre) of the mu best individuals and estimates the variance over all attributes of each individual. Once update() is called the distributions parameters are changed and a new population can be generated.
Two classes are needed, a minimization fitness and a individual that will combine the fitness and the real values. Moreover, we will use numpy.ndarray as base class for our individuals.
creator.create("FitnessMin", base.Fitness, weights=(-1.0,))
creator.create("Individual", numpy.ndarray, fitness=creator.FitnessMin) 
The eaGenerateUpdate() algorithm requires to set in a toolbox an evaluation function, an generation method and an update method. We will use the method of an initialized EDA. For the generate method, we set the class that the individuals are transferred in to our Individual class containing a fitness.
def main():
    N, LAMBDA = 30, 1000
    MU = int(LAMBDA/4)
    strategy = EDA(centroid=[5.0]*N, sigma=[5.0]*N, mu=MU, lambda_=LAMBDA)
    
    toolbox = base.Toolbox()
    toolbox.register("evaluate", benchmarks.rastrigin)
    toolbox.register("generate", strategy.generate, creator.Individual)
    toolbox.register("update", strategy.update)
    hof = tools.HallOfFame(1)
    stats = tools.Statistics(lambda ind: ind.fitness.values)
    stats.register("avg", tools.mean)
    stats.register("std", tools.std)
    stats.register("min", min)
    stats.register("max", max)
    
    algorithms.eaGenerateUpdate(toolbox, ngen=150, stats=stats, halloffame=hof)
    
    return hof[0].fitness.values[0]
The complete example : [source code].