All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.cicirello.search.sa.ModifiedLam Maven / Gradle / Ivy

Go to download

Chips-n-Salsa is a Java library of customizable, hybridizable, iterative, parallel, stochastic, and self-adaptive local search algorithms. The library includes implementations of several stochastic local search algorithms, including simulated annealing, hill climbers, as well as constructive search algorithms such as stochastic sampling. Chips-n-Salsa now also includes genetic algorithms as well as evolutionary algorithms more generally. The library very extensively supports simulated annealing. It includes several classes for representing solutions to a variety of optimization problems. For example, the library includes a BitVector class that implements vectors of bits, as well as classes for representing solutions to problems where we are searching for an optimal vector of integers or reals. For each of the built-in representations, the library provides the most common mutation operators for generating random neighbors of candidate solutions, as well as common crossover operators for use with evolutionary algorithms. Additionally, the library provides extensive support for permutation optimization problems, including implementations of many different mutation operators for permutations, and utilizing the efficiently implemented Permutation class of the JavaPermutationTools (JPT) library. Chips-n-Salsa is customizable, making extensive use of Java's generic types, enabling using the library to optimize other types of representations beyond what is provided in the library. It is hybridizable, providing support for integrating multiple forms of local search (e.g., using a hill climber on a solution generated by simulated annealing), creating hybrid mutation operators (e.g., local search using multiple mutation operators), as well as support for running more than one type of search for the same problem concurrently using multiple threads as a form of algorithm portfolio. Chips-n-Salsa is iterative, with support for multistart metaheuristics, including implementations of several restart schedules for varying the run lengths across the restarts. It also supports parallel execution of multiple instances of the same, or different, stochastic local search algorithms for an instance of a problem to accelerate the search process. The library supports self-adaptive search in a variety of ways, such as including implementations of adaptive annealing schedules for simulated annealing, such as the Modified Lam schedule, implementations of the simpler annealing schedules but which self-tune the initial temperature and other parameters, and restart schedules that adapt to run length.

There is a newer version: 7.0.1
Show newest version
/*
 * Chips-n-Salsa: A library of parallel self-adaptive local search algorithms.
 * Copyright (C) 2002-2021  Vincent A. Cicirello
 *
 * This file is part of Chips-n-Salsa (https://chips-n-salsa.cicirello.org/).
 *
 * Chips-n-Salsa is free software: you can redistribute it and/or modify
 * it under the terms of the GNU General Public License as published by
 * the Free Software Foundation, either version 3 of the License, or
 * (at your option) any later version.
 *
 * Chips-n-Salsa is distributed in the hope that it will be useful,
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 * GNU General Public License for more details.
 *
 * You should have received a copy of the GNU General Public License
 * along with this program.  If not, see .
 */

package org.cicirello.search.sa;

import java.util.concurrent.ThreadLocalRandom;

/**
 * This class implements an optimized variant of the Modified Lam annealing schedule. The Modified
 * Lam annealing schedule dynamically adjusts simulated annealing's temperature parameter up and
 * down to either decrease or increase the neighbor acceptance rate as necessary to attempt to match
 * a theoretically determined ideal. The Modified Lam annealing schedule is a practical realization
 * of Lam and Delosme's (1988) schedule, refined first by Swartz (1993) and then further by Boyan
 * (1998).
 *
 * 

This optimized version of the Modified Lam is described in the following article:
* Vincent A. Cicirello. 2020. Optimizing the Modified Lam * Annealing Schedule. Industrial Networks and Intelligent Systems, 7(25): 1-11, Article * e1 (December 2020). doi:10.4108/eai.16-12-2020.167653. * *

This optimized Java implementation is significantly faster than the implementation that would * result from a direct implementation as described originally by Boyan (1998). Specifically, in the * original Boyan's Modified Lam, the update of the target rate of acceptance involves an * exponentiation. This update occurs once for each iteration of simulated annealing. However, in * the Optimized Modified Lam of Cicirello (2020), the target rate is instead computed incrementally * from the prior rate. If the simulated annealing run is n evaluations in length, then the direct * implementation of Boyan's Modified Lam schedule performs n/2 exponentiations in total across all * updates of the target rate; while the Optimized Modified Lam instead perform only 2 * exponentiations and n/2 multiplications total across all updates of the target rate. The schedule * of target acceptance rates is otherwise the same. * *

For details of the original Modified Lam schedule, such as its origins and rationale, see the * following references: * *

    *
  • Lam, J., and Delosme, J. 1988. Performance of a new annealing schedule. In Proc. 25th * ACM/IEEE DAC, 306–311. *
  • Swartz, W. P. 1993. Automatic Layout of Analog and Digital Mixed Macro/Standard Cell * Integrated Circuits. Ph.D. Dissertation, Yale University. *
  • Boyan, J. A. 1998. Learning Evaluation Functions for Global Optimization. Ph.D. * Dissertation, Carnegie Mellon University, Pittsburgh, PA. *
* *

The Chips-n-Salsa library also includes an implementation of the original Modified Lam * schedule that is the result of a direct implementation of Boyan's description of the annealing * schedule, see the {@link ModifiedLamOriginal} class for that version. * *

The {@link #accept} methods of this class use the classic, and most common, Boltzmann * distribution for determining whether to accept a neighbor. * * @author Vincent A. Cicirello, https://www.cicirello.org/ * @version 1.22.2021 */ public final class ModifiedLam implements AnnealingSchedule { private double t; private double acceptRate; private double targetRate; private double phase1; private double phase2; private int iterationCount; private double termPhase1; private double multPhase1; private double multPhase3; private int lastMaxEvals; /** * Default constructor. The Modified Lam annealing schedule, unlike other annealing schedules, has * no control parameters other than the run length (the maxEvals parameter of the {@link #init} * method), so no parameters need be passed to the constructor. */ public ModifiedLam() { lastMaxEvals = -1; } @Override public void init(int maxEvals) { t = 0.5; acceptRate = 0.5; targetRate = 1.0; iterationCount = 0; termPhase1 = 0.56; if (lastMaxEvals != maxEvals) { // These don't change during the run, and only depend // on maxEvals. So initialize only if run length // has changed. phase1 = 0.15 * maxEvals; phase2 = 0.65 * maxEvals; multPhase1 = Math.pow(560, -1.0 / phase1); multPhase3 = Math.pow(440, -1.0 / (maxEvals - phase2)); lastMaxEvals = maxEvals; } } @Override public boolean accept(double neighborCost, double currentCost) { boolean doAccept = neighborCost <= currentCost || ThreadLocalRandom.current().nextDouble() < Math.exp((currentCost - neighborCost) / t); updateSchedule(doAccept); return doAccept; } @Override public ModifiedLam split() { return new ModifiedLam(); } private void updateSchedule(boolean doAccept) { if (doAccept) acceptRate = 0.998 * acceptRate + 0.002; else acceptRate = 0.998 * acceptRate; iterationCount++; if (iterationCount <= phase1) { // Original Modified Lam schedule indicates that targetRate should // be set in phase 1 (first 15% of run) // to: 0.44 + 0.56 * Math.pow(560, -1.0*iterationCount/phase1); // That involves a pow for each phase 1 iteration. We instead compute it // incrementally with 1 call to pow in the init, and 1 multiplication per // phase 1 update. termPhase1 *= multPhase1; targetRate = 0.44 + termPhase1; } else if (iterationCount > phase2) { // Original Modified Lam schedule indicates that targetRate should // be set in phase 3 (last 35% of run) // to: 0.44 * Math.pow(440, -(1.0*iterationCount/maxEvals - 0.65)/0.35); // That involves a pow for each phase 3 iteration. We instead compute it // incrementally with 1 call to pow in the init, and 1 multiplication per // phase 3 update. // Also note that at the end of phase 2, targetRate will equal 0.44, where phase 3 begins. targetRate *= multPhase3; } else { // Phase 2 (50% of run beginning after phase 1): constant targetRate at 0.44. targetRate = 0.44; } if (acceptRate > targetRate) t *= 0.999; else t *= 1.001001001001001; // 1.001001001001001 == 1.0 / 0.999 } /* * package-private for unit testing */ double getTargetRate() { return targetRate; } /* * package-private for unit testing */ double getAcceptRate() { return acceptRate; } /* * package-private for unit testing */ double getTemperature() { return t; } }





© 2015 - 2025 Weber Informatics LLC | Privacy Policy