The Story So Far... The central problem of this course: Smartness( X ) arg max X. Possibly with some constraints on X.

Similar documents
Local Search & Optimization

Local Search & Optimization

A.I.: Beyond Classical Search

Local Search and Optimization

CS 331: Artificial Intelligence Local Search 1. Tough real-world problems

LOCAL SEARCH. Today. Reading AIMA Chapter , Goals Local search algorithms. Introduce adversarial search 1/31/14

Local Search (Greedy Descent): Maintain an assignment of a value to each variable. Repeat:

22c:145 Artificial Intelligence

Local and Online search algorithms

School of EECS Washington State University. Artificial Intelligence

Chapter 4 Beyond Classical Search 4.1 Local search algorithms and optimization problems

CS 380: ARTIFICIAL INTELLIGENCE

Scaling Up. So far, we have considered methods that systematically explore the full search space, possibly using principled pruning (A* etc.).

Incremental Stochastic Gradient Descent

Pengju

Local search algorithms. Chapter 4, Sections 3 4 1

CSC 4510 Machine Learning

PROBLEM SOLVING AND SEARCH IN ARTIFICIAL INTELLIGENCE

Local Beam Search. CS 331: Artificial Intelligence Local Search II. Local Beam Search Example. Local Beam Search Example. Local Beam Search Example

Local search algorithms

Evolutionary computation

Genetic Algorithms & Modeling

Methods for finding optimal configurations

Local and Stochastic Search

Evolutionary Computation. DEIS-Cesena Alma Mater Studiorum Università di Bologna Cesena (Italia)

Introduction to Optimization

Computational statistics

Beyond Classical Search

Local search algorithms. Chapter 4, Sections 3 4 1

Summary. AIMA sections 4.3,4.4. Hill-climbing Simulated annealing Genetic algorithms (briey) Local search in continuous spaces (very briey)

Stochastic Search: Part 2. Genetic Algorithms. Vincent A. Cicirello. Robotics Institute. Carnegie Mellon University

Optimization. Announcements. Last Time. Searching: So Far. Complete Searching. Optimization

Optimization Methods via Simulation

Lecture 22. Introduction to Genetic Algorithms

Local search and agents

Lecture 9 Evolutionary Computation: Genetic algorithms

Contents. 1 Introduction. 1.1 History of Optimization ALG-ML SEMINAR LISSA: LINEAR TIME SECOND-ORDER STOCHASTIC ALGORITHM FEBRUARY 23, 2016

Support Vector Machines: Training with Stochastic Gradient Descent. Machine Learning Fall 2017

Evolutionary Computation

CSC242: Artificial Intelligence. Lecture 4 Local Search

Artificial Intelligence (AI) Common AI Methods. Training. Signals to Perceptrons. Artificial Neural Networks (ANN) Artificial Intelligence

CMA-ES a Stochastic Second-Order Method for Function-Value Free Numerical Optimization

Introduction to Optimization

CS60021: Scalable Data Mining. Large Scale Machine Learning

Finding optimal configurations ( combinatorial optimization)

Introduction to Reinforcement Learning. CMPT 882 Mar. 18

New rubric: AI in the news

Experiments with. Cascading Design. Ben Kovitz Fluid Analogies Research Group Indiana University START

CSC321 Lecture 8: Optimization

Introduction to Simulated Annealing 22c:145

Machine Learning Support Vector Machines. Prof. Matteo Matteucci

Fundamentals of Genetic Algorithms

Genetic Algorithms. Donald Richards Penn State University

Artificial Intelligence Heuristic Search Methods

Search. Search is a key component of intelligent problem solving. Get closer to the goal if time is not enough

Data Warehousing & Data Mining

Lecture 5: Logistic Regression. Neural Networks

CSC321 Lecture 7: Optimization

A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring

Gecco 2007 Tutorial / Grammatical Evolution

Neural networks and optimization

CSC242: Intro to AI. Lecture 5. Tuesday, February 26, 13

Artificial Intelligence

CSE 417T: Introduction to Machine Learning. Lecture 11: Review. Henry Chai 10/02/18

Data Mining Part 5. Prediction

Artificial Intelligence Methods (G5BAIM) - Examination

Lecture 5: Linear models for classification. Logistic regression. Gradient Descent. Second-order methods.

Iterative improvement algorithms. Beyond Classical Search. Outline. Example: Traveling Salesperson Problem

Overview. Optimization. Easy optimization problems. Monte Carlo for Optimization. 1. Survey MC ideas for optimization: (a) Multistart

how should the GA proceed?

Machine Learning for NLP

ECS171: Machine Learning

Grundlagen der Künstlichen Intelligenz

Balancing and Control of a Freely-Swinging Pendulum Using a Model-Free Reinforcement Learning Algorithm

Least Mean Squares Regression. Machine Learning Fall 2018

Lecture 9: Large Margin Classifiers. Linear Support Vector Machines

CSC321 Lecture 4: Learning a Classifier

Genetic Algorithms and Genetic Programming Lecture 17

Least Mean Squares Regression

Evolutionary Design I

Gaussian and Linear Discriminant Analysis; Multiclass Classification

Lecture 26: Neural Nets

11 More Regression; Newton s Method; ROC Curves

April 9, Depto. de Ing. de Sistemas e Industrial Universidad Nacional de Colombia, Bogotá. Linear Classification Models. Fabio A. González Ph.D.

STA141C: Big Data & High Performance Statistical Computing

Neural Networks: Backpropagation

Heuristic Optimisation

Mixture Models & EM. Nicholas Ruozzi University of Texas at Dallas. based on the slides of Vibhav Gogate

CSC321 Lecture 4: Learning a Classifier

Empirical Risk Minimization

Machine Learning. Support Vector Machines. Fabio Vandin November 20, 2017

Neural Networks. Nicholas Ruozzi University of Texas at Dallas

Optimization and Complexity

Artificial Intelligence Methods (G5BAIM) - Examination

Reinforcement Learning

Mixture Models & EM. Nicholas Ruozzi University of Texas at Dallas. based on the slides of Vibhav Gogate

18.6 Regression and Classification with Linear Models

Computational Intelligence Winter Term 2017/18

Warm up: risk prediction with logistic regression

Repeated Occurrences of the Baldwin Effect Can Guide Evolution on Rugged Fitness Landscapes

Transcription:

Heuristic Search

The Story So Far... The central problem of this course: arg max X Smartness( X ) Possibly with some constraints on X. (Alternatively: arg min Stupidness(X ) ) X

Properties of Smartness(X) Possible categories for arg max Smartness( X ), in X decreasing order of desirability: Efficient closed-form solution. Linear Regression etc.

Properties of Smartness(X) Possible categories for arg max Smartness( X ), in X decreasing order of desirability: Efficient closed-form solution. Linear Regression etc. Differentiable and convex. Logistic Regression Quadratic programming (SVMs) etc.

Properties of Smartness(X) Possible categories for arg max Smartness( X ), in X decreasing order of desirability: Efficient closed-form solution. Linear Regression etc. Differentiable and convex. Logistic Regression Quadratic programming (SVMs) etc. Differentiable and non-convex. Multi-layer Perceptrons etc.

Properties of Smartness(X) Possible categories for arg max Smartness( X ), in X decreasing order of desirability: Efficient closed-form solution. Linear Regression etc. Differentiable and convex. Logistic Regression Quadratic programming (SVMs) etc. Differentiable and non-convex. Multi-layer Perceptrons etc. Non differentiable??

Example 1 Constraint Satisfaction Problems Problems like N-Queens: place N queens on an NxN chess board so that no two are threatening. There is no derivative to follow. We can evaluate how close an assignment is to a solution The search space may be too large to search using the exhaustive algorithms we saw earlier.

Example 2 Policy Search Two approaches to reinforcement learning: Value Estimation Learn to predict the long term reward associated with states. (We ve been working on this.) Policy Search Learn a policy directly, without trying to predict reward. For example: a neural network where the input is the state information and the output is an action choice. (It is possible to estimate gradients, but expensive.)

Hill Climbing Assume that our search problem allows us to take one of a fixed number of moves. Each move transforms the current state to a successor state. States can be evaluated by an objective function. Hill Climbing Algorithm: Start in an arbitrary state. Choose the move that results in the best successor state. Repeat until converged.

Properties of Hill Climbing Easy to program. Often finds good solutions quickly. Very susceptible to local maxima. Can be inefficient if many moves are available. Lots of variations: Stochastic hill climbing randomly choose an uphill move. Random restart hill climbing redo search until satisfied with result. Simulated Annealing...

Simulated Annealing Avoiding local maxima requires us to take some steps downhill. Simulated Annealing attempts to find the right tradeoff between uphill and downhill moves: Start in a random state ** Select a random move * If the move results in improvement, keep it * If the move does not result in improvement, keep it with probability P SA. * Return to **.

Computing P SA E cur E move P SA =e T cur PSA decreases as E cur E move increases. I.e. The worse the move, the less likely we are to keep it. T cur PSA also decreases as (temperature) decreases. High temperature: random search. Low temperature: randomized hill climbing. If we decrease temperature infinitely slowly, we can guarantee that the global optimum will be found :)

Genetic Algorithm (Basic algorithm due to John Holland, 1975) First, contrive a mapping from bit strings to your problem space. Genotype -> Phenotype. Next, contrive an objective function for evaluating solutions: fitness function. F(phenotype) = fitness.

Genetic Algorithm Generate K bit strings randomly: a population of individuals. **Evaluate the fitness of each individual. Assign a probability to each individual proportional to it's fitness. Generate a new population: Select 2 individuals (parents) according to probability assigned above. Crossover: Pick a random bit position, and swap all bits after that position. Mutation: flip individual bits with a small independent probability. Repeat until we have K new individuals. Return to ** unless satisfied.

GA Illustration (Figure from Russel & Norvig, Artifical Intelligence a Modern Approach.)

Does it Work? How well it works has a lot to do with the structure of the fitness surface. Getting it to work well requires carefully engineering the genotype->phenotype mapping. MANY variations: Different methods for selecting individuals, rank ordered instead of proportional, keep the fittest, etc. etc... Co-evolving parasites evolving test cases. Independently evolving island populations.

More Recent/Practical Algorithms Covariance Matrix Adaptation Evolution Strategy (CMA-ES) Current population is selected from a multivariate normal distribution. Distribution is updated after the population is evaluated.

CMA-ES By Sentewolf (talk) (Uploads) Transferred from en.wikipedia to Commons., Public Domain, https://commons.wikimedia.org/w/index.php?curid=48100101