Dynamic Programming. Lecture 13 (5/31/2017)

Similar documents
Suggested solutions for the exam in SF2863 Systems Engineering. June 12,

CS 331 DESIGN AND ANALYSIS OF ALGORITHMS DYNAMIC PROGRAMMING. Dr. Daisy Tang

Computational issues surrounding the management of an ecological food web

Analysis of Discrete Time Queues (Section 4.6)

U.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017

Hidden Markov Models & The Multivariate Gaussian (10/26/04)

Winter 2008 CS567 Stochastic Linear/Integer Programming Guest Lecturer: Xu, Huan

Continuous Time Markov Chains

Outline and Reading. Dynamic Programming. Dynamic Programming revealed. Computing Fibonacci. The General Dynamic Programming Technique

Hidden Markov Models

6. Stochastic processes (2)

6. Stochastic processes (2)

Equilibrium with Complete Markets. Instructor: Dmytro Hryshko

Amiri s Supply Chain Model. System Engineering b Department of Mathematics and Statistics c Odette School of Business

EEL 6266 Power System Operation and Control. Chapter 3 Economic Dispatch Using Dynamic Programming

Google PageRank with Stochastic Matrix

Introduction to Algorithms

Simultaneous Optimization of Berth Allocation, Quay Crane Assignment and Quay Crane Scheduling Problems in Container Terminals

Introduction to Algorithms

8. Modelling Uncertainty

Lecture 14: Bandits with Budget Constraints

Design and Analysis of Algorithms

Hidden Markov Models

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Resource Allocation with a Budget Constraint for Computing Independent Tasks in the Cloud

Department of Computer Science Artificial Intelligence Research Laboratory. Iowa State University MACHINE LEARNING

CS : Algorithms and Uncertainty Lecture 17 Date: October 26, 2016

Intra-Domain Traffic Engineering

College of Computer & Information Science Fall 2009 Northeastern University 20 October 2009

A LINEAR PROGRAM TO COMPARE MULTIPLE GROSS CREDIT LOSS FORECASTS. Dr. Derald E. Wentzien, Wesley College, (302) ,

Markov decision processes

Limited Dependent Variables

Problem Set 9 - Solutions Due: April 27, 2005

A NOTE ON A PERIODIC REVIEW INVENTORY MODEL WITH UNCERTAIN DEMAND IN A RANDOM ENVIRONMENT. Hirotaka Matsumoto and Yoshio Tabata

Min Cut, Fast Cut, Polynomial Identities

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models

CS-433: Simulation and Modeling Modeling and Probability Review

CIE4801 Transportation and spatial modelling Trip distribution

Revenue comparison when the length of horizon is known in advance. The standard errors of all numbers are less than 0.1%.

COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS

k t+1 + c t A t k t, t=0

Hopfield networks and Boltzmann machines. Geoffrey Hinton et al. Presented by Tambet Matiisen

Random Partitions of Samples

Engineering Risk Benefit Analysis

Portfolios with Trading Constraints and Payout Restrictions

An Integrated Asset Allocation and Path Planning Method to to Search for a Moving Target in in a Dynamic Environment

Calculation of time complexity (3%)

The Minimum Universal Cost Flow in an Infeasible Flow Network

Annexes. EC.1. Cycle-base move illustration. EC.2. Problem Instances

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD

Lecture 21: Numerical methods for pricing American type derivatives

Continuous Time Markov Chain

ECE 534: Elements of Information Theory. Solutions to Midterm Exam (Spring 2006)

An Experiment/Some Intuition (Fall 2006): Lecture 18 The EM Algorithm heads coin 1 tails coin 2 Overview Maximum Likelihood Estimation

The Feynman path integral

Hidden Markov Models

Additional Codes using Finite Difference Method. 1 HJB Equation for Consumption-Saving Problem Without Uncertainty

Single-Facility Scheduling over Long Time Horizons by Logic-based Benders Decomposition

Absorbing Markov Chain Models to Determine Optimum Process Target Levels in Production Systems with Rework and Scrapping

PHYS 705: Classical Mechanics. Hamilton-Jacobi Equation

Topics in Probability Theory and Stochastic Processes Steven R. Dunbar. Classes of States and Stationary Distributions

Ryan (2009)- regulating a concentrated industry (cement) Firms play Cournot in the stage. Make lumpy investment decisions

Randomness and Computation

Module 9. Lecture 6. Duality in Assignment Problems

Other NN Models. Reinforcement learning (RL) Probabilistic neural networks

Course 395: Machine Learning - Lectures

Solutions HW #2. minimize. Ax = b. Give the dual problem, and make the implicit equality constraints explicit. Solution.

Lecture 3: Shannon s Theorem

The Geometry of Logit and Probit

INTRODUCTION TO MACHINE LEARNING 3RD EDITION

Complement of Type-2 Fuzzy Shortest Path Using Possibility Measure

ANSWERS CHAPTER 9. TIO 9.2: If the values are the same, the difference is 0, therefore the null hypothesis cannot be rejected.

MIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU

CS 3750 Machine Learning Lecture 6. Monte Carlo methods. CS 3750 Advanced Machine Learning. Markov chain Monte Carlo

Resource Allocation and Decision Analysis (ECON 8010) Spring 2014 Foundations of Regression Analysis

4DVAR, according to the name, is a four-dimensional variational method.

Linear Approximation with Regularization and Moving Least Squares

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore

PROBABILITY PRIMER. Exercise Solutions

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering /

COS 521: Advanced Algorithms Game Theory and Linear Programming

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

8.592J: Solutions for Assignment 7 Spring 2005

Text S1: Detailed proofs for The time scale of evolutionary innovation

Review: Fit a line to N data points

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Lecture 10 Support Vector Machines. Oct

Robust Adaptive Markov Decision Processes in Multivehicle

A Bayesian Approach to Arrival Rate Forecasting for Inhomogeneous Poisson Processes for Mobile Calls

An Optimization Model for Routing in Low Earth Orbit Satellite Constellations

VARIATION OF CONSTANT SUM CONSTRAINT FOR INTEGER MODEL WITH NON UNIFORM VARIABLES

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models

PROBLEM SET 7 GENERAL EQUILIBRIUM

Sampling Theory MODULE VII LECTURE - 23 VARYING PROBABILITY SAMPLING

FUZZY FINITE ELEMENT METHOD

Dynamic Programming. Preview. Dynamic Programming. Dynamic Programming. Dynamic Programming (Example: Fibonacci Sequence)

Lecture 10 Support Vector Machines II

Reinforcement learning

Credit Card Pricing and Impact of Adverse Selection

Transcription:

Dynamc Programmng Lecture 13 (5/31/2017)

- A Forest Thnnng Example - Projected yeld (m3/ha) at age 20 as functon of acton taken at age 10 Age 10 Begnnng Volume Resdual Ten-year Volume volume thnned volume growth at age 20 260 0 260 390 650 260 50 210 325 535 260 100 160 250 410 Projected yeld (m3/ha) at age 30 as functon of begnnng volume at age 20 and acton taken at age 20 Age 20 Begnnng Volume Resdual Ten-year Volume volume thnned volume growth at age 30 0 650 200 850 650 150 500 250 750 200 450 200 650 0 535 215 750 535 100 435 215 650 175 360 140 500 0 410 190 600 410 75 335 165 500 150 260 140 400 Age 10 1 260 0 Stage 1 50 100 Age 20 2 650 0 3 535 4 410 150 Stage 2 0 150 200 100 175 0 75 Age 30 5 850 6 750 7 650 8 600 9 500 850 750 650 600 500 Stage 3 Age 10 11 260 Source: Dykstra s Mathematcal Programmng for Natural Resource Management (1984) 10 400 400

Dynamc Programmng Cont. Stages, states and actons (decsons) Backward and forward recurson x decson varable: mmedate destnaton node at stage ; f ( s, x ) the maxmum total volume harvested durng all remanng stages, gven that the stand has reached the state correspondng to node s at stage and the mmedate destnaton node s x ; x the value of x that maxmzes f ( s, x ); and * f ( s) the maxmum value of f ( s, x ),.e., f ( s) f ( s, x ). * * *

Solvng Dynamc Programs Recursve relaton at stage (Bellman Equaton): f () s max H f ( x ), * * s, x 1 x where H s the volume harvested by thnnng at the begnnng of stage sx, n order to move the stand from state s at the begnnng of stage to state x at the end of stage.

Dynamc Programmng Structural requrements of DP The problem can be dvded nto stages Each stage has a fnte set of assocated states (dscrete state DP) The mpact of a polcy decson to transform a state n a gven stage to another state n the subsequent stage s determnstc Prncple of optmalty: gven the current state of the system, the optmal polcy for the remanng stages s ndependent of any pror polcy adopted

Examples of DP The Floyd-Warshall Algorthm (used n the Bucket formulaton of ARM): Let c (, j) denote the length (or weght) of the path between node and j; and Let s (, jk, ) denote the length (or total weght) of the shortest path between nodes and j gong through ntermedate node 1, 2,...,k. Then, the followng recurson wll gve the length of the shortest paths between all pars of nodes n a graph: s (, jk, ) c (, j) f k1 mn s (, jk, 1), skk (,, 1) sk (, jk, 1) otherwse

Examples of DP cont. Mnmzng the rsk of losng an endangered speces (non-lnear DP): Probablty of Project Falure Fundng level (state) Project (stage) $1M 1 50% 30% 20% 2 70% 50% 30% 3 80% 50% 40% $1M $1M $1M $1M $1M $1M Stage 1 Stage 2 Stage 3 Source: Buongorno and Glles (2003) Decson Methods for Forest Resource Management

Mnmzng the rsk of losng an endangered speces (DP example) Stages (t=1,2,3) represent $ allocaton decsons for Project 1, 2 and 3; States (=0,1,2) represent the budgets avalable at each stage t; Decsons (j=0,1,2) represent the budgets avalable at stage t+1; pt (, j) denotes the probablty of falure of Project t f decson j s made (.e., -j s spent on Project t); and V t* () s the smallest probablty of total falure from stage t onward, startng n state and makng the best decson j*. Then, the recursve relaton can be stated as: V () mn p (, j) V ( j) * * t t t1 j

Markov Chans (based on Buongorno and Glles 2003) Volume by stand states State Volume (m3/ha) L <400 M 400-700 H 700< 20-yr Transton probabltes w/o management Start State L End State j M H L 40% 60% 0% M 0% 30% 70% H 5% 5% 90% Transton probablty matrx P pt = probablty dstrbuton of stand states n perod t p p P t t t-1 Vector p t converges to a vector of steady-state probabltes p*. p* s ndependent of p 0!

Markov Chans cont. Berger-Parker Landscape Index BP t plt pmt pht max( p, p, p ) Lt Mt Ht,where s the probablty that the stand p t s n state L, M or H n perod t. Mean Resdence Tmes m D (1 p ),where D s the length of each perod, and p s the probablty that a stand n state n the begnnng of perod t stays there tll the end of perod t. Mean Recurrence Tmes m D,where s the steady-state probablty of state.

Markov Chans cont. Forest dynamcs (expected revenues or bodversty wth vs. w/o management: 20-yr Transton probabltes w/o management Start State L End State j M H L 40% 60% 0% M 0% 30% 70% H 5% 5% 90% 20-yr Transton probabltes w/ management Start State L End State j M H L 40% 60% 0% M 0% 30% 70% H 40% 60% 0% Expected long-term bodversty: Expected long-term perodc ncome: B B B B L L M M H H R R R R L L M M H H

Markov Chans cont. Present value of expected returns 1 V R p V p V p V (1 r) t, 1 20 L Lt M Mt H Ht where V s the present value of the expected return from t, a stand n state n state = L, M or H managed wth a specfc harvest polcy wth t perods to go before the end of the plannng horzon. R s the mmedate return from managng a stand n state wth the gven harvest polcy. p, p, p are the probabltes that a stand n state moves to L M H state L, M or H, respectvely.

Markov Decson Processes 20-yr Transton probabltes dependng on the harvest decson No Cut Cut Start End State j Start End State j State L M H State L M H L 40% 60% 0% L 40% 60% 0% M 0% 30% 70% M 40% 60% 0% H 5% 5% 90% H 40% 60% 0% 1 V max R p V p V p V (1 r) * * * *, t1 j 20 Lj Lt Mj Mt Hj Ht j where V s the hghest present value of the expected return from * t, 1 a stand n state n state = L, M or H managed wth a specfc harvest polcy wth t perods to go before the end of the plannng horzon. R s the mmedate return from j managng a stand n state wth harvest polcy j. p, p, p are the probabltes that a stand n state moves to Lj Mj Hj state L, M or H, respectvely f the harvest polcy s j.