6.231 Dynamic Programming Midterm, Fall Instructions
|
|
|
- Hillary Leonard
- 10 years ago
- Views:
Transcription
1 6.231 Dynamic Programming Midterm, Fall 2008 Instructions The midterm comprises three problems. Problem 1 is worth 60 points, problem 2 is worth 40 points, and problem 3 is worth 40 points. Your grade G 0, 100] is given by the following formula: G = 60 f min1, f 2 + f 3 ], where f i 0, 1], i = 1, 2, 3, is the fraction of problem i that you solved correctly. Notice that if you solve correctly both problems 2 and 3, but you do not solve problem 1, your grade will only be G = 40. Thus, try to solve problem 1 first, then choose between problems 2 and 3 whichever you prefer, and if you still have time, try to do the remaining problem. Credit is given only for clear logic and careful reasoning. GOOD LUCK! 1
2 Problem 1 Infinite Horizon Problem, 60 points An engineer has invented a better mouse trap and is interested in selling it for the right price. At the beginning of each period, he receives a sale offer that takes one of the values s 1,..., s n with corresponding probabilities p 1,..., p n, independently of prior offers. If he accepts the offer he retires from engineering. If he refuses the offer, he may accept subsequent offers but he also runs the risk that a competitor will invent an even better mouse trap, rendering his own unsaleable; this happens with probability β > 0 at each time period, independently of earlier time periods. While he is overtaken by the competitor, at each time period, he may choose to retire from engineering, or he may choose to invest an amount v 0, in which case he has a probability γ to improve his mouse trap, overtake his competitor, and start receiving offers as earlier. The problem is to determine the engineer s strategy to maximize his discounted expected payoff minus investment cost, assuming a discount factor α < 1. a 15 points] Formulate the problem as an infinite horizon discounted cost problem and write the corresponding Bellman s equation. b 15 points] Characterize as best as you can an optimal policy. c 10 points] Describe as best as you can how policy iteration would work for this problem. What kind of initial policy would lead to an efficient policy iteration algorithm? d 10 points] Assume that there is no discount factor. Does the problem make sense as an average cost per stage problem? e 10 points] Assume that there is no discount factor and that the investment cost v is equal to 0. Does the problem make sense as a stochastic shortest path problem, and what is then the optimal policy? Solution to Problem 1 a Let the set of possible offers be {x i i = 1,..., n} {t, r}, where being in a state has the following meaning: x i : receiving offer s i ; t: overtaken; r: retired. Define the control space by {Aaccept offer, Iinvest, Rretire, Rejreject}. The state r is absorbing, and for the other states, the set of admissible controls is Ux i = {A, Rej}, Ut = {I, R}. 2
3 Define the corresponding transition probabilities, and per-stage costs as described in the problem. Bellman s equation for α < 1 is n J x i = max s i, α 1 β p j J x j + βj t, 1 j=1 n J t = max 0, v + α γ p j J x j + 1 γj t. 2 j=1 b From Eq. 1, we see that since the second term in the minimization of the right-hand side does not depend on the current state x i, the optimal policy is a threshold policy at states x i, i = 1,..., n. A single threshold is needed since J s is clearly monotonically nondecreasing with s. From Eq. 2, we see that once the inventor has been overtaken, it is either optimal to retire immediately, or to keep investing until his mouse trap is improved. c Consider policy iteration. If the current policy µ is a threshold policy, the cost J µ s is monotonically nondecreasing with s, so the cost of the improved policy is also monotonically nondecreasing with s. Hence, starting with a nonoptimal threshold policy, the method will generate a new and improved policy for at least one offer. It follows that starting with a threshold policy, after at most n iterations, policy iteration will terminate with an optimal threshold policy. d Even without discounting, all policies have finite total cost, except the one that never sells and always invests, which incurs infinite total cost. Thus, we see that under the average cost criterion, all of the finite total cost policies will be optimal. Thus, the average cost criterion does not discriminate enough between different policies and makes no sense. e Since there is no discounting and v = 0, there is no penalty for waiting as long as necessary until the maximum possible offer is received, which is going to happen with probability 1. So a stochastic shortest path formulation makes limited sense, since it excludes from consideration all offers except the maximum possible, no matter how unlikely this maximum offer is. Problem 2 Imperfect State Information Problem, 40 points A machine tosses a coin N times, but each time may switch the probability of heads between two possible values h and h. The switch occurs with probability q at each time, independently of the switches in previous times. A gambler may bet on the outcome of each coin toss knowing h, h, and q, the outcomes of previous tosses, and the probability r that h is used at the initial toss. If the gambler bets on a coin toss, he wins $ 2 if a head occurs, while he loses $ 1 if a tail occurs. The gambler may also permanently stop betting prior to any coin toss. 3
4 a 20 points] Define r for k = 0, p k = P rob{coin with a probability of heads h is used at toss k zk 1,..., z 0 } for k = 1,..., N 1, where z k 1,..., z 0 are the preceding observed tosses. Write a equation for the evolution of p k as the coin tosses are observed. b 10 points] Use the equation of part a to write a DP algorithm to find the gambler s optimal policy for betting or stopping to bet prior to any coin toss. You will receive credit if you do this part correctly, assuming a generic functional form of this equation. c 10 points] Repeat part b for the variant of the problem where the gambler may resume play at any time after he stops betting, and he keeps observing the coin tosses also when he is not betting. Solution to Problem 2 a The machine uses two coins, one with a probability of heads equal to h, and the the other one with a probability of heads equal to h. We call the first coin a type h coin, while we call the second coin a type h coin. Let f h z, z {head, tail}, be the probability mass function for the type h coin clearly, f h head = h. Similarly, let f h z, z {head, tail}, be the probability mass function for the type h coin clearly, f hhead = h. We are faced with an imperfect state information problem where the state at stage k, that we call x k, is the type of coin that the machine uses at stage k. We write x k = h if at stage k the machine uses a type h coin, and similarly we write x k = h if at stage k the machine uses a type h coin. The conditional probability p k is generated recursively according to the following equation with p 0 = r: p k+1 = = P r{x k+1 = h z k, z k 1,..., z 0 } P r {x k+1 = h, z k z k 1,..., z 0 } = P r{z k z k 1,..., z 0 } P r{z k x k+1 = h, z k 1,..., z 0 } P r{x k+1 = h z k 1,..., z 0 } = P r{zk x k = h, z k 1,..., z 0 }P r{x k = h z k 1,..., z 0 } + P r{z k x k = h, z k 1,..., z 0 }P r {x k = h z k 1,..., z 0 } = P r{z k x k+1 = h, z k 1,..., z 0 } P r{x k+1 = h x k = h, z k 1,..., z 0 }p k + P r{x k+1 = h x k = h, z k 1,..., z 0 }1 p k f h z k p k + f z k 1 p k h 4
5 = 1 qp k q1 pk f h zk 1 q p 1 p + fh zk 1 qp k +q1 p k 1 q p k + q1 p k. = φp k, z k, k+q k f h z k p k + f h z k 1 p k ] since P r{z k x k+1 = h, z k 1,..., z 0 } = = P r{z k x k+1 = h, x k = h, z k 1,..., z0} P r{x k = h x k+1 = h, z k 1,..., z 0} + + P r{z k x k+1 = h, x k = h, z k 1,..., z 0 } P r{x k = h x k+1 = h, z k 1,..., z0} = P r{x k+1 = h x k = h, z k 1,..., z0}p k = f h z k + P r{x k+1 = h x k = h, z k 1,..., z 0 }p k + P r{x k+1 = h x k = h, z k 1,..., z 0} 1 p k P r{x k+1 = h x k = h, z k 1,..., z 0}1 p k + f h z k = P r{x k+1 = h x k = h, z k 1,..., z 0 }p k + P r{x k+1 = h 1 x k = h, z k 1,..., z 0 } p k 1 qpk q1 p k = f h z k + f h zk 1 q p k + q1 p k 1 qp k + q1 p k b The DP algorithm is J k p k = max }{{} 0, 2 p k h+1 p k h 1 p k 1 h+1 p k 1 h+e zk {J }] k+1 φk p k, z k, k = 0,..., N 1, with J N p N 0. stop c In this case we have for k = 0,..., N 1 J k p k = max 0 + E zk {J k+1 φk p k, z k } {, 2 p k h+1 p k h 1 p k 1 h+1 p k 1 h+e zk J k+1 φk p k, z k }], } {{ } with J N p N 0. Thus, we do not play if do not play 0 > 2 p k h + 1 p k h 1 p k 1 h + 1 p k 1 h, that is, assuming without loss of generality that h > h, we do not play if 1 3h p k <. 3h h Problem 3 Optimal Stopping Problem, 40 points 5
6 A driver is looking for parking on the way to his destination. Each parking place is free with probability p independently of whether other parking places are free or not. The driver cannot observe whether a parking place is free until he reaches it. If he parks k places from his destination, he incurs a cost k. If he reaches the destination without having parked the cost is C. The driver wishes to find an optimal parking policy. a 20 points] Formulate the problem as a DP problem, clearly identifying, states, controls, transition probabilities and costs. b 20 points] Write the DP algorithm and show that it can be represented by the sequence {C k } generated as follows C k = p mink, C k 1 ] + qc k, k = 1, 2,..., 1 where C 0 = C and q = 1 p. Characterize as best as you can the optimal policy. Solution to Problem 3 a Let the state xk {T, T } where T represents the driver having parked before reaching the kth spot. Let the control at each parking spot uk {P, P } where P represents the choice to park in the kth spot. Let the disturbance w k equal 1 if the kth spot is free; otherwise it equals 0. Clearly, we have the control constraint that u k = P, if x k = T or w k = 0. The cost associated with parking in the kth spot is: gkt, P, 1 = k If the driver has not parked upon reaching his destination, he incurs a cost g N T = C. All other costs are zero. The system evolves according to: { T, if xk = T or uk = P x k+1 = T, otherwise b Once the driver has parked, his remaining cost is zero. Thus, we can define C k to be the expected remaining cost, given that the driver has not parked before the kth spot. Note that this is simply J k T. The DP algorithm starts with C 0 = C, and for k = 1, 2,..., generates C k by C k = min E { g k T, u k, w k + J k 1 x k 1 } = min p k + J k 1 T ] ] + qj 1 T k, u k {P,P } w k } } } {{ } } J k 1 {{ T {{ park, free Since J k T = 0 for all k, we have ] ] C k = min pk + qc k 1, C k 1 = p min k, C k 1 + qc k 1. Thus the sequence {C k } is the sequence generated by DP. 6 park, not free don t par} k
7 MIT OpenCourseWare Dynamic Programming and Stochastic Control Fall 2011 For information about citing these materials or our Terms of Use, visit:
6.231 Dynamic Programming and Stochastic Control Fall 2008
MIT OpenCourseWare http://ocw.mit.edu 6.231 Dynamic Programming and Stochastic Control Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 6.231
Reinforcement Learning
Reinforcement Learning LU 2 - Markov Decision Problems and Dynamic Programming Dr. Martin Lauer AG Maschinelles Lernen und Natürlichsprachliche Systeme Albert-Ludwigs-Universität Freiburg [email protected]
Reinforcement Learning
Reinforcement Learning LU 2 - Markov Decision Problems and Dynamic Programming Dr. Joschka Bödecker AG Maschinelles Lernen und Natürlichsprachliche Systeme Albert-Ludwigs-Universität Freiburg [email protected]
6.042/18.062J Mathematics for Computer Science. Expected Value I
6.42/8.62J Mathematics for Computer Science Srini Devadas and Eric Lehman May 3, 25 Lecture otes Expected Value I The expectation or expected value of a random variable is a single number that tells you
Economics 1011a: Intermediate Microeconomics
Lecture 11: Choice Under Uncertainty Economics 1011a: Intermediate Microeconomics Lecture 11: Choice Under Uncertainty Tuesday, October 21, 2008 Last class we wrapped up consumption over time. Today we
Optimization under uncertainty: modeling and solution methods
Optimization under uncertainty: modeling and solution methods Paolo Brandimarte Dipartimento di Scienze Matematiche Politecnico di Torino e-mail: [email protected] URL: http://staff.polito.it/paolo.brandimarte
Lecture 25: Money Management Steven Skiena. http://www.cs.sunysb.edu/ skiena
Lecture 25: Money Management Steven Skiena Department of Computer Science State University of New York Stony Brook, NY 11794 4400 http://www.cs.sunysb.edu/ skiena Money Management Techniques The trading
.4 120 +.1 80 +.5 100 = 48 + 8 + 50 = 106.
Chapter 16. Risk and Uncertainty Part A 2009, Kwan Choi Expected Value X i = outcome i, p i = probability of X i EV = pix For instance, suppose a person has an idle fund, $100, for one month, and is considering
Principles of demand management Airline yield management Determining the booking limits. » A simple problem» Stochastic gradients for general problems
Demand Management Principles of demand management Airline yield management Determining the booking limits» A simple problem» Stochastic gradients for general problems Principles of demand management Issues:»
Real Time Scheduling Basic Concepts. Radek Pelánek
Real Time Scheduling Basic Concepts Radek Pelánek Basic Elements Model of RT System abstraction focus only on timing constraints idealization (e.g., zero switching time) Basic Elements Basic Notions task
Part I. Gambling and Information Theory. Information Theory and Networks. Section 1. Horse Racing. Lecture 16: Gambling and Information Theory
and Networks Lecture 16: Gambling and Paul Tune http://www.maths.adelaide.edu.au/matthew.roughan/ Lecture_notes/InformationTheory/ Part I Gambling and School of Mathematical
15.407 Sample Mid-Term Examination Fall 2008. Some Useful Formulas
15.407 Sample Mid-Term Examination Fall 2008 Please check to be certain that your copy of this examination contains 18 pages (including this one). Write your name and MIT ID number on every page. You are
Classification - Examples
Lecture 2 Scheduling 1 Classification - Examples 1 r j C max given: n jobs with processing times p 1,...,p n and release dates r 1,...,r n jobs have to be scheduled without preemption on one machine taking
LOOKING FOR A GOOD TIME TO BET
LOOKING FOR A GOOD TIME TO BET LAURENT SERLET Abstract. Suppose that the cards of a well shuffled deck of cards are turned up one after another. At any time-but once only- you may bet that the next card
Improve Net Present Value using cash flow weight
2011 2 nd International Conference on Construction and Project Management IPEDR vol.15 (2011) (2011) IACSIT Press, Singapore Improve Net Present Value using cash flow weight Vacharee Tantisuvanichkul 1
The Discrete Binomial Model for Option Pricing
The Discrete Binomial Model for Option Pricing Rebecca Stockbridge Program in Applied Mathematics University of Arizona May 4, 2008 Abstract This paper introduces the notion of option pricing in the context
Kelly s Criterion for Option Investment. Ron Shonkwiler ([email protected])
Kelly s Criterion for Option Investment Ron Shonkwiler ([email protected]) 1 1 Kelly s Criterion for Option Investment Ron Shonkwiler ([email protected]) Outline: Review Kelly s Problem
4.2 Description of the Event operation Network (EON)
Integrated support system for planning and scheduling... 2003/4/24 page 39 #65 Chapter 4 The EON model 4. Overview The present thesis is focused in the development of a generic scheduling framework applicable
Finite Horizon Investment Risk Management
History Collaborative research with Ralph Vince, LSP Partners, LLC, and Marcos Lopez de Prado, Guggenheim Partners. Qiji Zhu Western Michigan University Workshop on Optimization, Nonlinear Analysis, Randomness
5. Factoring by the QF method
5. Factoring by the QF method 5.0 Preliminaries 5.1 The QF view of factorability 5.2 Illustration of the QF view of factorability 5.3 The QF approach to factorization 5.4 Alternative factorization by the
Lecture notes: single-agent dynamics 1
Lecture notes: single-agent dynamics 1 Single-agent dynamic optimization models In these lecture notes we consider specification and estimation of dynamic optimization models. Focus on single-agent models.
Computing Near Optimal Strategies for Stochastic Investment Planning Problems
Computing Near Optimal Strategies for Stochastic Investment Planning Problems Milos Hauskrecfat 1, Gopal Pandurangan 1,2 and Eli Upfal 1,2 Computer Science Department, Box 1910 Brown University Providence,
1 Gambler s Ruin Problem
Coyright c 2009 by Karl Sigman 1 Gambler s Ruin Problem Let N 2 be an integer and let 1 i N 1. Consider a gambler who starts with an initial fortune of $i and then on each successive gamble either wins
Decision Theory. 36.1 Rational prospecting
36 Decision Theory Decision theory is trivial, apart from computational details (just like playing chess!). You have a choice of various actions, a. The world may be in one of many states x; which one
How To Price A Call Option
Now by Itô s formula But Mu f and u g in Ū. Hence τ θ u(x) =E( Mu(X) ds + u(x(τ θ))) 0 τ θ u(x) E( f(x) ds + g(x(τ θ))) = J x (θ). 0 But since u(x) =J x (θ ), we consequently have u(x) =J x (θ ) = min
The Kelly criterion for spread bets
IMA Journal of Applied Mathematics 2007 72,43 51 doi:10.1093/imamat/hxl027 Advance Access publication on December 5, 2006 The Kelly criterion for spread bets S. J. CHAPMAN Oxford Centre for Industrial
Chapter 4 Lecture Notes
Chapter 4 Lecture Notes Random Variables October 27, 2015 1 Section 4.1 Random Variables A random variable is typically a real-valued function defined on the sample space of some experiment. For instance,
Section 7C: The Law of Large Numbers
Section 7C: The Law of Large Numbers Example. You flip a coin 00 times. Suppose the coin is fair. How many times would you expect to get heads? tails? One would expect a fair coin to come up heads half
Options pricing in discrete systems
UNIVERZA V LJUBLJANI, FAKULTETA ZA MATEMATIKO IN FIZIKO Options pricing in discrete systems Seminar II Mentor: prof. Dr. Mihael Perman Author: Gorazd Gotovac //2008 Abstract This paper is a basic introduction
Periodic Capacity Management under a Lead Time Performance Constraint
Periodic Capacity Management under a Lead Time Performance Constraint N.C. Buyukkaramikli 1,2 J.W.M. Bertrand 1 H.P.G. van Ooijen 1 1- TU/e IE&IS 2- EURANDOM INTRODUCTION Using Lead time to attract customers
1 Interest rates, and risk-free investments
Interest rates, and risk-free investments Copyright c 2005 by Karl Sigman. Interest and compounded interest Suppose that you place x 0 ($) in an account that offers a fixed (never to change over time)
1/3 1/3 1/3 0.4 0.4 0.4 0.4 0.4 0.4 0.4 0 1 2 3 4 5 6 7 8 0.6 0.6 0.6 0.6 0.6 0.6 0.6
HOMEWORK 4: SOLUTIONS. 2. A Markov chain with state space {, 2, 3} has transition probability matrix /3 /3 /3 P = 0 /2 /2 0 0 Show that state 3 is absorbing and, starting from state, find the expected
arxiv:1112.0829v1 [math.pr] 5 Dec 2011
How Not to Win a Million Dollars: A Counterexample to a Conjecture of L. Breiman Thomas P. Hayes arxiv:1112.0829v1 [math.pr] 5 Dec 2011 Abstract Consider a gambling game in which we are allowed to repeatedly
Stochastic Processes and Advanced Mathematical Finance. Duration of the Gambler s Ruin
Steven R. Dunbar Department of Mathematics 203 Avery Hall University of Nebraska-Lincoln Lincoln, NE 68588-0130 http://www.math.unl.edu Voice: 402-472-3731 Fax: 402-472-8466 Stochastic Processes and Advanced
פרויקט מסכם לתואר בוגר במדעים )B.Sc( במתמטיקה שימושית
המחלקה למתמטיקה Department of Mathematics פרויקט מסכם לתואר בוגר במדעים )B.Sc( במתמטיקה שימושית הימורים אופטימליים ע"י שימוש בקריטריון קלי אלון תושיה Optimal betting using the Kelly Criterion Alon Tushia
A Profit-Maximizing Production Lot sizing Decision Model with Stochastic Demand
A Profit-Maximizing Production Lot sizing Decision Model with Stochastic Demand Kizito Paul Mubiru Department of Mechanical and Production Engineering Kyambogo University, Uganda Abstract - Demand uncertainty
How to Gamble If You Must
How to Gamble If You Must Kyle Siegrist Department of Mathematical Sciences University of Alabama in Huntsville Abstract In red and black, a player bets, at even stakes, on a sequence of independent games
V. RANDOM VARIABLES, PROBABILITY DISTRIBUTIONS, EXPECTED VALUE
V. RANDOM VARIABLES, PROBABILITY DISTRIBUTIONS, EXPETED VALUE A game of chance featured at an amusement park is played as follows: You pay $ to play. A penny and a nickel are flipped. You win $ if either
TD(0) Leads to Better Policies than Approximate Value Iteration
TD(0) Leads to Better Policies than Approximate Value Iteration Benjamin Van Roy Management Science and Engineering and Electrical Engineering Stanford University Stanford, CA 94305 [email protected] Abstract
Phantom bonuses. November 22, 2004
Phantom bonuses November 22, 2004 1 Introduction The game is defined by a list of payouts u 1, u 2,..., u l, and a list of probabilities p 1, p 2,..., p l, p i = 1. We allow u i to be rational numbers,
Betting with the Kelly Criterion
Betting with the Kelly Criterion Jane June 2, 2010 Contents 1 Introduction 2 2 Kelly Criterion 2 3 The Stock Market 3 4 Simulations 5 5 Conclusion 8 1 Page 2 of 9 1 Introduction Gambling in all forms,
MONT 107N Understanding Randomness Solutions For Final Examination May 11, 2010
MONT 07N Understanding Randomness Solutions For Final Examination May, 00 Short Answer (a) (0) How are the EV and SE for the sum of n draws with replacement from a box computed? Solution: The EV is n times
A New Interpretation of Information Rate
A New Interpretation of Information Rate reproduced with permission of AT&T By J. L. Kelly, jr. (Manuscript received March 2, 956) If the input symbols to a communication channel represent the outcomes
Betting systems: how not to lose your money gambling
Betting systems: how not to lose your money gambling G. Berkolaiko Department of Mathematics Texas A&M University 28 April 2007 / Mini Fair, Math Awareness Month 2007 Gambling and Games of Chance Simple
Expected Value and the Game of Craps
Expected Value and the Game of Craps Blake Thornton Craps is a gambling game found in most casinos based on rolling two six sided dice. Most players who walk into a casino and try to play craps for the
Linear Programming Notes V Problem Transformations
Linear Programming Notes V Problem Transformations 1 Introduction Any linear programming problem can be rewritten in either of two standard forms. In the first form, the objective is to maximize, the material
Dynamic Programming 11.1 AN ELEMENTARY EXAMPLE
Dynamic Programming Dynamic programming is an optimization approach that transforms a complex problem into a sequence of simpler problems; its essential characteristic is the multistage nature of the optimization
Risk Management for IT Security: When Theory Meets Practice
Risk Management for IT Security: When Theory Meets Practice Anil Kumar Chorppath Technical University of Munich Munich, Germany Email: [email protected] Tansu Alpcan The University of Melbourne Melbourne,
Optimal f :Calculating the expected growth-optimal fraction for discretely-distributed outcomes
Optimal f :Calculating the expected growth-optimal fraction for discretely-distributed outcomes Ralph Vince November 12, 2015 First version: January, 2012 The president of LSP Partners, LLC, 35185 Holbrook
Option Properties. Liuren Wu. Zicklin School of Business, Baruch College. Options Markets. (Hull chapter: 9)
Option Properties Liuren Wu Zicklin School of Business, Baruch College Options Markets (Hull chapter: 9) Liuren Wu (Baruch) Option Properties Options Markets 1 / 17 Notation c: European call option price.
24. The Branch and Bound Method
24. The Branch and Bound Method It has serious practical consequences if it is known that a combinatorial problem is NP-complete. Then one can conclude according to the present state of science that no
OPTIMAL DESIGN OF A MULTITIER REWARD SCHEME. Amir Gandomi *, Saeed Zolfaghari **
OPTIMAL DESIGN OF A MULTITIER REWARD SCHEME Amir Gandomi *, Saeed Zolfaghari ** Department of Mechanical and Industrial Engineering, Ryerson University, Toronto, Ontario * Tel.: + 46 979 5000x7702, Email:
Stochastic Models for Inventory Management at Service Facilities
Stochastic Models for Inventory Management at Service Facilities O. Berman, E. Kim Presented by F. Zoghalchi University of Toronto Rotman School of Management Dec, 2012 Agenda 1 Problem description Deterministic
On Adaboost and Optimal Betting Strategies
On Adaboost and Optimal Betting Strategies Pasquale Malacaria School of Electronic Engineering and Computer Science Queen Mary, University of London Email: [email protected] Fabrizio Smeraldi School of
Review Horse Race Gambling and Side Information Dependent horse races and the entropy rate. Gambling. Besma Smida. ES250: Lecture 9.
Gambling Besma Smida ES250: Lecture 9 Fall 2008-09 B. Smida (ES250) Gambling Fall 2008-09 1 / 23 Today s outline Review of Huffman Code and Arithmetic Coding Horse Race Gambling and Side Information Dependent
Alok Gupta. Dmitry Zhdanov
RESEARCH ARTICLE GROWTH AND SUSTAINABILITY OF MANAGED SECURITY SERVICES NETWORKS: AN ECONOMIC PERSPECTIVE Alok Gupta Department of Information and Decision Sciences, Carlson School of Management, University
Formulations of Model Predictive Control. Dipartimento di Elettronica e Informazione
Formulations of Model Predictive Control Riccardo Scattolini Riccardo Scattolini Dipartimento di Elettronica e Informazione Impulse and step response models 2 At the beginning of the 80, the early formulations
Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay
Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 17 Shannon-Fano-Elias Coding and Introduction to Arithmetic Coding
36 Odds, Expected Value, and Conditional Probability
36 Odds, Expected Value, and Conditional Probability What s the difference between probabilities and odds? To answer this question, let s consider a game that involves rolling a die. If one gets the face
Elementary Statistics and Inference. Elementary Statistics and Inference. 17 Expected Value and Standard Error. 22S:025 or 7P:025.
Elementary Statistics and Inference S:05 or 7P:05 Lecture Elementary Statistics and Inference S:05 or 7P:05 Chapter 7 A. The Expected Value In a chance process (probability experiment) the outcomes of
Statistics 100A Homework 3 Solutions
Chapter Statistics 00A Homework Solutions Ryan Rosario. Two balls are chosen randomly from an urn containing 8 white, black, and orange balls. Suppose that we win $ for each black ball selected and we
Integrated support system for planning and scheduling... 2003/4/24 page 75 #101. Chapter 5 Sequencing and assignment Strategies
Integrated support system for planning and scheduling... 2003/4/24 page 75 #101 Chapter 5 Sequencing and assignment Strategies 5.1 Overview This chapter is dedicated to the methodologies used in this work
LECTURE 15: AMERICAN OPTIONS
LECTURE 15: AMERICAN OPTIONS 1. Introduction All of the options that we have considered thus far have been of the European variety: exercise is permitted only at the termination of the contract. These
When Machine Learning Meets AI and Game Theory
1 When Machine Learning Meets AI and Game Theory Anurag Agrawal, Deepak Jaiswal Abstract We study the problem of development of intelligent machine learning applications to exploit the problems of adaptation
I. Readings and Suggested Practice Problems. II. Risks Associated with Default-Free Bonds
Prof. Alex Shapiro Lecture Notes 13 Bond Portfolio Management I. Readings and Suggested Practice Problems II. Risks Associated with Default-Free Bonds III. Duration: Details and Examples IV. Immunization
TRAFFIC ENGINEERING OF DISTRIBUTED CALL CENTERS: NOT AS STRAIGHT FORWARD AS IT MAY SEEM. M. J. Fischer D. A. Garbin A. Gharakhanian D. M.
TRAFFIC ENGINEERING OF DISTRIBUTED CALL CENTERS: NOT AS STRAIGHT FORWARD AS IT MAY SEEM M. J. Fischer D. A. Garbin A. Gharakhanian D. M. Masi January 1999 Mitretek Systems 7525 Colshire Drive McLean, VA
Week_11: Network Models
Week_11: Network Models 1 1.Introduction The network models include the traditional applications of finding the most efficient way to link a number of locations directly or indirectly, finding the shortest
A hierarchical multicriteria routing model with traffic splitting for MPLS networks
A hierarchical multicriteria routing model with traffic splitting for MPLS networks João Clímaco, José Craveirinha, Marta Pascoal jclimaco@inesccpt, jcrav@deecucpt, marta@matucpt University of Coimbra
Wald s Identity. by Jeffery Hein. Dartmouth College, Math 100
Wald s Identity by Jeffery Hein Dartmouth College, Math 100 1. Introduction Given random variables X 1, X 2, X 3,... with common finite mean and a stopping rule τ which may depend upon the given sequence,
We consider the optimal production and inventory control of an assemble-to-order system with m components,
MANAGEMENT SCIENCE Vol. 52, No. 12, December 2006, pp. 1896 1912 issn 0025-1909 eissn 1526-5501 06 5212 1896 informs doi 10.1287/mnsc.1060.0588 2006 INFORMS Production and Inventory Control of a Single
Load Balancing and Switch Scheduling
EE384Y Project Final Report Load Balancing and Switch Scheduling Xiangheng Liu Department of Electrical Engineering Stanford University, Stanford CA 94305 Email: [email protected] Abstract Load
2004 Networks UK Publishers. Reprinted with permission.
Riikka Susitaival and Samuli Aalto. Adaptive load balancing with OSPF. In Proceedings of the Second International Working Conference on Performance Modelling and Evaluation of Heterogeneous Networks (HET
A Simple Model of Price Dispersion *
Federal Reserve Bank of Dallas Globalization and Monetary Policy Institute Working Paper No. 112 http://www.dallasfed.org/assets/documents/institute/wpapers/2012/0112.pdf A Simple Model of Price Dispersion
KEELE UNIVERSITY MID-TERM TEST, 2007 BA BUSINESS ECONOMICS BA FINANCE AND ECONOMICS BA MANAGEMENT SCIENCE ECO 20015 MANAGERIAL ECONOMICS II
KEELE UNIVERSITY MID-TERM TEST, 2007 Thursday 22nd NOVEMBER, 12.05-12.55 BA BUSINESS ECONOMICS BA FINANCE AND ECONOMICS BA MANAGEMENT SCIENCE ECO 20015 MANAGERIAL ECONOMICS II Candidates should attempt
Algorithms for optimal allocation of bets on many simultaneous events
Appl. Statist. (2007) 56, Part 5, pp. 607 623 Algorithms for optimal allocation of bets on many simultaneous events Chris Whitrow Imperial College London, UK [Received September 2006. Revised June 2007]
Scheduling Algorithms for Downlink Services in Wireless Networks: A Markov Decision Process Approach
Scheduling Algorithms for Downlink Services in Wireless Networks: A Markov Decision Process Approach William A. Massey ORFE Department Engineering Quadrangle, Princeton University Princeton, NJ 08544 K.
Lecture 6: Portfolios with Stock Options Steven Skiena. http://www.cs.sunysb.edu/ skiena
Lecture 6: Portfolios with Stock Options Steven Skiena Department of Computer Science State University of New York Stony Brook, NY 11794 4400 http://www.cs.sunysb.edu/ skiena Portfolios with Options The
The overall size of these chance errors is measured by their RMS HALF THE NUMBER OF TOSSES NUMBER OF HEADS MINUS 0 400 800 1200 1600 NUMBER OF TOSSES
INTRODUCTION TO CHANCE VARIABILITY WHAT DOES THE LAW OF AVERAGES SAY? 4 coins were tossed 1600 times each, and the chance error number of heads half the number of tosses was plotted against the number
OPTIMAL CONTROL OF A COMMERCIAL LOAN REPAYMENT PLAN. E.V. Grigorieva. E.N. Khailov
DISCRETE AND CONTINUOUS Website: http://aimsciences.org DYNAMICAL SYSTEMS Supplement Volume 2005 pp. 345 354 OPTIMAL CONTROL OF A COMMERCIAL LOAN REPAYMENT PLAN E.V. Grigorieva Department of Mathematics
Linear Risk Management and Optimal Selection of a Limited Number
How to build a probability-free casino Adam Chalcraft CCR La Jolla [email protected] Chris Freiling Cal State San Bernardino [email protected] Randall Dougherty CCR La Jolla [email protected] Jason
Dynamic programming. Doctoral course Optimization on graphs - Lecture 4.1. Giovanni Righini. January 17 th, 2013
Dynamic programming Doctoral course Optimization on graphs - Lecture.1 Giovanni Righini January 1 th, 201 Implicit enumeration Combinatorial optimization problems are in general NP-hard and we usually
ALGORITHMIC TRADING WITH MARKOV CHAINS
June 16, 2010 ALGORITHMIC TRADING WITH MARKOV CHAINS HENRIK HULT AND JONAS KIESSLING Abstract. An order book consists of a list of all buy and sell offers, represented by price and quantity, available
STRATEGIC CAPACITY PLANNING USING STOCK CONTROL MODEL
Session 6. Applications of Mathematical Methods to Logistics and Business Proceedings of the 9th International Conference Reliability and Statistics in Transportation and Communication (RelStat 09), 21
Current Accounts in Open Economies Obstfeld and Rogoff, Chapter 2
Current Accounts in Open Economies Obstfeld and Rogoff, Chapter 2 1 Consumption with many periods 1.1 Finite horizon of T Optimization problem maximize U t = u (c t ) + β (c t+1 ) + β 2 u (c t+2 ) +...
Motivated by a problem faced by a large manufacturer of a consumer product, we
A Coordinated Production Planning Model with Capacity Expansion and Inventory Management Sampath Rajagopalan Jayashankar M. Swaminathan Marshall School of Business, University of Southern California, Los
1 Portfolio Selection
COS 5: Theoretical Machine Learning Lecturer: Rob Schapire Lecture # Scribe: Nadia Heninger April 8, 008 Portfolio Selection Last time we discussed our model of the stock market N stocks start on day with
CS91.543 MidTerm Exam 4/1/2004 Name: KEY. Page Max Score 1 18 2 11 3 30 4 15 5 45 6 20 Total 139
CS91.543 MidTerm Exam 4/1/2004 Name: KEY Page Max Score 1 18 2 11 3 30 4 15 5 45 6 20 Total 139 % INTRODUCTION, AI HISTORY AND AGENTS 1. [4 pts. ea.] Briefly describe the following important AI programs.
