Lecture notes: single-agent dynamics 1
|
|
|
- Julia Clark
- 10 years ago
- Views:
Transcription
1 Lecture notes: single-agent dynamics 1 Single-agent dynamic optimization models In these lecture notes we consider specification and estimation of dynamic optimization models. Focus on single-agent models. 1 Rust (1987) Rust (1987) is one of the first papers in this literature. Model is quite simple, but empirical framework introduced in this paper for dynamic discrete-choice (DDC) models is still widely applied. Agent is Harold Zurcher, manager of bus depot in Madison, Wisconsin. Each week, HZ must decide whether to replace the bus engine, or keep it running for another week. This engine replacement problem is an example of an optimal stopping problem, which features the usual tradeoff: (i) there are large fixed costs associated with stopping (replacing the engine), but new engine has lower associated future maintenance costs; (ii) by not replacing the engine, you avoid the fixed replacement costs, but suffer higher future maintenance costs. 1.1 Behavioral Model At the end of each week t, HZ decides whether or not to replace engine. Control variable defined as: i t = 1 if HZ replaces 0 otherwise. For simplicity. we describe the case where there is only one bus (in the paper, buses are treated as independent entities). HZ chooses the (infinite) sequence i 1, i 2, i 3,..., i t, i t+1,...} to maximize discounted expected utility stream: where max E i 1,i 2,i 3,...,i t,i t+1,...} β t 1 u (x t, ɛ t, i t ; θ) (1)
2 Lecture notes: single-agent dynamics 2 The state variables of this problem are: 1. x t : the mileage. Both HZ and the econometrician observe this, so we call this the observed state variable 2. ɛ t : the utility shocks. Econometrician does not observe this, so we call it the unobserved state variable x t is the mileage of the bus at the end of week t. Assume that evolution of mileage is stochastic (from HZ s point of view) and follows G(x x t ) if i t = 0 (don t replace engine in period t) x t+1 (2) G(x 0) if i t = 1: once replaced, mileage gets reset to zero and G(x x) is the conditional probability distribution of next period s mileage x given that current mileage is x. HZ knows G; econometrician knows the form of G, up to a vector of parameters which are estimated. 1 ɛ t denotes shocks in period t, which affect HZ s choice of whether to replace the engine. These are the structural errors of the model (they are observed by HZ, but not by us), and we will discuss them in more detail below. Define value function: V (x t, ɛ t ) = max i τ, τ=t+1,t+2,... E t [ τ=t+1 β τ t u (x t, ɛ t, i t ; θ) x t ] where maximum is over all possible sequences of i t+1, i t+2,...}. Note that we have imposed stationarity, so that the value function V ( ) is a function of t only indirectly, through the value that the state variable x takes during period t. 2 1 Since mileage evolves randomly, this implies that even given a sequence of replacement choices i 1, i 2, i 3,..., i t, i t+1,...}, the corresponding sequence of mileages x 1, x 2, x 3,..., x t, x t+1,...} is still random. The expectation in Eq. (1) is over this stochastic sequence of mileages and over the shocks ɛ 1, ɛ 2,...}. 2 An important distinction between empirical papers with dynamic optimization models is whether agents have infinite-horizon, or finite-horizon. Stationarity (or time homogeneity) is assumed for infinite-horizon problems, and they are solved using value function iteration. Finite-horizon problems are non-stationary, and solved by backward induction starting from the final period.
3 Lecture notes: single-agent dynamics 3 Using the Bellman equation, we can break down the DO problem into an (infinite) sequence of single-period decisions: i t = i (x t, ɛ t ; θ) = argmax i u(xt, ɛ t, i; θ) + βe x,ɛ x t,ɛ t,i t V (x, ɛ ) } where the value function is V (x, ɛ) = max u(x, ɛ, i; θ) + βex i=1,0,ɛ x t,ɛ t,i t V (x, ɛ ) } = max u(x, ɛ, 0; θ) + βe x,ɛ x t,ɛ t,i t=0v (x, ɛ ), u(x, ɛ, 1; θ) + βe x,ɛ 0,ɛ t,i V (x, ɛ ). } } = max Ṽ (x, ɛ, 1), Ṽ (x, ɛ, 0). In the above, we define the choice-specific value function u(x, ɛ, 1; θ) + βe x Ṽ (x, ɛ, i) =,ɛ x=0,ɛ,i=1v (x, ɛ ) if i = 1 u(x, ɛ, 0; θ) + βe x,ɛ x,ɛ,i=0v (x, ɛ ) if i = 0. We make the following parametric assumptions on utility flow: (3) where u(x t, ɛ t, i; θ) = c ((1 i t ) x t ; θ) i RC + ɛ it c( ) is the maintenance cost function, which is presumably increasing in x (higher x means higher costs) RC denotes the lumpy fixed costs of adjustment. The presence of these costs implies that HZ won t want to replace the engine every period. ɛ it, i = 0, 1 are structural errors, which represents factors which affect HZ s replacement choice i t in period t, but are unobserved by the econometrician. Define ɛ t (ɛ 0t, ɛ 1t ). As Rust remarks (bottom, pg. 1008), you need this in order to generate a positive likelihood for your observed data. Without these ɛ s, we observe as much as HZ does, and i t = i (x t ; θ), so that replacement decision should be perfectly explained by mileage. Hence, model will not be able to explain situations where
4 Lecture notes: single-agent dynamics 4 there are two periods with identical mileage, but in one period HZ replaced, and in the other HZ doesn t replace. (Tension between this empirical practice and falsifiability: of model) As remarked earlier, these assumption imply a very simple type of optimal decision rule i (x, ɛ; θ): in any period t, you replace when x t x (ɛ t ), where x (ɛ t ) is some optimal cutoff mileage level, which depends on the value of the shocks ɛ t. Parameters to be estimated are: 1. parameters of maintenance cost function c( ); 2. replacement cost RC; 3. parameters of mileage transition function G(x x). Remark: Distinguishing myopic from forward-looking behavior. In these models, the discount factor β is typically not estimated. Essentially, the time series data on i t, x t } could be equally well explained by a myopic model, which posits that i t = argmax i 0,1} u(x t, ɛ t, i)}, or a forward-looking model, which posits that } i t = argmax i 0,1} Ṽ (xt, ɛ t, i). In both models, the choice i t depends just on the current state variables x t, ɛ t. Indeed, Magnac and Thesmar (2002) shows that in general, DDC models are nonparametrically underidentified, without knowledge of β and F (ɛ), the distribution of the ɛ shocks. (Below, we show how knowledge of β and F, along with an additional normalization, permits nonparametric identification of the utility functions in this model.) Intuitively, in this model, it is difficult to identify β apart from fixed costs. In this model, if HZ were myopic (ie. β close to zero) and replacement costs RC were low, his decisions may look similar as when he were forward-looking (ie. β close to 1) and RC
5 Lecture notes: single-agent dynamics 5 were large. Reduced-form tests for forward-looking behavior exploit scenarios in which some variables which affect future utility are known in period t: consumers are deemed forward-looking if their period t decisions depends on these variables. Examples: Chevalier and Goolsbee (2009) examine whether students choices of purchasing a textbook now depend on the possibility that a new edition will be released soon. Becker, Grossman, and Murphy (1994) argue that cigarette addiction is rational by showing that cigarette consumption is response to permanent future changes in cigarette prices. 1.2 Econometric Model Data: observe i t, x t }, t = 1,..., T for 62 buses. Treat buses as homogeneous and independent (ie. replacement decision on bus j is not affected by replacement decision on bus j ). Rust makes the following conditional independence assumption, on the Markovian transition probabilities in the Bellman equation above: Assumption 1 (x t, ɛ t ) is a stationary controlled first-order Markov process, with transition p(x, ɛ x, ɛ, i) = p(ɛ x, x, ɛ, i) p(x x, e, i) = p(ɛ x ) p(x x, i). (4) The first line is just factoring the joint density into a conditional times a marginal. The second line shows the simplifications from Rust s assumptions. Namely, two types of conditional independence: (i) given x, ɛ s are independent over time; and (ii) conditional on x and i, x is independent of ɛ.
6 Lecture notes: single-agent dynamics 6 Likelihood function for a single bus: l (x 1,..., x T, i 1,..., i T x 0, i 0 ; θ) T = P rob (i t, x t x 0, i 0,..., x t 1, i t 1 ; θ) = = T P rob (i t, x t x t 1, i t 1 ; θ) T P rob (i t x t ; θ) P rob (x t x t 1, i t 1 ; θ 3 ). Both the third and fourth lines arise from the conditional independence assumption. Note that, in the dynamic optimization problem, the optimal choice of i t depends on the state variables (x t, ɛ t ). Hence the third line (implying that x t, i t } evolves as 1-order Markov) relies on the conditional serial independence of ɛ t. The last equality also arises from this conditional serial independence assumption. Hence, the log likelihood is additively separable in the two components: log l = T log P rob (i t x t ; θ) + T log P rob (x t x t 1, i t 1 ; θ 3 ). Here θ 3 θ denotes the subset of parameters which enter G, the transition probability function for mileage. Because θ 3 θ, we can maximize the likelihood function above in two steps. First step: Estimate θ 3, the parameters of the Markov transition probabilities for mileage. We assume a discrete distribution for mileage x, taking K distinct and equally-spaced values x [1], x [2],..., x [K] }, in increasing order, where x[k ] x [k] = (k k), where is a mileage increment (Rust considers = 5000). Also assume that given the current state x t = x [k], the mileage in the next period can more up to at most x [k+j]. (When i t = 1 so that engine is replaced, we reset x t = 0 = x [0].) Then the mileage transition probabilities can be expressed as: p j if 0 j J P (x [k+j] x [k], d = 0) = 0 otherwise so that θ 3 p 0, p 1,..., p J }, with 0 < p 0,..., p J < 1 and J j=1 p j = 1. (5) (6)
7 Lecture notes: single-agent dynamics 7 This first step can be executed separately from the substantial second step. θ 3 estimated just by empirical frequencies: ˆp j = freq x t+1 x t = j}, for all 0 j J. Second step: Estimate the remaining parameters θ\θ 3, parameters of maintenance cost function c( ) and engine replacement costs. Here, we make a further assumption: Assumption 2 The ɛ s are distributed i.i.d. (across choices and periods), according to the Type I extreme value distribution. So this implies that in Eq. (4) above, p(ɛ x ) = p(ɛ ), for all x. Expand the expression for P rob(i t = 1 x t ; θ) equals P rob c(0; θ) RC + ɛ 1t + βe x,ɛ 0V (x, ɛ ) > c(x t ; θ) + ɛ 0t + βe x,ɛ x t V (x, ɛ ) } =P rob ɛ 1t ɛ 0t > c(0; θ) c(x t ; θ) + β [ E x,ɛ x t V (x, ɛ) E x,ɛ 0V (x, ɛ ) ] + RC } Because of the logit assumptions on ɛ t, the replacement probability simplifies to a multinomial logit-like expression: = exp ( c(0; θ) RC + βe x,ɛ x t=0v (x, ɛ ) ) exp ( c(0; θ) RC + βe x,ɛ x t=0v (x, ɛ ) ) + exp ( c(x t ; θ) + βe x,ɛ x t V (x, ɛ ) ). This is called a dynamic logit model, in the literature. Defining ū(x, i; θ) u(x, ɛ, i; θ) ɛ i the choice probability takes the form P rob (i t x t ; θ) = exp ( ū(x t, i t, θ) + βe x,ɛ x t,i t V (x, ɛ ) ) i=0,1 exp ( ). (7) ū(x t, i, θ) + βe x,ɛ x t,iv (x, ɛ ) Estimation method for second step: Nested fixed-point algorithm The second-step of the estimation procedures is via a nested fixed point algorithm. Outer loop: search over different parameter values ˆθ. Inner loop: For ˆθ, we need to compute the value function V (x, ɛ; ˆθ). After V (x, ɛ; ˆθ) is obtained, we can compute the LL fxn in Eq. (7).
8 Lecture notes: single-agent dynamics Computational details for inner loop Compute value function V (x, ɛ; ˆθ) by iterating over Bellman s equation (3). A clever and computationally convenient feature in Rust s paper is that he iterates over the expected value function EV (x, i) E x,ɛ x,iv (x, ɛ ; θ). The reason for this is that you avoid having to calculate the value function at values of ɛ 0 and ɛ 1, which are additional state variables. He iterates over the following equation (which is Eq in his paper): EV (x, i) = y log j C(y) exp [ū(y, j; θ) + βev (y, j)] p(dy x, i) (8) Somewhat awkward notation: here EV denotes a function. Here x, i denotes the previous period s mileage and replacement choice, and y, j denote the current period s mileage and choice (as will be clear below). This equation can be derived from Bellman s equation (3): V (y, ɛ; θ) = max [ū(y, j; θ) + ɛ + βev (y, j)] j 0,1 } E y,ɛ [V (y, ɛ; θ) x, i] EV (x, i; θ) =E y,ɛ x,i max [ū(y, j; θ) + ɛ + βev (y, j)] j 0,1 } =E y x,i E ɛ y,x,i max [ū(y, j; θ) + ɛ + βev (y, j)] j 0,1 } =E y x,i log exp [ū(y, j; θ) + βev (y, j)] = log y j=0,1 } exp [ū(y, j; θ) + βev (y, j)] p(dy x, i). The next-to-last equality uses the closed-form expression for the expectation of the maximum, for extreme-value variates. 3 Once the EV (x, i; θ) function is computed for θ, the choice probabilities p(i t x t ) can be constructed as j=0,1 exp (ū(x t, i t ; θ) + βev (x t, i t ; θ)) i=0,1 exp (ū(x t, i; θ) + βev (x t, i; θ)). 3 See Chiong, Galichon, and Shum (2013) for the most general treatment of this.
9 Lecture notes: single-agent dynamics 9 The value iteration procedure: The expected value function EV ( ; θ) will be computed for each value of the parameters θ. The computational procedure is iterative. Let τ index the iterations. Let EV τ (x, i) denote the expected value function during the τ-th iteration. (We suppress the functional dependence of EV on θ for convenience.) Here Rust assumes that mileage is discrete- (finite-) valued, and takes K values, each spaced 5000 miles apart, consistently with earlier modeling of mileage transition function in Eq. (6). Let the values of the state variable x be discretized into a grid of points, which we denote r. Because of this assumption that x is discrete, the EV (x, i) function is now finite dimensional, having 2 K elements. τ = 0: Start from an initial guess of the expected value function EV (x, i). Common way is to start with EV (x, i) = 0, for all x r, and i = 0, 1. τ = 1: Use Eq. (8) and EV 0 (x; θ) to calculate, at each x r, and i 0, 1}. EV 1 (x, i) = log y r j C(y) exp [ ū(y, j; θ) + βev 0 (y, j) ] p(y x, i) where the transition probabilities p(y x, i) are given by Eq. (6) above. Now check: is EV 1 (x, i) close to EV 0 (x, i)? Check whether sup x,i EV 1 (x, i) EV 0 (x, i) < η where η is some very small number (eg ). If so, then you are done. If not, then go to next iteration τ = 2.
10 Lecture notes: single-agent dynamics 10 References Becker, G., M. Grossman, and K. Murphy (1994): An Empirical Analysis of Cigarette Addiction, American Economic Review, 84, Chevalier, J., and A. Goolsbee (2009): Are Durable Goods Consumers Forward Looking? Evidence from the College Textbook Market, Quarterly Journal of Economics, 124, Chiong, K., A. Galichon, and M. Shum (2013): Estimating Dynamic Discrete Choice Models via Convex Analysis, mimeo, Caltech. Magnac, T., and D. Thesmar (2002): Identifying Dynamic Discrete Decision Processes, Econometrica, 70, Rust, J. (1987): Optimal Replacement of GMC Bus Engines: An Empirical Model of Harold Zurcher, Econometrica, 55,
Lecture notes: single-agent dynamics 1
Lecture notes: single-agent dynamics 1 Single-agent dynamic optimization models In these lecture notes we consider specification and estimation of dynamic optimization models. Focus on single-agent models.
A HYBRID GENETIC ALGORITHM FOR THE MAXIMUM LIKELIHOOD ESTIMATION OF MODELS WITH MULTIPLE EQUILIBRIA: A FIRST REPORT
New Mathematics and Natural Computation Vol. 1, No. 2 (2005) 295 303 c World Scientific Publishing Company A HYBRID GENETIC ALGORITHM FOR THE MAXIMUM LIKELIHOOD ESTIMATION OF MODELS WITH MULTIPLE EQUILIBRIA:
INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition)
INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition) Abstract Indirect inference is a simulation-based method for estimating the parameters of economic models. Its
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem
LECTURE 15: AMERICAN OPTIONS
LECTURE 15: AMERICAN OPTIONS 1. Introduction All of the options that we have considered thus far have been of the European variety: exercise is permitted only at the termination of the contract. These
Monte Carlo-based statistical methods (MASM11/FMS091)
Monte Carlo-based statistical methods (MASM11/FMS091) Jimmy Olsson Centre for Mathematical Sciences Lund University, Sweden Lecture 5 Sequential Monte Carlo methods I February 5, 2013 J. Olsson Monte Carlo-based
CSCI567 Machine Learning (Fall 2014)
CSCI567 Machine Learning (Fall 2014) Drs. Sha & Liu {feisha,yanliu.cs}@usc.edu September 22, 2014 Drs. Sha & Liu ({feisha,yanliu.cs}@usc.edu) CSCI567 Machine Learning (Fall 2014) September 22, 2014 1 /
Basics of Statistical Machine Learning
CS761 Spring 2013 Advanced Machine Learning Basics of Statistical Machine Learning Lecturer: Xiaojin Zhu [email protected] Modern machine learning is rooted in statistics. You will find many familiar
Monte Carlo and Empirical Methods for Stochastic Inference (MASM11/FMS091)
Monte Carlo and Empirical Methods for Stochastic Inference (MASM11/FMS091) Magnus Wiktorsson Centre for Mathematical Sciences Lund University, Sweden Lecture 5 Sequential Monte Carlo methods I February
ECON20310 LECTURE SYNOPSIS REAL BUSINESS CYCLE
ECON20310 LECTURE SYNOPSIS REAL BUSINESS CYCLE YUAN TIAN This synopsis is designed merely for keep a record of the materials covered in lectures. Please refer to your own lecture notes for all proofs.
1 Teaching notes on GMM 1.
Bent E. Sørensen January 23, 2007 1 Teaching notes on GMM 1. Generalized Method of Moment (GMM) estimation is one of two developments in econometrics in the 80ies that revolutionized empirical work in
Lecture 3: Linear methods for classification
Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,
Lecture 3: Growth with Overlapping Generations (Acemoglu 2009, Chapter 9, adapted from Zilibotti)
Lecture 3: Growth with Overlapping Generations (Acemoglu 2009, Chapter 9, adapted from Zilibotti) Kjetil Storesletten September 10, 2013 Kjetil Storesletten () Lecture 3 September 10, 2013 1 / 44 Growth
Application of Fourier Transform to PDE (I) Fourier Sine Transform (application to PDEs defined on a semi-infinite domain)
Application of Fourier Transform to PDE (I) Fourier Sine Transform (application to PDEs defined on a semi-infinite domain) The Fourier Sine Transform pair are F. T. : U = 2/ u x sin x dx, denoted as U
Statistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
Stochastic Inventory Control
Chapter 3 Stochastic Inventory Control 1 In this chapter, we consider in much greater details certain dynamic inventory control problems of the type already encountered in section 1.3. In addition to the
Reject Inference in Credit Scoring. Jie-Men Mok
Reject Inference in Credit Scoring Jie-Men Mok BMI paper January 2009 ii Preface In the Master programme of Business Mathematics and Informatics (BMI), it is required to perform research on a business
Notes from Week 1: Algorithms for sequential prediction
CS 683 Learning, Games, and Electronic Markets Spring 2007 Notes from Week 1: Algorithms for sequential prediction Instructor: Robert Kleinberg 22-26 Jan 2007 1 Introduction In this course we will be looking
Reinforcement Learning
Reinforcement Learning LU 2 - Markov Decision Problems and Dynamic Programming Dr. Martin Lauer AG Maschinelles Lernen und Natürlichsprachliche Systeme Albert-Ludwigs-Universität Freiburg [email protected]
CHAPTER 2 Estimating Probabilities
CHAPTER 2 Estimating Probabilities Machine Learning Copyright c 2016. Tom M. Mitchell. All rights reserved. *DRAFT OF January 24, 2016* *PLEASE DO NOT DISTRIBUTE WITHOUT AUTHOR S PERMISSION* This is a
1 Short Introduction to Time Series
ECONOMICS 7344, Spring 202 Bent E. Sørensen January 24, 202 Short Introduction to Time Series A time series is a collection of stochastic variables x,.., x t,.., x T indexed by an integer value t. The
Logistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression
Logistic Regression Department of Statistics The Pennsylvania State University Email: [email protected] Logistic Regression Preserve linear classification boundaries. By the Bayes rule: Ĝ(x) = arg max
Linear Threshold Units
Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear
An Extension Model of Financially-balanced Bonus-Malus System
An Extension Model of Financially-balanced Bonus-Malus System Other : Ratemaking, Experience Rating XIAO, Yugu Center for Applied Statistics, Renmin University of China, Beijing, 00872, P.R. China Phone:
Measuring consumer switching costs in the. television industry
Measuring consumer switching costs in the television industry Oleksandr Shcherbakov January 2008 Abstract This paper develops and estimates a dynamic model of consumer behavior with switching costs in
HETEROGENEOUS AGENTS AND AGGREGATE UNCERTAINTY. Daniel Harenberg [email protected]. University of Mannheim. Econ 714, 28.11.
COMPUTING EQUILIBRIUM WITH HETEROGENEOUS AGENTS AND AGGREGATE UNCERTAINTY (BASED ON KRUEGER AND KUBLER, 2004) Daniel Harenberg [email protected] University of Mannheim Econ 714, 28.11.06 Daniel Harenberg
Language Modeling. Chapter 1. 1.1 Introduction
Chapter 1 Language Modeling (Course notes for NLP by Michael Collins, Columbia University) 1.1 Introduction In this chapter we will consider the the problem of constructing a language model from a set
Maximum Likelihood Estimation
Math 541: Statistical Theory II Lecturer: Songfeng Zheng Maximum Likelihood Estimation 1 Maximum Likelihood Estimation Maximum likelihood is a relatively simple method of constructing an estimator for
Using Markov Decision Processes to Solve a Portfolio Allocation Problem
Using Markov Decision Processes to Solve a Portfolio Allocation Problem Daniel Bookstaber April 26, 2005 Contents 1 Introduction 3 2 Defining the Model 4 2.1 The Stochastic Model for a Single Asset.........................
6. Budget Deficits and Fiscal Policy
Prof. Dr. Thomas Steger Advanced Macroeconomics II Lecture SS 2012 6. Budget Deficits and Fiscal Policy Introduction Ricardian equivalence Distorting taxes Debt crises Introduction (1) Ricardian equivalence
Non-Stationary Time Series andunitroottests
Econometrics 2 Fall 2005 Non-Stationary Time Series andunitroottests Heino Bohn Nielsen 1of25 Introduction Many economic time series are trending. Important to distinguish between two important cases:
Supplement to Call Centers with Delay Information: Models and Insights
Supplement to Call Centers with Delay Information: Models and Insights Oualid Jouini 1 Zeynep Akşin 2 Yves Dallery 1 1 Laboratoire Genie Industriel, Ecole Centrale Paris, Grande Voie des Vignes, 92290
Lecture notes on Moral Hazard, i.e. the Hidden Action Principle-Agent Model
Lecture notes on Moral Hazard, i.e. the Hidden Action Principle-Agent Model Allan Collard-Wexler April 19, 2012 Co-Written with John Asker and Vasiliki Skreta 1 Reading for next week: Make Versus Buy in
Reinforcement Learning
Reinforcement Learning LU 2 - Markov Decision Problems and Dynamic Programming Dr. Joschka Bödecker AG Maschinelles Lernen und Natürlichsprachliche Systeme Albert-Ludwigs-Universität Freiburg [email protected]
Single Agent Dynamics: Dynamic Discrete Choice Models
Single Agent Dynamics: Dynamic Discrete Choice Models Part : Estimation Günter J. Hitsch The University of Chicago Booth School of Business 013 1/83 Overview The Estimation Problem; Structural vs Reduced
Hedging Options In The Incomplete Market With Stochastic Volatility. Rituparna Sen Sunday, Nov 15
Hedging Options In The Incomplete Market With Stochastic Volatility Rituparna Sen Sunday, Nov 15 1. Motivation This is a pure jump model and hence avoids the theoretical drawbacks of continuous path models.
Credit Risk Models: An Overview
Credit Risk Models: An Overview Paul Embrechts, Rüdiger Frey, Alexander McNeil ETH Zürich c 2003 (Embrechts, Frey, McNeil) A. Multivariate Models for Portfolio Credit Risk 1. Modelling Dependent Defaults:
Principle of Data Reduction
Chapter 6 Principle of Data Reduction 6.1 Introduction An experimenter uses the information in a sample X 1,..., X n to make inferences about an unknown parameter θ. If the sample size n is large, then
Uncovering Consumer Decision Rules under Complex Dynamic Environments: The Case of Coalition Loyalty Programs
Uncovering Consumer Decision Rules under Complex Dynamic Environments: The Case of Coalition Loyalty Programs Andrew Ching Masakazu Ishihara Very Preliminary and Incomplete April 18, 2014 Abstract We propose
Optimal Paternalism: Sin Taxes and Health Subsidies
Optimal Paternalism: Sin Taxes and Health Subsidies Thomas Aronsson and Linda Thunström Department of Economics, Umeå University SE - 901 87 Umeå, Sweden April 2005 Abstract The starting point for this
The Heat Equation. Lectures INF2320 p. 1/88
The Heat Equation Lectures INF232 p. 1/88 Lectures INF232 p. 2/88 The Heat Equation We study the heat equation: u t = u xx for x (,1), t >, (1) u(,t) = u(1,t) = for t >, (2) u(x,) = f(x) for x (,1), (3)
Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay
Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 17 Shannon-Fano-Elias Coding and Introduction to Arithmetic Coding
Estimating the random coefficients logit model of demand using aggregate data
Estimating the random coefficients logit model of demand using aggregate data David Vincent Deloitte Economic Consulting London, UK [email protected] September 14, 2012 Introduction Estimation
A Uniform Asymptotic Estimate for Discounted Aggregate Claims with Subexponential Tails
12th International Congress on Insurance: Mathematics and Economics July 16-18, 2008 A Uniform Asymptotic Estimate for Discounted Aggregate Claims with Subexponential Tails XUEMIAO HAO (Based on a joint
Current Accounts in Open Economies Obstfeld and Rogoff, Chapter 2
Current Accounts in Open Economies Obstfeld and Rogoff, Chapter 2 1 Consumption with many periods 1.1 Finite horizon of T Optimization problem maximize U t = u (c t ) + β (c t+1 ) + β 2 u (c t+2 ) +...
Bayesian logistic betting strategy against probability forecasting. Akimichi Takemura, Univ. Tokyo. November 12, 2012
Bayesian logistic betting strategy against probability forecasting Akimichi Takemura, Univ. Tokyo (joint with Masayuki Kumon, Jing Li and Kei Takeuchi) November 12, 2012 arxiv:1204.3496. To appear in Stochastic
Chapter 4 Lecture Notes
Chapter 4 Lecture Notes Random Variables October 27, 2015 1 Section 4.1 Random Variables A random variable is typically a real-valued function defined on the sample space of some experiment. For instance,
Why Do Life Insurance Policyholders Lapse? The Roles of Income, Health and Bequest Motive Shocks
Why Do Life Insurance Policyholders Lapse? The Roles of Income, Health and Bequest Motive Shocks Hanming Fang Edward Kung March 22, 2012 Abstract We present and empirically implement a dynamic discrete
E 4101/5101 Lecture 8: Exogeneity
E 4101/5101 Lecture 8: Exogeneity Ragnar Nymoen 17 March 2011 Introduction I Main references: Davidson and MacKinnon, Ch 8.1-8,7, since tests of (weak) exogeneity build on the theory of IV-estimation Ch
Multiple Choice Models II
Multiple Choice Models II Laura Magazzini University of Verona [email protected] http://dse.univr.it/magazzini Laura Magazzini (@univr.it) Multiple Choice Models II 1 / 28 Categorical data Categorical
Dynamics of Small Open Economies
Dynamics of Small Open Economies Lecture 2, ECON 4330 Tord Krogh January 22, 2013 Tord Krogh () ECON 4330 January 22, 2013 1 / 68 Last lecture The models we have looked at so far are characterized by:
NBER WORKING PAPER SERIES WHY DO LIFE INSURANCE POLICYHOLDERS LAPSE? THE ROLES OF INCOME, HEALTH AND BEQUEST MOTIVE SHOCKS. Hanming Fang Edward Kung
NBER WORKING PAPER SERIES WHY DO LIFE INSURANCE POLICYHOLDERS LAPSE? THE ROLES OF INCOME, HEALTH AND BEQUEST MOTIVE SHOCKS Hanming Fang Edward Kung Working Paper 17899 http://www.nber.org/papers/w17899
Real Business Cycles. Federal Reserve Bank of Minneapolis Research Department Staff Report 370. February 2006. Ellen R. McGrattan
Federal Reserve Bank of Minneapolis Research Department Staff Report 370 February 2006 Real Business Cycles Ellen R. McGrattan Federal Reserve Bank of Minneapolis and University of Minnesota Abstract:
Natural Language Processing. Today. Logistic Regression Models. Lecture 13 10/6/2015. Jim Martin. Multinomial Logistic Regression
Natural Language Processing Lecture 13 10/6/2015 Jim Martin Today Multinomial Logistic Regression Aka log-linear models or maximum entropy (maxent) Components of the model Learning the parameters 10/1/15
Optimal Health Insurance for Prevention and Treatment
Optimal Health Insurance for Prevention and Treatment Randall P. Ellis Department of Economics Boston University Willard G. Manning Harris School of Public Policy Studies The University of Chicago We thank
LECTURE 4. Last time: Lecture outline
LECTURE 4 Last time: Types of convergence Weak Law of Large Numbers Strong Law of Large Numbers Asymptotic Equipartition Property Lecture outline Stochastic processes Markov chains Entropy rate Random
Online Appendix to Stochastic Imitative Game Dynamics with Committed Agents
Online Appendix to Stochastic Imitative Game Dynamics with Committed Agents William H. Sandholm January 6, 22 O.. Imitative protocols, mean dynamics, and equilibrium selection In this section, we consider
Average Redistributional Effects. IFAI/IZA Conference on Labor Market Policy Evaluation
Average Redistributional Effects IFAI/IZA Conference on Labor Market Policy Evaluation Geert Ridder, Department of Economics, University of Southern California. October 10, 2006 1 Motivation Most papers
1 The Brownian bridge construction
The Brownian bridge construction The Brownian bridge construction is a way to build a Brownian motion path by successively adding finer scale detail. This construction leads to a relatively easy proof
Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.
Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C
Markups and Firm-Level Export Status: Appendix
Markups and Firm-Level Export Status: Appendix De Loecker Jan - Warzynski Frederic Princeton University, NBER and CEPR - Aarhus School of Business Forthcoming American Economic Review Abstract This is
Introduction to General and Generalized Linear Models
Introduction to General and Generalized Linear Models General Linear Models - part I Henrik Madsen Poul Thyregod Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby
Inflation. Chapter 8. 8.1 Money Supply and Demand
Chapter 8 Inflation This chapter examines the causes and consequences of inflation. Sections 8.1 and 8.2 relate inflation to money supply and demand. Although the presentation differs somewhat from that
Machine Learning and Pattern Recognition Logistic Regression
Machine Learning and Pattern Recognition Logistic Regression Course Lecturer:Amos J Storkey Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh Crichton Street,
Real Business Cycle Models
Phd Macro, 2007 (Karl Whelan) 1 Real Business Cycle Models The Real Business Cycle (RBC) model introduced in a famous 1982 paper by Finn Kydland and Edward Prescott is the original DSGE model. 1 The early
Comparing Features of Convenient Estimators for Binary Choice Models With Endogenous Regressors
Comparing Features of Convenient Estimators for Binary Choice Models With Endogenous Regressors Arthur Lewbel, Yingying Dong, and Thomas Tao Yang Boston College, University of California Irvine, and Boston
Martingale Pricing Applied to Options, Forwards and Futures
IEOR E4706: Financial Engineering: Discrete-Time Asset Pricing Fall 2005 c 2005 by Martin Haugh Martingale Pricing Applied to Options, Forwards and Futures We now apply martingale pricing theory to the
Optimization under uncertainty: modeling and solution methods
Optimization under uncertainty: modeling and solution methods Paolo Brandimarte Dipartimento di Scienze Matematiche Politecnico di Torino e-mail: [email protected] URL: http://staff.polito.it/paolo.brandimarte
CCNY. BME I5100: Biomedical Signal Processing. Linear Discrimination. Lucas C. Parra Biomedical Engineering Department City College of New York
BME I5100: Biomedical Signal Processing Linear Discrimination Lucas C. Parra Biomedical Engineering Department CCNY 1 Schedule Week 1: Introduction Linear, stationary, normal - the stuff biology is not
Impulse Response Functions
Impulse Response Functions Wouter J. Den Haan University of Amsterdam April 28, 2011 General definition IRFs The IRF gives the j th -period response when the system is shocked by a one-standard-deviation
Racetrack Betting and Consensus of Subjective Probabilities
Racetrack Betting and Consensus of Subective Probabilities Lawrence D. Brown 1 and Yi Lin 2 University of Pennsylvania and University of Wisconsin, Madison Abstract In this paper we consider the dynamic
A simple analysis of the TV game WHO WANTS TO BE A MILLIONAIRE? R
A simple analysis of the TV game WHO WANTS TO BE A MILLIONAIRE? R Federico Perea Justo Puerto MaMaEuSch Management Mathematics for European Schools 94342 - CP - 1-2001 - DE - COMENIUS - C21 University
Detection of changes in variance using binary segmentation and optimal partitioning
Detection of changes in variance using binary segmentation and optimal partitioning Christian Rohrbeck Abstract This work explores the performance of binary segmentation and optimal partitioning in the
These slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher Bishop
Music and Machine Learning (IFT6080 Winter 08) Prof. Douglas Eck, Université de Montréal These slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher
Risk Preferences and Demand Drivers of Extended Warranties
Risk Preferences and Demand Drivers of Extended Warranties Online Appendix Pranav Jindal Smeal College of Business Pennsylvania State University July 2014 A Calibration Exercise Details We use sales data
1. (First passage/hitting times/gambler s ruin problem:) Suppose that X has a discrete state space and let i be a fixed state. Let
Copyright c 2009 by Karl Sigman 1 Stopping Times 1.1 Stopping Times: Definition Given a stochastic process X = {X n : n 0}, a random time τ is a discrete random variable on the same probability space as
Probabilistic Models for Big Data. Alex Davies and Roger Frigola University of Cambridge 13th February 2014
Probabilistic Models for Big Data Alex Davies and Roger Frigola University of Cambridge 13th February 2014 The State of Big Data Why probabilistic models for Big Data? 1. If you don t have to worry about
An Environment Model for N onstationary Reinforcement Learning
An Environment Model for N onstationary Reinforcement Learning Samuel P. M. Choi Dit-Yan Yeung Nevin L. Zhang pmchoi~cs.ust.hk dyyeung~cs.ust.hk lzhang~cs.ust.hk Department of Computer Science, Hong Kong
2WB05 Simulation Lecture 8: Generating random variables
2WB05 Simulation Lecture 8: Generating random variables Marko Boon http://www.win.tue.nl/courses/2wb05 January 7, 2013 Outline 2/36 1. How do we generate random variables? 2. Fitting distributions Generating
Lecture 4 Online and streaming algorithms for clustering
CSE 291: Geometric algorithms Spring 2013 Lecture 4 Online and streaming algorithms for clustering 4.1 On-line k-clustering To the extent that clustering takes place in the brain, it happens in an on-line
Financial Risk Forecasting Chapter 8 Backtesting and stresstesting
Financial Risk Forecasting Chapter 8 Backtesting and stresstesting Jon Danielsson London School of Economics 2015 To accompany Financial Risk Forecasting http://www.financialriskforecasting.com/ Published
Normalization and Mixed Degrees of Integration in Cointegrated Time Series Systems
Normalization and Mixed Degrees of Integration in Cointegrated Time Series Systems Robert J. Rossana Department of Economics, 04 F/AB, Wayne State University, Detroit MI 480 E-Mail: [email protected]
Pricing of an Exotic Forward Contract
Pricing of an Exotic Forward Contract Jirô Akahori, Yuji Hishida and Maho Nishida Dept. of Mathematical Sciences, Ritsumeikan University 1-1-1 Nojihigashi, Kusatsu, Shiga 525-8577, Japan E-mail: {akahori,
Schooling, Political Participation, and the Economy. (Online Supplementary Appendix: Not for Publication)
Schooling, Political Participation, and the Economy Online Supplementary Appendix: Not for Publication) Filipe R. Campante Davin Chor July 200 Abstract In this online appendix, we present the proofs for
ALGORITHMIC TRADING WITH MARKOV CHAINS
June 16, 2010 ALGORITHMIC TRADING WITH MARKOV CHAINS HENRIK HULT AND JONAS KIESSLING Abstract. An order book consists of a list of all buy and sell offers, represented by price and quantity, available
Numerical methods for American options
Lecture 9 Numerical methods for American options Lecture Notes by Andrzej Palczewski Computational Finance p. 1 American options The holder of an American option has the right to exercise it at any moment
