Single Agent Dynamics: Dynamic Discrete Choice Models
|
|
|
- Vernon Buddy Carter
- 10 years ago
- Views:
Transcription
1 Single Agent Dynamics: Dynamic Discrete Choice Models Part : Estimation Günter J. Hitsch The University of Chicago Booth School of Business 013 1/83 Overview The Estimation Problem; Structural vs Reduced Form of the Model dentification Nested Fixed Point (NFP) Estimators Unobserved State Variables Example: The Bayesian Learning Model Model Estimation in the Presence of Unobserved State Variables Unobserved Heterogeneity Postscript: Sequential vs. Myopic Learning The MPEC Approach to ML Estimation Example: Estimation of the Durable Goods Adoption Problem Using NFP and MPEC Approaches Other Topics References /83
2 Overview The Estimation Problem; Structural vs Reduced Form of the Model dentification Nested Fixed Point (NFP) Estimators Unobserved State Variables Example: The Bayesian Learning Model Model Estimation in the Presence of Unobserved State Variables Unobserved Heterogeneity Postscript: Sequential vs. Myopic Learning The MPEC Approach to ML Estimation Example: Estimation of the Durable Goods Adoption Problem Using NFP and MPEC Approaches Other Topics References 3/83 Data We observe the choice behavior of N individuals For each individual i, define a vector of states and choices for periods t =0,...,T i : Q i =(x it,a it ) T i i=1 The full data vector is Q =(Q 1,...,Q N ) We will initially assume that all individuals are identical and that the components of the state x are observed to us We will discuss the case of (permanent) heterogeneity and unobserved state variables later 4/83
3 Data-generating process We assume that the choice data are generated based on the choice-specific value functions Z v j (x) =u j (x)+ w(x 0 )f(x 0 x, j) dx 0 We observe action a it conditional on the state x it if and only if v k (x it )+ kit v j (x it )+ jit for all j A, j6= k The CCP k (x it ) is the probability that the inequality above holds, given the distribution of the latent utility components g( ) 5/83 Static vs dynamic discrete choice models Using the concept of choice-specific value functions, the predictions of a dynamic discrete choice model can be expressed in the same manner as the predictions of a static discrete choice model Therefore, it appears that we can simply estimate each v j (x) by approximating it using a flexible functional form, e.g.a polynomial or a linear combination of basis functions: v j (x) j (x; ) n a static discrete choice model, we typically start directly with a parametric specification of each choice-specific utility function, u j (x; ) Unlike u j (x), v j (x) is not a structural object, but the solution of a dynamic decision process that depends on the model primitives, u j (x), f(x 0 x, a),, and g( ). Why is this important? Because it affects the questions that the estimated model can answer 6/83
4 Example: The storable goods demand problem Remember that x t (i t,p t ) it {0, 1,...,}, andp t {P (1),...,P (L) } The random utility components are Type Extreme Value, hence the discrete values v j (x) can be estimated just as in a standard multinomial logit model Suppose we estimate v j (x) based on data generated from the high/low promotion process discussed in the example in Part We want to evaluate a policy where the price is permanently set at the low price level Will knowledge of v j (x) allow us to answer this question? 7/83 CCP s Pricing with promotions Small pack size Promoted price Regular price Large pack size Promoted price Regular price Purchase probability Purchase probability nventory units nventory units 8/83
5 CCP s only low price Small pack size Promoted price Large pack size Promoted price Purchase probability Purchase probability nventory units nventory units 9/83 Example: The storable goods demand problem Based on the data generated under the price promotion process, the sales volume ratio between promoted and regular price periods is 3.78/0.358 = 10.6 However, when we permanently lower the price the sales ratio is only 0.991/0.358 =.8 10 / 83
6 Example: The storable goods demand problem Evaluation of a policy where the price is permanently set at the low price level: We are not just changing a component of x (the price), but also the price expectations, f(x 0 x, a) Hence, vj (x) will also change to ṽ j (x) Unless consumer choices under the alternative price process are observed in our data, we cannot estimate ṽ j (x) nstead, we must predict ṽj (x) using knowledge of the model primitives, u j (x), f(x 0 x, a),, and g( ) The same problem does not arise in a static discrete choice model (if the static discrete choice model accurately describes consumer behavior) 11 / 83 Structural and reduced form of the model (u j (x),f(x 0 x, a),,g( )) is the structural form of the dynamic discrete choice model (v j (x),g( )) is the reduced form of the model The reduced form describes the joint distribution of the data, but typically cannot predict the causal effect of a marketing policy intervention Policy predictions are therefore not only dependent on the statistical properties of the model and model parameters, but also on the behavioral assumptions we make about how decisions are made Good background reading on structural estimation and the concept of inferring causality from data that are non-experimentally generated: Reiss and Wolak (007) 1 / 83
7 Overview The Estimation Problem; Structural vs Reduced Form of the Model dentification Nested Fixed Point (NFP) Estimators Unobserved State Variables Example: The Bayesian Learning Model Model Estimation in the Presence of Unobserved State Variables Unobserved Heterogeneity Postscript: Sequential vs. Myopic Learning The MPEC Approach to ML Estimation Example: Estimation of the Durable Goods Adoption Problem Using NFP and MPEC Approaches Other Topics References 13 / 83 dentification All behavioral implications of the model are given by the collection of CCP s { j (x) :j A,x X} Suppose we have infinitely many data points, so that we can observe the CCP s { j (x)} For example, if X is finite we can estimate j(x) as the frequency of observing j conditional on x We assume we know the distribution of the random utility components, g( ) Could we then uniquely infer the model primitives describing behavior from the data: u j (x), f(x 0 x, j),?.e., is the structural form of the model identified (in a non-parametric sense)? 14 / 83
8 dentification Hotz and Miller (1993): f has a density with respect to the Lebesgue measure on R K+1 and is strictly positive, then we can invert the observed CCP s to infer the choice-specific value function differences: v j (x) v 0 (x) = 1 j ( (x)) for all j A f j is Type Extreme Value distributed, this inversion has a closed form: v j (x) v 0 (x) = log( j (x)) log( 0 (x)) We see that =0,u 0 (x) 0, and u j (x) log( j (x)) log( 0 (x)) completely rationalize the data! 15 / 83 dentification Theorem Suppose we know the distribution of the random utility components, g( ), the consumers beliefs about the evolution of the state vector, f(x 0 x, j), and the discount factor. Assume that u 0 (x) 0. Let the CCP s, j(x), be given for all x and j A. Then: (i) We can infer the unique choice-specific value functions, v j (x), consistent with the consumer decision model. (ii) The utilities u j (x) are identified for all states x and choices j. 16 / 83
9 (Non)identification proof Define the difference in choice-specific value functions: ṽ j (x) v j (x) v 0 (x) Hotz-Miller (1993) inversion theorem: ṽ j (x) = 1 j ( (x)) for all j A The expected value function can be expressed as a function of the data and the reference alternative: Z w(x) = max {v k(x)+ k }g( )d ka Z = max {ṽ k(x)+ k }g( )d + v 0 (x) ka 17 / 83 (Non)identification proof Choice-specific value of reference alternative: Z v 0 (x) =u 0 (x)+ w(x 0 )f(x 0 x, 0)dx 0 Z = max {ṽ k(x 0 )+ k }g( )f(x 0 x, 0)d dx 0 ka Z + v 0 (x 0 )f(x 0 x, 0)dx 0 Defines a contraction mapping and hence has a unique solution Recover all choice-specific value functions: v j (x) = j( (x)) + v 0 (x) Recover the utility functions: u j (x) =v j (x) Z w(x 0 )f(x 0 x, j)dx 18 / 83
10 (Non)identification The proof of the proposition shows how to calculate the utilities, u j (x), from the data and knowledge of and f(x 0 x, j) Proof shows that if 0 6= or f 0 (x 0 x, j) 6= f(x 0 x, j) ) u 0 j (x) 6= u j(x) in general mplications f either the discount factor or the consumer s belief about the state evolution is unknown, the utility function is not identified.e., the model primitives uj (x),,f(x 0 x, j) are non-parametrically unidentified 19 / 83 mplications of (non)identification result n practice: Researchers assume a given discount factor calibrated from some overall interest rate r, such that =1/(1 + r) u 0 (c t )= E t [(1 + r)u 0 (c t+1 )] Typically, discount factor corresponding to 5%-10% interest rate is used Assume rational expectations: the consumer s subjective belief f(x 0 x, j) coincides with the actual transition process of x t Allows us to estimate f(x 0 x, j) from the data 0 / 83
11 dentifying based on exclusion restrictions A folk theorem : is identified if there are states that do not affect the current utility but the transition probability of x More formally: Suppose there are two states x 1 and x such that u j (x 1 )=u j (x ) but f(x 0 x 1,j) 6= f(x 0 x,j) ntuition: Variation in x does not change the current utility but the future expected value, and thus is identified: Z v j (x) =u j (x)+ w(x 0 )f(x 0 x, j) dx 0 1 / 83 dentifying based on exclusion restrictions The folk theorem above is usually attributed to Magnac and Thesmar (00), but the actual statement provided in their paper is more complicated Define the current value function, the expected difference between (i) choosing action j today, action 0 tomorrow, and then behaving optimally afterwards, and (ii) choosing action 0 today and tomorrow and behaving optimally afterwards Z U j (x) u j (x)+ v 0 (x 0 )p(x 0 x, j)dx 0 Z u 0 (x)+ v 0 (x 0 )p(x 0 x, 0)dx 0 Also, define Z V j (x) (w(x 0 ) v 0 (x 0 )) p(x 0 x, j)dx 0 / 83
12 dentifying based on exclusion restrictions The proposition proved in Magnac and Thesmar (00): Theorem Suppose there are states x 1 and x,x 1 6= x, such that U j (x 1 )=U j (x ) for some action j. Furthermore, suppose that (V j (x 1 ) V 0 (x 1 )) (V j (x )+V 0 (x )) 6= 0. Then the discount factor is identified. Note: The assumptions of the theorem require knowledge of the solution of the decision process and are thus difficult to verify 3 / 83 dentifying based on exclusion restrictions s the folk theorem true? Although widely credited to Magnac and Thesmar (00), believe this attribution is false However, a recent paper by Fang and Wang (013), Estimating Dynamic Discrete Choice Models with Hyperbolic Discounting, with an Application to Mammography Decisions, seems to prove the claim in the folk theorem 4 / 83
13 dentification of from data on static and dynamic decisions Yao et al. (01) Observe consumer choice data across two scenarios Examples: Static: Current choice does not affect future payoffs Dynamic Cell phone customers are initially on a linear usage plan, and are then switched to a three-part tariff Under the three-part tariff current cell phone usage affects future per-minute rate Any finite horizon problem Yao et al. prove identification of for continuous controls No proof provided for discrete choices, but (my guess) statement is true more generally 5 / 83 dentification of using stated choice data Dubé, Hitsch, and Jindal (013) Conjoint design to infer product adoption choices Present subjects with forecasts of future states (e.g. prices) Collect stated choice data on adoption timing dentification assumption: Subjects take the forecast of future states as given Allows identification of discount factor discount function (t) ntuition, or more generally a Treatments: Manipulations of states that change current period utilities by the same amount in period t =0and t>0 Relative effect on choice probabilities and hence choice-specific value differences at t>0 versus t =0identifies (t) 6 / 83
14 Overview The Estimation Problem; Structural vs Reduced Form of the Model dentification Nested Fixed Point (NFP) Estimators Unobserved State Variables Example: The Bayesian Learning Model Model Estimation in the Presence of Unobserved State Variables Unobserved Heterogeneity Postscript: Sequential vs. Myopic Learning The MPEC Approach to ML Estimation Example: Estimation of the Durable Goods Adoption Problem Using NFP and MPEC Approaches Other Topics References 7 / 83 Goal of estimation We would like to recover the structural form of the dynamic discrete choice model, (u j (x),f(x 0 x, a),,g( )) We assume: g( ) is known The decision makers have rational expectations, and thus f(x 0 x, a) is the true transition density of the data Typically we also assume that is known Assume that the utility functions and transition densities are parametric functions indexed by : u j (x; ) and f(x 0 x, a; ) Goal: Develop an estimator for 8 / 83
15 Likelihood function The likelihood contribution of individual i is l i (Q i ; ) = YT i t=0 The likelihood function is Pr{a it x it ; } f(x it x i,t 1,a i,t 1 ; ) Pr{a i0 x i0 ; } (x i0 ; ) l(q; ) = NY l i (Q i ; ) i=1 Define the maximum likelihood estimator, NFP = arg max{log(l(q; ))}!... 9 / 83 The nested fixed point estimator NFP is a nested fixed point estimator We employ a maximization algorithm that searches over possible values of Given, we first solve for w(x; ) as the fixed point of the integrated Bellman equation Given w(x; ), we calculate the choice-specific value functions and then the CCP s j(x; ) Allows us to assemble l(q; ) The solution of the fixed point w(x; ) is nested in the maximization algorithm This estimator is computationally intensive, as we need to solve for the expected value function at each! 30 / 83
16 Estimating in two steps Separate =( u, f ) into components that affect the utility functions, u j (x; u ), and the transition densities f(x 0 x, a; f ) Note the log-likelihood contribution of individual i: XT i log(l i (Q i ; )) = log(pr{a it x it ; u, f })+... t=1 XT i t= log(f(x it x i,t 1,a i,t 1 ; f )) This expression suggests that we can estimate in two steps: 1. Find a consistent estimator for f (need not be a ML estimator). Conditional on ˆ f, maximize the sum in the first row above to find ˆ u (adjust standard errors to account for sampling error in first step) 31 / 83 Overview The Estimation Problem; Structural vs Reduced Form of the Model dentification Nested Fixed Point (NFP) Estimators Unobserved State Variables Example: The Bayesian Learning Model Model Estimation in the Presence of Unobserved State Variables Unobserved Heterogeneity Postscript: Sequential vs. Myopic Learning The MPEC Approach to ML Estimation Example: Estimation of the Durable Goods Adoption Problem Using NFP and MPEC Approaches Other Topics References 3 / 83
17 Bayesian learning models in marketing Large literature on demand for experience goods Following Erdem and Keane (1996) Product (or service) needs to be consumed or used to fully ascertain its utility Examples Product with unknown flavor or texture Pharmaceutical drug with unknown match value, e.g. effectiveness or side effects mportance: Consumer learning may cause inertia in brand choices nertia = state dependence in a purely statistical sense f true, has implications for pricing and other marketing actions Optimal sequential learning about demand Hitsch (006) 33 / 83 Bayesian learning: General model structure Learning about an unknown parameter vector # R N nformation acquisition: Sampling (choosing) j A yields a signal j f j ( #) Knowledge: Prior t (#) Knowledge about # before any additional information in period t is sampled Learning through Bayesian updating: t+1 (#) t (# jt ) / f j ( jt #) t (#) Posterior at end of period t is prior at the beginning of period t / 83
18 General model structure Notation: t t (#) State vector x t =( t,z t ) gnore for now that t is infinite-dimensional in general Utility: Z u j (x t )=E(u j (z t, jt, #) x t )= u j (z t,, #)f j ( #) t (#) d d# Expected utility given zt and the agents belief t (#) about # 35 / 83 General model structure We assume the decision maker is able to anticipate how her knowledge evolves, conditional on the potential information that she may receive in this period Allows to define a corresponding Markov transition probability f( t+1 t,a t ) ) learning model is a special case of the dynamic discrete choice framework 36 / 83
19 Review: Normal linear regression model with conjugate priors Sample y N(X, ), y R m, R k Suppose is known Can be generalized with inverse-wishart prior on, but rarely (never?) used in extant consumer learning literature Goal: nference about Prior: N( 0, 0 ), Then the posterior is also normal: p( y, X, ) =N( n, n ) n =( X T 1 X) 1 n =( X T 1 X) 1 ( X T 1 y) 37 / 83 The Erdem and Keane (1996) learning model Consumers make purchase (= consumption) decisions over time, t =0, 1,... # j is the mean attribute level (quality) of product j Affects utility Consumers face uncertainty over #j Realized utility from consumption in period t affected by realized attribute level: jt = # j + jt, jt N(0, ) nherent variability in attribute level Variability in consumer s perception of attribute level Utility: u j (z t, jt )= jt r jt P jt r>0:risk aversion 38 / 83
20 nformation sources Assume for notational simplicity that there is only one product with unknown attribute level, and hence we can drop the j index Learning from consumption: t = # + t, t N(0, ) Let t 1,t,... denote the time periods when a consumer receives a consumption signal Let H t = {t k : t k <t} be the time periods prior to t when a consumption signal was received, and N t = H t be the corresponding number of signals Mean consumption signal prior to t : t = 1 N t X H t 39 / 83 Prior Prior in the first decision period: 0 (#) =N(µ 0, 0 )=N( #, 0 ) # is the product class mean attribute level Refer back to the slide on the normal linear regression model with conjugate priors, and define: = # X =(1,...,1) (vector of length N t ) = diagonal matrix with elements 0 = # 0 = 0 Elements in correspond to the consumption and advertising signals in H t (order does not matter) 40 / 83
21 Posterior Results on normal linear regression model show that prior t = posterior given the signals for <tis also normal: t (#) =N(µ t, t ) t = n = 1 0 µ t = n = N t + N t µ 0 + N t t Remember: nverse of covariance matrix also called the precision matrix Precisions add up in the normal linear regression model with conjugate priors Shows that we can equate the prior in period t with the mean and variance of a normal distribution: t =(µ t, t ) 41 / 83 Transition of prior Conditional on the current prior t =(µ t, t ), and only in periods t when the consumer receives a consumption signal t : t+1 = 1 t µ t+1 = 1 t t µ t + 1 t Can correspondingly derive the Markov transition density f( t+1 t,a t ) t evolves deterministically µt+1 is normally distributed with mean E(µ t+1 t,a t )=µ t Note that var(µ t+1 t,a t ) 6= t+1 4 / 83
22 Model solution Reintroduce j subscript Priors are independent, and t =( 1t,..., Jt ) Recall that jt = # j + jt Expected utility: u j (x t )=E(u j (z t, jt ) t,z t )= µ jt rµ jt r( jt + ) P jt State transition: f(x t+1 x t,a t )=f(z t+1 z t,a t ) JY f( j,t+1 jt,a t ) j=1 ) well-defined dynamic discrete choice model with unique choice-specific value functions characterizing optimal decisions 43 / 83 Estimation Conditional on x t =(z t, t ) we can calculate the CCP s Pr{a t = j z t, t } = j (z t, t ) Data Qt = a t, Q t =(Q 0,...,Q t ),andq =(Q 0,...,Q T ) zt =(z 0,...,z t ) and z =(z 0,...,z T ) Assume f(z t+1 z t,a t ) is known Estimated from data in a preliminary estimation step Difficulty in constructing the likelihood: t is an unobserved state Need to integrate out the unobserved states from the likelihood function 44 / 83
23 Constructing the priors Define t =( 1t,..., Jt ) and =( 0,..., T 1 ) Conditional on # j (an estimated parameter) and, we know the signals jt = # j + jt By assumption j0 (# j )=N(µ j0, j0 )=N( #, 0 ) Given we can then infer the sequence of priors 0, 1,..., T 1 : j,t+1 = µ j,t+1 = 1 jt 1 jt + 1! 1 + 1! 1! 1 µ jt + 1 jt jt 45 / 83 Likelihood function is a vector of model parameters We know from the discussion above that Pr{a t = j z t, t ; } =Pr{a t = j Q t 1, z t, ; } Define Z l( Q, z) f(q z; ) = T Y t=0 Pr{Q t Q t 1, z t, ; }! f( ; )d High-dimensional integral Simulation estimator Draw (r) Average over draws to simulate the likelihood l( Q, z) =l i ( Q i, z i ) is the likelihood contribution for one household i Likelihood components conditionally independent across households given ) joint likelihood is l( Q, z) = Q i l i( Q i, z i ) 46 / 83
24 Recap: Unobserved states and the likelihood function General approach Note 1. Find a vector of random variables with density p( ; ) that allows to reconstruct the sequence of unobserved states. Formulate the likelihood conditional on 3. ntegrate over to calculate the likelihood conditional on data and only Requires that numerical integration with respect to or simulation from p( ; ) is possible 47 / 83 nitial conditions Assumption so far: The first sample period is the first period when the consumers start buying the products What if the consumers made product choices before the sample period? Would j0 ( ) =N( #, Solution in Erdem and Keane (1996) Note 0) be a valid assumption? Split the sample periods into two parts, t =0,...,T0 1 and t = T 0,...,T Make some assumption about the prior in period t =0 Simulate the evolution of the priors conditional on the observed product choices and for periods t apple T 0 Conditional on the simulated draw of j0 ( ) formulate the likelihood using data from periods t = T 0,...,T The underlying assumption is that T0 is large enough such that the effect of the initial condition in t =0vanishes 48 / 83
25 Unobserved heterogeneity The most common methods to incorporate unobserved heterogeneity in dynamic discrete choice models are based on a finite mixture or latent class approach Assume there is a finite number of M consumer/household types, each characterized by a parameter m Let m be the fraction of consumers of type m in the population f there is no correlation between a consumer s type and the initial state, we can define the likelihood contribution for i as l i (Q i ;, ) = MX l i (Q i ; m ) m m=1 Here, =( 1,..., M 1 ) Obviously, the computational overhead increases in M 49 / 83 Unobserved heterogeneity Matters are more complicated if individuals are systematically in specific states depending on their type For example, preference for a brand may be correlated with product experience or stockpiling of that brand n that case, m (x i1 )=Pr(i is of type m x i1 ) 6= m An easy solution (Heckman 1981) is to form an auxiliary model for this probability, i.e. make m (x i1 ; ) a parametrically specified function with to be estimated Example: m (x i1 ; ) = exp(x i1 m ) 1+ P M 1 n=1 exp(x i1 n ) for m =1,...,M 1 50 / 83
26 Continuous types We now consider a more general form of consumer heterogeneity, which allows for a continuous distribution of types Consumer i s parameter vector # i is a function of a common component and some idiosyncratic component! i, which is drawn from a distribution with density ( ) is known, i.e., does not depend on any parameters that are estimated Example: =(µ, L), and # i = µ + L! i,! i N(0,) 51 / 83 Continuous types Given (! i, ), we can compute the choice specific value functions, calculate the CCP s Pr{a it x it ;! i, }, and then obtain the likelihood contribution for individual i : Z l i (Q i ; ) = l i (Q i ;!, ) (!)d! f the integral is high-dimensional, we need to calculate it by simulation (! (r) ) l i (Q i ; ) 1 R RX l i (Q i ;! (r), ) The problem with this approach: instead of just re-solving for the value functions once when we change, we need to re-solve for the value function R times r=1 5 / 83
27 Change of variables and importance sampling Ackerberg (009) proposes a method based on a change of variables and importance sampling to overcome this computational challenge (see Hartmann 006 for an application) Assume # i = (! i, ) (could additionally allow parameters to be a function of household characteristics) # i fully summarizes the behavior (choice-specific value functions) of household i Let p(# i ) be the density of # i The support of p must be the same for each n our example, # i N(µ, LL T ) 53 / 83 Change of variables and importance sampling Let g( ) be a density that is positive on the support of p Then For example, we could choose g( ) =p( 0 ) where 0 is some arbitrary parameter Z l i (Q i ; ) = Z = Z = l i (Q i ;!, ) (!)d! l i (Q i ; #)p(# )d# l i (Q i ; #) p(# ) g(#) g(#)d# n the second line we use a change of variables n the third line we use importance sampling 54 / 83
28 Change of variables and importance sampling We can now simulate the integral as follows: l i (Q i ; ) 1 R RX r=1 l i (Q i ; # (r) ) p(#(r) ) g(# (r) ) Note that # (r) is drawn from g, a distribution that does not depend on the parameter vector Hence, as we change, only the weights p(# (r) )/g(# (r) ) change, but not l i! We only need to calculate the value functions R times ntuition behind this approach: Change only the weights on different possible household types, not directly the households as we search for an optimal Based on the simulated likelihood function, we can define an SML (simulated maximum likelihood) estimator for 55 / 83 Understanding sequential, forward-looking learning Example: A firm launches a product, but is uncertain about its profitability Firm s profits from the product given by # # {# L, # H } could be either negative, # L < 0, or positive, # H > 0 t : prior probability that profits are negative, t =Pr{# = # L } Firm decisions a t {0, 1} at =0denotes that the firm scraps the product, and receives the payoff 0 at =1denotes that the firm stays in the market, and receives the payoff # Assumption: f the firm stays in the market it observes profits and immediately learns the true value of # 56 / 83
29 Expected profit: u 0 ( t )=0 u 1 ( t )=E(# t )= t # L +(1 t )# H No latent payoff terms jt in this model Suppose that the firm has information that the product is not profitable, that is E(# t )= t # L +(1 t )# H < 0 What action should the firm take scrap or stay in the market? 57 / 83 Myopic vs. forward-looking decision making f the firm only cares about current profits, scrap the product: u 1 ( t ) < 0=u 0 ( t ) =0, then it should But what if > 0? Let s approach this considering the impact of the current decision on future information, and the optimal decision that the firm can take based on future information 58 / 83
30 Choice-specific value functions f the firm stays in the market and sells the product, it learns the true level of profits Hence, in the next period, t +1, either t+1 =0or t+1 =1 No more uncertainty about the profit level For the case of certainty we can easily calculate the value function: v(1) = 0 v(0) = 1 1 # H Choice-specific value functions for arbitrary t : v 0 ( t )=0 v 1 ( t )=u 1 ( t )+ E (v( t+1 ) t,a 1 = 1) =( t # L +(1 t )# H )+ t 0+(1 t ) 1 = t # L +(1 1 t ) 1 # H 1 # H 59 / 83 Optimal forward-looking decision making under learning Keep the product in the market if and only if v 1 ( t )= t # L +(1 t ) 1 1 # H > 0=v 0 ( t ) Reduces to myopic decision rule if =0 Example #L = 1 and # H =1 Static decision making: Product will be scrapped iff t 0.5 Suppose the firm is forward-looking, =0.9. Then the product will be scrapped iff t 10/ / 83
31 Overview The Estimation Problem; Structural vs Reduced Form of the Model dentification Nested Fixed Point (NFP) Estimators Unobserved State Variables Example: The Bayesian Learning Model Model Estimation in the Presence of Unobserved State Variables Unobserved Heterogeneity Postscript: Sequential vs. Myopic Learning The MPEC Approach to ML Estimation Example: Estimation of the Durable Goods Adoption Problem Using NFP and MPEC Approaches Other Topics References 61 / 83 Overview Su and Judd (01) propose an alternative approach to obtain the ML estimator for a dynamic decision process Their method also works and has additional advantages for games They note that the nested fixed point approach is really only a special way of finding the solution of a more general constrained optimization problem MPEC (mathematical programming with equilibrium constraints) approach There will often be better algorithms to solve an MPEC problem, which allows for faster and/or more robust computation of the ML estimate compared to the NFP approach 6 / 83
32 mplementation details Let s be precise on the computational steps we take to calculate the likelihood function We calculate the expected value function, w(x; ), in order to derive the choice probabilities which allow us to match model predictions and observations On our computer, we will use some interpolation or approximation method to represent w Representation will depend on a set of parameters, represents the value of w at specific state points (interpolation), or coefficients on basis functions (approximation) On our computer, completely defines w, w, 63 / 83 mplementation details The expected value function satisfies w = (w; ) We can alternatively express this relationship as = ( ; ) Value function iteration on our computer proceeds by computing the sequence (n+1) = ( (n) (0) ; ), given some starting value For each parameter vector, there is a unique that satisfies = ( ; ) Denote this relationship as = ( ) 64 / 83
33 mplementation details Recall how we calculate the likelihood function: Using the expected value function, w,, calculate the choice-specific values, Z v j (x; ) =u j (x; )+ w(x 0 ; )f(x 0 x, j; ) dx 0 Then calculate the CCP s The equation above presumes that we use w = the choice-specific values But we could alternatively use some arbitrary guess for w (w; ) to calculate Let be the parameter vector that summarizes some arbitrary w Let vj (x;, ) be the corresponding choice-specific value function 65 / 83 MPEC approach We can use v j (x;, ) to calculate the CCP s and then the augmented likelihood function, l(q;, ) This expressions clearly shows what the likelihood function depends on: Parameter values The expected value function w, describing how the decision maker behaves Our assumption of rational behavior imposes that the decision maker does not follow some arbitrary decision rule, but rather the rule corresponding to = ( ; ), denoted by = ( ) 66 / 83
34 MPEC approach We can thus express the likelihood estimation problem in its general form: max (, ) log(l(q;, )) s.t. ( ; ) =0 Solve for the parameter vector (, ) The constraint is an equilibrium constraint, thus the term MPEC The nested fixed point algorithm is a special way of formulating this problem: max log(l(q; )) log(l(q;, ( ))) Note that both problem formulations define the same ML estimator 67 / 83 Discussion Benefits of MPEC estimation: Avoid having to find an exact solution of the value function, = ( ; ), for each ) speed advantage The MPEC formulation allows for more robust convergence to the solution, because derivatives (of the objective and constraint) are easier to compute than in the NFP approach 68 / 83
35 Discussion sn t a large state space ) with many elements (say 100,000) an obvious obstacle to using the MPEC approach? Good solvers, such as SNOPT or KNTRO can handle such problems Will require a sparse Jacobian of the constraint, i.e. r ( ; ) needs to be sparse Will require that we are able to compute the Jacobian, r ( ; ), which is an L L matrix, where L is the number of elements in 69 / 83 Discussion How to calculate the Jacobian of the constraint? Analytic derivatives! can be cumbersome and error-prone Automatic differentiation (AD), available in C, C++, FORTRAN, and MATLAB! virtually no extra programming effort Works fine for small-medium scale problems May be difficult to implement (computer speed/memory requirements) given currently available AD software for large-scale problems 70 / 83
36 Overview The Estimation Problem; Structural vs Reduced Form of the Model dentification Nested Fixed Point (NFP) Estimators Unobserved State Variables Example: The Bayesian Learning Model Model Estimation in the Presence of Unobserved State Variables Unobserved Heterogeneity Postscript: Sequential vs. Myopic Learning The MPEC Approach to ML Estimation Example: Estimation of the Durable Goods Adoption Problem Using NFP and MPEC Approaches Other Topics References 71 / 83 Overview Data are generated from the durable goods adoption model with two products discussed in Part of the lecture We solve for w using bilinear interpolation Exercise: Re-write the code using Chebyshev approximation, which is much more efficient in this example 7 / 83
37 Code documentation Main.m Defines model parameters and price process parameters Sets values for interpolator and initializes Gauss-Hermite quadrature information f create_data=1, solves the decision process given parameters and simulates a new data set. Call to simulate_price_process to simulate prices and simulate_adoptions to simulate the corresponding adoption decisions. Plots aggregate adoption data based on the output of calculate_sales Use script display_dp_solution to show choice probabilities and relative choice-specific value functions, v j (x) v 0 (x) 73 / 83 Code documentation Main.m contains three estimation approaches: 1. Use NFP estimator and MATLAB s built-in Nelder-Meade simplex search algorithm. Use the TOMLAB package to find the NFP estimator. Numerical gradients are used 3. Use TOMLAB and the MPEC approach Allows for use of automatic differentiation using the TOMLAB/MAD module TOMLAB can be obtained at (ask for trial license) 74 / 83
38 Code documentation Solution of the optimal decision rule Based on iteration on the integrated value function to solve for the expect value function Bellman_equation_rhs updates the current guess of the expected value function Bellman_equation_rhs uses interpolate_d when taking the expectation of the future value 75 / 83 Code documentation NFP algorithm Calls log_likelihood, which solves for the expected value function to calculate the CCP s MPEC algorithm Calls log_likelihood_augmented with, w supplied as parameters Bellman_equation_constraint implements the equilibrium constraint, ( ; ) =0 76 / 83
39 Overview The Estimation Problem; Structural vs Reduced Form of the Model dentification Nested Fixed Point (NFP) Estimators Unobserved State Variables Example: The Bayesian Learning Model Model Estimation in the Presence of Unobserved State Variables Unobserved Heterogeneity Postscript: Sequential vs. Myopic Learning The MPEC Approach to ML Estimation Example: Estimation of the Durable Goods Adoption Problem Using NFP and MPEC Approaches Other Topics References 77 / 83 Omitted topics This lecture omits two recent estimation approaches: 1. Two-step estimators: Attempt to alleviate the computational burden inherent in NFP and MPEC estimation approaches (e.g. Pesendorfer and Schmidt-Dengler 008). Bayesian estimation using a new algorithm combining MCMC and value function iteration (mai, Jain, and Ching 009; Norets 009) 78 / 83
40 Overview The Estimation Problem; Structural vs Reduced Form of the Model dentification Nested Fixed Point (NFP) Estimators Unobserved State Variables Example: The Bayesian Learning Model Model Estimation in the Presence of Unobserved State Variables Unobserved Heterogeneity Postscript: Sequential vs. Myopic Learning The MPEC Approach to ML Estimation Example: Estimation of the Durable Goods Adoption Problem Using NFP and MPEC Approaches Other Topics References 79 / 83 Estimation and dentification Abbring, J. H. (010): dentification of Dynamic Discrete Choice Models, Annual Review of Economics,, Arcidiacono, P., and P. E. Ellickson (011): Practical Methods for Estimation of Dynamic Discrete Choice Models, Annual Review of Economics, 3, Ackerberg, D. A. (009): "A new use of importance sampling to reduce computational burden in simulation estimation," Quantitative Marketing and Economics, 7(4), Aguirregabiria, V. and Mira, P. (010): Dynamic discrete choice structural models: A survey, Journal of Econometrics, 156, Dubé, J.-P., Hitsch, G. J., and Jindal, P. (013): The Joint dentification of Utility and Discount Functions from Stated Choice Data: An Application to Durable Goods Adoption, manuscript Fang, H., and Y. Wang (013): Estimating Dynamic Discrete Choice Models with Hyperbolic Discounting, with an Application to Mammography Decisions, manuscript Heckman, J. (1981): The ncidental Parameters Problem and the Problem of nitial Conditions in Estimating a Discrete Time-Discrete Data Stochastic Process, in C. Manski and D. L. McFadden (eds.): Structural Analysis of Discrete Data with Econometric Applications. MT Press 80 / 83
41 Estimation and dentification (Cont.) Hotz, J. V., and R. A. Miller (1993): Conditional Choice Probabilities and the Estimation of Dynamic Models, Review of Economic Studies, 60, Magnac, T. and Thesmar, D. (00): dentifying Dynamic Discrete Decision Processes, Econometrica, 70(), Reiss, P. C., and F. A. Wolak (007): Structural Econometric Modeling: Rationales and Examples from ndustrial Organization, in Handbook of Econometrics, Vol. 6A, ed. by J. J. Heckman and E. E. Leamer. Elsevier B. V. Rust, J. (1987): Optimal Replacement of GMC Bus Engines: An Empirical Model of Harold Zurcher, Econometrica, 55 (5), Rust, J. (1994): Structural Estimation of Markov Decision Processes, in Handbook of Econometrics, Vol. V, ed. by R. F. Engle and D. L. McFadden. Amsterdam and New York: North-Holland Su, C.-L., and K. L. Judd (01): Constrained Optimization Approaches to Estimation of Structural Models, Econometrica, 80(5), Yao, S., C. F. Mela, J. Chiang, and Y. Chen (01): Determining Consumers Discount Rates with Field Studies, Journal of Marketing Research, 49(6), / 83 Other Estimation Approaches (Two-Step, Bayesian,...) Aguirregabiria, V., and P. Mira (00): Swapping the Nested Fixed Point Algorithm: A Class of Estimators for Discrete Markov Decision Models, Econometrica, 70 (4), Aguirregabiria, V. and Mira, P. (007): Sequential Estimation of Dynamic Discrete Games, Econometrica, 75(1), 1-53 Arcidiacono, P. and Miller, R. (008): CCP Estimation of Dynamic Discrete Choice Models with Unobserved Heterogeneity, manuscript Bajari, P., L. Benkard, and J. Levin (007): Estimating Dynamic Models of mperfect Competition, Econometrica, 75 (5), mai, S., Jain, N., and Ching, A. (009): Bayesian Estimation of Dynamic Discrete Choice Models, Econometrica, 77(6), Norets, A. (009): nference in Dynamic Discrete Choice Models With Serially Correlated Unobserved State Variables, Econometrica, 77(5), Pesendorfer, M. and Schmidt-Dengler, P. (008): Asymptotic Least Squares Estimators for Dynamic Games, Review of Economic Studies, 75, / 83
42 Other Applications Erdem, T., and M. P. Keane (1996): Decision-Making Under Uncertainty: Capturing Brand Choice Processes in Turbulent Consumer Goods Markets, Marketing Science, 15 (1), 1 0 Hartmann, W. R. (006): ntertemporal Effects of Consumption and Their mplications for Demand Elasticity Estimates, Quantitative Marketing and Economics, 4, Hartmann, W. R. and Viard, V. B. (008): "Do frequency reward programs create switching costs? A dynamic structural analysis of demand in a reward program," Quantitative Marketing and Economics, 6(), Hitsch, G. J. (006): "An Empirical Model of Optimal Dynamic Product Launch and Exit Under Demand Uncertainty," Marketing Science, 5(1), 5-50 Misra, S. and Nair, H. (011): "A Structural Model of Sales-Force Compensation Dynamics: Estimation and Field mplementation," Quantitative Marketing and Economics, 9, / 83
Lecture notes: single-agent dynamics 1
Lecture notes: single-agent dynamics 1 Single-agent dynamic optimization models In these lecture notes we consider specification and estimation of dynamic optimization models. Focus on single-agent models.
STA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! [email protected]! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct
INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition)
INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition) Abstract Indirect inference is a simulation-based method for estimating the parameters of economic models. Its
Lecture notes: single-agent dynamics 1
Lecture notes: single-agent dynamics 1 Single-agent dynamic optimization models In these lecture notes we consider specification and estimation of dynamic optimization models. Focus on single-agent models.
A HYBRID GENETIC ALGORITHM FOR THE MAXIMUM LIKELIHOOD ESTIMATION OF MODELS WITH MULTIPLE EQUILIBRIA: A FIRST REPORT
New Mathematics and Natural Computation Vol. 1, No. 2 (2005) 295 303 c World Scientific Publishing Company A HYBRID GENETIC ALGORITHM FOR THE MAXIMUM LIKELIHOOD ESTIMATION OF MODELS WITH MULTIPLE EQUILIBRIA:
Statistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
Uncovering Consumer Decision Rules under Complex Dynamic Environments: The Case of Coalition Loyalty Programs
Uncovering Consumer Decision Rules under Complex Dynamic Environments: The Case of Coalition Loyalty Programs Andrew Ching Masakazu Ishihara Very Preliminary and Incomplete April 18, 2014 Abstract We propose
The primary goal of this thesis was to understand how the spatial dependence of
5 General discussion 5.1 Introduction The primary goal of this thesis was to understand how the spatial dependence of consumer attitudes can be modeled, what additional benefits the recovering of spatial
Lecture 3: Linear methods for classification
Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,
On the Efficiency of Competitive Stock Markets Where Traders Have Diverse Information
Finance 400 A. Penati - G. Pennacchi Notes on On the Efficiency of Competitive Stock Markets Where Traders Have Diverse Information by Sanford Grossman This model shows how the heterogeneous information
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written
1 Short Introduction to Time Series
ECONOMICS 7344, Spring 202 Bent E. Sørensen January 24, 202 Short Introduction to Time Series A time series is a collection of stochastic variables x,.., x t,.., x T indexed by an integer value t. The
Multivariate Normal Distribution
Multivariate Normal Distribution Lecture 4 July 21, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Lecture #4-7/21/2011 Slide 1 of 41 Last Time Matrices and vectors Eigenvalues
Moral Hazard. Itay Goldstein. Wharton School, University of Pennsylvania
Moral Hazard Itay Goldstein Wharton School, University of Pennsylvania 1 Principal-Agent Problem Basic problem in corporate finance: separation of ownership and control: o The owners of the firm are typically
Marketing Mix Modelling and Big Data P. M Cain
1) Introduction Marketing Mix Modelling and Big Data P. M Cain Big data is generally defined in terms of the volume and variety of structured and unstructured information. Whereas structured data is stored
Mathematics Course 111: Algebra I Part IV: Vector Spaces
Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 1996-7 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are
Choice under Uncertainty
Choice under Uncertainty Part 1: Expected Utility Function, Attitudes towards Risk, Demand for Insurance Slide 1 Choice under Uncertainty We ll analyze the underlying assumptions of expected utility theory
Applications to Data Smoothing and Image Processing I
Applications to Data Smoothing and Image Processing I MA 348 Kurt Bryan Signals and Images Let t denote time and consider a signal a(t) on some time interval, say t. We ll assume that the signal a(t) is
Linear Programming Notes V Problem Transformations
Linear Programming Notes V Problem Transformations 1 Introduction Any linear programming problem can be rewritten in either of two standard forms. In the first form, the objective is to maximize, the material
CHAPTER 2 Estimating Probabilities
CHAPTER 2 Estimating Probabilities Machine Learning Copyright c 2016. Tom M. Mitchell. All rights reserved. *DRAFT OF January 24, 2016* *PLEASE DO NOT DISTRIBUTE WITHOUT AUTHOR S PERMISSION* This is a
Logistic Regression (1/24/13)
STA63/CBB540: Statistical methods in computational biology Logistic Regression (/24/3) Lecturer: Barbara Engelhardt Scribe: Dinesh Manandhar Introduction Logistic regression is model for regression used
Linear Threshold Units
Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear
STATISTICA Formula Guide: Logistic Regression. Table of Contents
: Table of Contents... 1 Overview of Model... 1 Dispersion... 2 Parameterization... 3 Sigma-Restricted Model... 3 Overparameterized Model... 4 Reference Coding... 4 Model Summary (Summary Tab)... 5 Summary
Logistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression
Logistic Regression Department of Statistics The Pennsylvania State University Email: [email protected] Logistic Regression Preserve linear classification boundaries. By the Bayes rule: Ĝ(x) = arg max
Inner Product Spaces
Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and
FIXED EFFECTS AND RELATED ESTIMATORS FOR CORRELATED RANDOM COEFFICIENT AND TREATMENT EFFECT PANEL DATA MODELS
FIXED EFFECTS AND RELATED ESTIMATORS FOR CORRELATED RANDOM COEFFICIENT AND TREATMENT EFFECT PANEL DATA MODELS Jeffrey M. Wooldridge Department of Economics Michigan State University East Lansing, MI 48824-1038
IDENTIFICATION IN A CLASS OF NONPARAMETRIC SIMULTANEOUS EQUATIONS MODELS. Steven T. Berry and Philip A. Haile. March 2011 Revised April 2011
IDENTIFICATION IN A CLASS OF NONPARAMETRIC SIMULTANEOUS EQUATIONS MODELS By Steven T. Berry and Philip A. Haile March 2011 Revised April 2011 COWLES FOUNDATION DISCUSSION PAPER NO. 1787R COWLES FOUNDATION
ECON20310 LECTURE SYNOPSIS REAL BUSINESS CYCLE
ECON20310 LECTURE SYNOPSIS REAL BUSINESS CYCLE YUAN TIAN This synopsis is designed merely for keep a record of the materials covered in lectures. Please refer to your own lecture notes for all proofs.
Inference of Probability Distributions for Trust and Security applications
Inference of Probability Distributions for Trust and Security applications Vladimiro Sassone Based on joint work with Mogens Nielsen & Catuscia Palamidessi Outline 2 Outline Motivations 2 Outline Motivations
Introduction to Logistic Regression
OpenStax-CNX module: m42090 1 Introduction to Logistic Regression Dan Calderon This work is produced by OpenStax-CNX and licensed under the Creative Commons Attribution License 3.0 Abstract Gives introduction
The VAR models discussed so fare are appropriate for modeling I(0) data, like asset returns or growth rates of macroeconomic time series.
Cointegration The VAR models discussed so fare are appropriate for modeling I(0) data, like asset returns or growth rates of macroeconomic time series. Economic theory, however, often implies equilibrium
We shall turn our attention to solving linear systems of equations. Ax = b
59 Linear Algebra We shall turn our attention to solving linear systems of equations Ax = b where A R m n, x R n, and b R m. We already saw examples of methods that required the solution of a linear system
Financial Time Series Analysis (FTSA) Lecture 1: Introduction
Financial Time Series Analysis (FTSA) Lecture 1: Introduction Brief History of Time Series Analysis Statistical analysis of time series data (Yule, 1927) v/s forecasting (even longer). Forecasting is often
PS 271B: Quantitative Methods II. Lecture Notes
PS 271B: Quantitative Methods II Lecture Notes Langche Zeng [email protected] The Empirical Research Process; Fundamental Methodological Issues 2 Theory; Data; Models/model selection; Estimation; Inference.
3 Introduction to Assessing Risk
3 Introduction to Assessing Risk Important Question. How do we assess the risk an investor faces when choosing among assets? In this discussion we examine how an investor would assess the risk associated
Betting Strategies, Market Selection, and the Wisdom of Crowds
Betting Strategies, Market Selection, and the Wisdom of Crowds Willemien Kets Northwestern University [email protected] David M. Pennock Microsoft Research New York City [email protected]
General Framework for an Iterative Solution of Ax b. Jacobi s Method
2.6 Iterative Solutions of Linear Systems 143 2.6 Iterative Solutions of Linear Systems Consistent linear systems in real life are solved in one of two ways: by direct calculation (using a matrix factorization,
Notes V General Equilibrium: Positive Theory. 1 Walrasian Equilibrium and Excess Demand
Notes V General Equilibrium: Positive Theory In this lecture we go on considering a general equilibrium model of a private ownership economy. In contrast to the Notes IV, we focus on positive issues such
BayesX - Software for Bayesian Inference in Structured Additive Regression
BayesX - Software for Bayesian Inference in Structured Additive Regression Thomas Kneib Faculty of Mathematics and Economics, University of Ulm Department of Statistics, Ludwig-Maximilians-University Munich
Statistics in Retail Finance. Chapter 6: Behavioural models
Statistics in Retail Finance 1 Overview > So far we have focussed mainly on application scorecards. In this chapter we shall look at behavioural models. We shall cover the following topics:- Behavioural
Basics of Statistical Machine Learning
CS761 Spring 2013 Advanced Machine Learning Basics of Statistical Machine Learning Lecturer: Xiaojin Zhu [email protected] Modern machine learning is rooted in statistics. You will find many familiar
(Quasi-)Newton methods
(Quasi-)Newton methods 1 Introduction 1.1 Newton method Newton method is a method to find the zeros of a differentiable non-linear function g, x such that g(x) = 0, where g : R n R n. Given a starting
Nonparametric adaptive age replacement with a one-cycle criterion
Nonparametric adaptive age replacement with a one-cycle criterion P. Coolen-Schrijner, F.P.A. Coolen Department of Mathematical Sciences University of Durham, Durham, DH1 3LE, UK e-mail: [email protected]
HETEROGENEOUS AGENTS AND AGGREGATE UNCERTAINTY. Daniel Harenberg [email protected]. University of Mannheim. Econ 714, 28.11.
COMPUTING EQUILIBRIUM WITH HETEROGENEOUS AGENTS AND AGGREGATE UNCERTAINTY (BASED ON KRUEGER AND KUBLER, 2004) Daniel Harenberg [email protected] University of Mannheim Econ 714, 28.11.06 Daniel Harenberg
Master s Theory Exam Spring 2006
Spring 2006 This exam contains 7 questions. You should attempt them all. Each question is divided into parts to help lead you through the material. You should attempt to complete as much of each problem
Detection of changes in variance using binary segmentation and optimal partitioning
Detection of changes in variance using binary segmentation and optimal partitioning Christian Rohrbeck Abstract This work explores the performance of binary segmentation and optimal partitioning in the
Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay
Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 17 Shannon-Fano-Elias Coding and Introduction to Arithmetic Coding
PUTNAM TRAINING POLYNOMIALS. Exercises 1. Find a polynomial with integral coefficients whose zeros include 2 + 5.
PUTNAM TRAINING POLYNOMIALS (Last updated: November 17, 2015) Remark. This is a list of exercises on polynomials. Miguel A. Lerma Exercises 1. Find a polynomial with integral coefficients whose zeros include
11. Time series and dynamic linear models
11. Time series and dynamic linear models Objective To introduce the Bayesian approach to the modeling and forecasting of time series. Recommended reading West, M. and Harrison, J. (1997). models, (2 nd
OPTIMAL DESIGN OF A MULTITIER REWARD SCHEME. Amir Gandomi *, Saeed Zolfaghari **
OPTIMAL DESIGN OF A MULTITIER REWARD SCHEME Amir Gandomi *, Saeed Zolfaghari ** Department of Mechanical and Industrial Engineering, Ryerson University, Toronto, Ontario * Tel.: + 46 979 5000x7702, Email:
Factor analysis. Angela Montanari
Factor analysis Angela Montanari 1 Introduction Factor analysis is a statistical model that allows to explain the correlations between a large number of observed correlated variables through a small number
Introduction to time series analysis
Introduction to time series analysis Margherita Gerolimetto November 3, 2010 1 What is a time series? A time series is a collection of observations ordered following a parameter that for us is time. Examples
The Basics of Graphical Models
The Basics of Graphical Models David M. Blei Columbia University October 3, 2015 Introduction These notes follow Chapter 2 of An Introduction to Probabilistic Graphical Models by Michael Jordan. Many figures
How To Find Out What Search Strategy Is Used In The U.S. Auto Insurance Industry
Simultaneous or Sequential? Search Strategies in the U.S. Auto Insurance Industry Current version: April 2014 Elisabeth Honka Pradeep Chintagunta Abstract We show that the search method consumers use when
APPLIED MISSING DATA ANALYSIS
APPLIED MISSING DATA ANALYSIS Craig K. Enders Series Editor's Note by Todd D. little THE GUILFORD PRESS New York London Contents 1 An Introduction to Missing Data 1 1.1 Introduction 1 1.2 Chapter Overview
Models for Product Demand Forecasting with the Use of Judgmental Adjustments to Statistical Forecasts
Page 1 of 20 ISF 2008 Models for Product Demand Forecasting with the Use of Judgmental Adjustments to Statistical Forecasts Andrey Davydenko, Professor Robert Fildes [email protected] Lancaster
2DI36 Statistics. 2DI36 Part II (Chapter 7 of MR)
2DI36 Statistics 2DI36 Part II (Chapter 7 of MR) What Have we Done so Far? Last time we introduced the concept of a dataset and seen how we can represent it in various ways But, how did this dataset came
Chapter 21: The Discounted Utility Model
Chapter 21: The Discounted Utility Model 21.1: Introduction This is an important chapter in that it introduces, and explores the implications of, an empirically relevant utility function representing intertemporal
Lecture 2: Universality
CS 710: Complexity Theory 1/21/2010 Lecture 2: Universality Instructor: Dieter van Melkebeek Scribe: Tyson Williams In this lecture, we introduce the notion of a universal machine, develop efficient universal
Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus
Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus Tihomir Asparouhov and Bengt Muthén Mplus Web Notes: No. 15 Version 8, August 5, 2014 1 Abstract This paper discusses alternatives
Econometrics Simple Linear Regression
Econometrics Simple Linear Regression Burcu Eke UC3M Linear equations with one variable Recall what a linear equation is: y = b 0 + b 1 x is a linear equation with one variable, or equivalently, a straight
Chapter 6: Multivariate Cointegration Analysis
Chapter 6: Multivariate Cointegration Analysis 1 Contents: Lehrstuhl für Department Empirische of Wirtschaftsforschung Empirical Research and und Econometrics Ökonometrie VI. Multivariate Cointegration
Lecture 8: Signal Detection and Noise Assumption
ECE 83 Fall Statistical Signal Processing instructor: R. Nowak, scribe: Feng Ju Lecture 8: Signal Detection and Noise Assumption Signal Detection : X = W H : X = S + W where W N(, σ I n n and S = [s, s,...,
Joint Product Signals of Quality. James E. McClure and Lee C. Spector. Ball State University Muncie, Indiana. September 1991.
Joint Product Signals of Quality By James E. McClure and Lee C. Spector Ball State University Muncie, Indiana September 1991 Abstract In the absence of other information about the quality of an experience
Notes on Factoring. MA 206 Kurt Bryan
The General Approach Notes on Factoring MA 26 Kurt Bryan Suppose I hand you n, a 2 digit integer and tell you that n is composite, with smallest prime factor around 5 digits. Finding a nontrivial factor
C(t) (1 + y) 4. t=1. For the 4 year bond considered above, assume that the price today is 900$. The yield to maturity will then be the y that solves
Economics 7344, Spring 2013 Bent E. Sørensen INTEREST RATE THEORY We will cover fixed income securities. The major categories of long-term fixed income securities are federal government bonds, corporate
Hedging Illiquid FX Options: An Empirical Analysis of Alternative Hedging Strategies
Hedging Illiquid FX Options: An Empirical Analysis of Alternative Hedging Strategies Drazen Pesjak Supervised by A.A. Tsvetkov 1, D. Posthuma 2 and S.A. Borovkova 3 MSc. Thesis Finance HONOURS TRACK Quantitative
Monte Carlo Methods in Finance
Author: Yiyang Yang Advisor: Pr. Xiaolin Li, Pr. Zari Rachev Department of Applied Mathematics and Statistics State University of New York at Stony Brook October 2, 2012 Outline Introduction 1 Introduction
Math 4310 Handout - Quotient Vector Spaces
Math 4310 Handout - Quotient Vector Spaces Dan Collins The textbook defines a subspace of a vector space in Chapter 4, but it avoids ever discussing the notion of a quotient space. This is understandable
Dirichlet Processes A gentle tutorial
Dirichlet Processes A gentle tutorial SELECT Lab Meeting October 14, 2008 Khalid El-Arini Motivation We are given a data set, and are told that it was generated from a mixture of Gaussian distributions.
LOGISTIC REGRESSION. Nitin R Patel. where the dependent variable, y, is binary (for convenience we often code these values as
LOGISTIC REGRESSION Nitin R Patel Logistic regression extends the ideas of multiple linear regression to the situation where the dependent variable, y, is binary (for convenience we often code these values
Analysis of Bayesian Dynamic Linear Models
Analysis of Bayesian Dynamic Linear Models Emily M. Casleton December 17, 2010 1 Introduction The main purpose of this project is to explore the Bayesian analysis of Dynamic Linear Models (DLMs). The main
Probabilistic Models for Big Data. Alex Davies and Roger Frigola University of Cambridge 13th February 2014
Probabilistic Models for Big Data Alex Davies and Roger Frigola University of Cambridge 13th February 2014 The State of Big Data Why probabilistic models for Big Data? 1. If you don t have to worry about
A linear algebraic method for pricing temporary life annuities
A linear algebraic method for pricing temporary life annuities P. Date (joint work with R. Mamon, L. Jalen and I.C. Wang) Department of Mathematical Sciences, Brunel University, London Outline Introduction
Adaptive Online Gradient Descent
Adaptive Online Gradient Descent Peter L Bartlett Division of Computer Science Department of Statistics UC Berkeley Berkeley, CA 94709 bartlett@csberkeleyedu Elad Hazan IBM Almaden Research Center 650
NEW YORK STATE TEACHER CERTIFICATION EXAMINATIONS
NEW YORK STATE TEACHER CERTIFICATION EXAMINATIONS TEST DESIGN AND FRAMEWORK September 2014 Authorized for Distribution by the New York State Education Department This test design and framework document
The HB. How Bayesian methods have changed the face of marketing research. Summer 2004
The HB How Bayesian methods have changed the face of marketing research. 20 Summer 2004 Reprinted with permission from Marketing Research, Summer 2004, published by the American Marketing Association.
Random graphs with a given degree sequence
Sourav Chatterjee (NYU) Persi Diaconis (Stanford) Allan Sly (Microsoft) Let G be an undirected simple graph on n vertices. Let d 1,..., d n be the degrees of the vertices of G arranged in descending order.
Statistics Graduate Courses
Statistics Graduate Courses STAT 7002--Topics in Statistics-Biological/Physical/Mathematics (cr.arr.).organized study of selected topics. Subjects and earnable credit may vary from semester to semester.
Model-based Synthesis. Tony O Hagan
Model-based Synthesis Tony O Hagan Stochastic models Synthesising evidence through a statistical model 2 Evidence Synthesis (Session 3), Helsinki, 28/10/11 Graphical modelling The kinds of models that
MATLAB and Big Data: Illustrative Example
MATLAB and Big Data: Illustrative Example Rick Mansfield Cornell University August 19, 2014 Goals Use a concrete example from my research to: Demonstrate the value of vectorization Introduce key commands/functions
Machine Learning and Pattern Recognition Logistic Regression
Machine Learning and Pattern Recognition Logistic Regression Course Lecturer:Amos J Storkey Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh Crichton Street,
Using simulation to calculate the NPV of a project
Using simulation to calculate the NPV of a project Marius Holtan Onward Inc. 5/31/2002 Monte Carlo simulation is fast becoming the technology of choice for evaluating and analyzing assets, be it pure financial
The Method of Least Squares
The Method of Least Squares Steven J. Miller Mathematics Department Brown University Providence, RI 0292 Abstract The Method of Least Squares is a procedure to determine the best fit line to data; the
Big Data - Lecture 1 Optimization reminders
Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Schedule Introduction Major issues Examples Mathematics
Should we Really Care about Building Business. Cycle Coincident Indexes!
Should we Really Care about Building Business Cycle Coincident Indexes! Alain Hecq University of Maastricht The Netherlands August 2, 2004 Abstract Quite often, the goal of the game when developing new
(Refer Slide Time: 1:42)
Introduction to Computer Graphics Dr. Prem Kalra Department of Computer Science and Engineering Indian Institute of Technology, Delhi Lecture - 10 Curves So today we are going to have a new topic. So far
In recent years, Federal Reserve (Fed) policymakers have come to rely
Long-Term Interest Rates and Inflation: A Fisherian Approach Peter N. Ireland In recent years, Federal Reserve (Fed) policymakers have come to rely on long-term bond yields to measure the public s long-term
ON THE ROBUSTNESS OF FIXED EFFECTS AND RELATED ESTIMATORS IN CORRELATED RANDOM COEFFICIENT PANEL DATA MODELS
ON THE ROBUSTNESS OF FIXED EFFECTS AND RELATED ESTIMATORS IN CORRELATED RANDOM COEFFICIENT PANEL DATA MODELS Jeffrey M. Wooldridge THE INSTITUTE FOR FISCAL STUDIES DEPARTMENT OF ECONOMICS, UCL cemmap working
ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2015
ECON 459 Game Theory Lecture Notes Auctions Luca Anderlini Spring 2015 These notes have been used before. If you can still spot any errors or have any suggestions for improvement, please let me know. 1
Asian Option Pricing Formula for Uncertain Financial Market
Sun and Chen Journal of Uncertainty Analysis and Applications (215) 3:11 DOI 1.1186/s4467-15-35-7 RESEARCH Open Access Asian Option Pricing Formula for Uncertain Financial Market Jiajun Sun 1 and Xiaowei
Introduction to General and Generalized Linear Models
Introduction to General and Generalized Linear Models General Linear Models - part I Henrik Madsen Poul Thyregod Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby
Chapter 7. BANDIT PROBLEMS.
Chapter 7. BANDIT PROBLEMS. Bandit problems are problems in the area of sequential selection of experiments, and they are related to stopping rule problems through the theorem of Gittins and Jones (974).
Missing Data: Part 1 What to Do? Carol B. Thompson Johns Hopkins Biostatistics Center SON Brown Bag 3/20/13
Missing Data: Part 1 What to Do? Carol B. Thompson Johns Hopkins Biostatistics Center SON Brown Bag 3/20/13 Overview Missingness and impact on statistical analysis Missing data assumptions/mechanisms Conventional
CPC/CPA Hybrid Bidding in a Second Price Auction
CPC/CPA Hybrid Bidding in a Second Price Auction Benjamin Edelman Hoan Soo Lee Working Paper 09-074 Copyright 2008 by Benjamin Edelman and Hoan Soo Lee Working papers are in draft form. This working paper
