Bayesian analysis of non-gaussian Long-Range Dependent processes
|
|
- Kelly Burke
- 7 years ago
- Views:
Transcription
1 Bayesian analysis of non-gaussian Long-Range Dependent processes Tim Graves 1, Christian Franzke 2, Nick Watkins 2, and Bobby Gramacy 3 Statistical Laboratory, University of Cambridge, Cambridge, UK British Antarctic Survey, Cambridge, UK Booth School of Business, The University of Chicago, Chicago, USA May 18, 2012
2 Outline 1 Introduction 2 Exact Bayesian analysis for Gaussian case 3 Approximate Bayesian analysis for general case
3 Outline 1 Introduction 2 Exact Bayesian analysis for Gaussian case 3 Approximate Bayesian analysis for general case
4 Long Range Dependence Definition A Long Range Dependent process is a stationary process for which k= ρ(k) =.
5 Long Range Dependence Definition A Long Range Dependent process is a stationary process for which k= ρ(k) =.... [T]he stationary long memory processes form a layer among the stationary processes that is near the boundary with non-stationary processes, or, alternatively, as the layer separating the non-stationary processes from the usual stationary processes. [Samorodnitsky, 2006]
6 Long Range Dependence Definition A Long Range Dependent process is a stationary process for which ρ(k) ck 2d 1, 0 < d < 1 2.
7 Long Range Dependence Definition A Long Range Dependent process is a stationary process for which ρ(k) ck 2d 1, 0 < d < 1 2. Definition A Long Range Dependent process is a stationary process for which X t = ψ k ɛ t k, k=0 ψ k ck d 1, 0 < d < 1 2.
8 ARFIMA processes Definition A process {X t } is an ARFIMA(p, d, q) process if it is the solution to: Φ(B)(1 B) d X t = Θ(B)ε t, p where Φ(z) = 1 + φ j z j and Θ(z) = 1 + j=1 q θ j z j, and the innovations {ε t } are iid with 0 mean and variance σ 2 <. We say that {X t } is an ARFIMA(p, d, q) process with mean µ, if {X t µ} is an ARFIMA(p, d, q) process. j=1
9 ARFIMA parameters µ location parameter σ scale parameter d long memory parameter (long memory process iff 0 < d < 0.5 ) φ p-dimensional short memory parameter θ q-dimensional short memory parameter Which parameters are of interest? When considering long memory processes, we are usually primarily interested in the parameter d (and possibly µ). The parameters σ, φ, θ (and even p, q) are essentially nuisance parameters.
10 ARFIMA parameters µ location parameter σ scale parameter d long memory parameter (long memory process iff 0 < d < 0.5 ) φ p-dimensional short memory parameter θ q-dimensional short memory parameter Which parameters are of interest? When considering long memory processes, we are usually primarily interested in the parameter d (and possibly µ). The parameters σ, φ, θ (and even p, q) are essentially nuisance parameters.
11 ARFIMA parameters µ location parameter σ scale parameter d long memory parameter (long memory process iff 0 < d < 0.5 ) φ p-dimensional short memory parameter θ q-dimensional short memory parameter Which parameters are of interest? When considering long memory processes, we are usually primarily interested in the parameter d (and possibly µ). The parameters σ, φ, θ (and even p, q) are essentially nuisance parameters.
12 ARFIMA parameters µ location parameter σ scale parameter d long memory parameter (long memory process iff 0 < d < 0.5 ) φ p-dimensional short memory parameter θ q-dimensional short memory parameter Which parameters are of interest? When considering long memory processes, we are usually primarily interested in the parameter d (and possibly µ). The parameters σ, φ, θ (and even p, q) are essentially nuisance parameters.
13 ARFIMA parameters µ location parameter σ scale parameter d long memory parameter (long memory process iff 0 < d < 0.5 ) φ p-dimensional short memory parameter θ q-dimensional short memory parameter Which parameters are of interest? When considering long memory processes, we are usually primarily interested in the parameter d (and possibly µ). The parameters σ, φ, θ (and even p, q) are essentially nuisance parameters.
14 Outline 1 Introduction 2 Exact Bayesian analysis for Gaussian case 3 Approximate Bayesian analysis for general case
15 Outline of problem Assume Gaussian distribution for the innovations Bayesian: use flat priors for µ, log(σ), and d...
16 Outline of problem Assume Gaussian distribution for the innovations Bayesian: use flat priors for µ, log(σ), and d but can use any set of (independent) priors if desired.
17 Outline of problem Assume Gaussian distribution for the innovations Bayesian: use flat priors for µ, log(σ), and d but can use any set of (independent) priors if desired. Even assuming Gaussianity, the likelihood for d is very complex impossible to find analytic posterior
18 Outline of problem Assume Gaussian distribution for the innovations Bayesian: use flat priors for µ, log(σ), and d but can use any set of (independent) priors if desired. Even assuming Gaussianity, the likelihood for d is very complex impossible to find analytic posterior Must resort to MCMC methods in order to obtain samples from the posterior
19 Outline of problem Assume Gaussian distribution for the innovations Bayesian: use flat priors for µ, log(σ), and d but can use any set of (independent) priors if desired. Even assuming Gaussianity, the likelihood for d is very complex impossible to find analytic posterior Must resort to MCMC methods in order to obtain samples from the posterior Don t want to assume form of short memory (i.e. p, q) must use Reversible-Jump (RJ) MCMC [Green, 1995]
20 Outline of method Re-parameterisation of model to enforce stationarity constraints on Φ and Θ
21 Outline of method Re-parameterisation of model to enforce stationarity constraints on Φ and Θ Efficient calculation of Gaussian likelihood (long memory correlation structure prevents use of standard quick methods)
22 Outline of method Re-parameterisation of model to enforce stationarity constraints on Φ and Θ Efficient calculation of Gaussian likelihood (long memory correlation structure prevents use of standard quick methods) Necessary use of Metropolis Hastings algorithm requires careful selection of proposal distributions
23 Outline of method Re-parameterisation of model to enforce stationarity constraints on Φ and Θ Efficient calculation of Gaussian likelihood (long memory correlation structure prevents use of standard quick methods) Necessary use of Metropolis Hastings algorithm requires careful selection of proposal distributions Correlation between parameters (e.g. φ and d) requires blocking.
24 Example: Pure Gaussian Long Range Dependence (1 B) 0.25 X t = ε t
25 Example: Pure Gaussian Long Range Dependence Density estimate of π(d) d
26 Example: Pure Gaussian Long Range Dependence Density estimate of π(d) d Similarly good results for µ and σ The posterior model probability for the (0, d, 0) model was 70%
27 Example: Corrupted Gaussian Long Range Dependence ( B)(1 B) 0.25 X t = ( B)ε t
28 Example: Corrupted Gaussian Long Range Dependence Density estimate of π(d) d
29 Example: Corrupted Gaussian Long Range Dependence Density estimate of π(d) d The posterior model probability for the (1, d, 1) model was 77% The posterior model probability for the (0, d, 0) model was 0%
30 Dependence of posterior variance on n
31 Dependence of posterior variance on n σ d σ d n n
32 Dependence of posterior variance on n σ d σ d n n σ d n 1/2
33 Outline 1 Introduction 2 Exact Bayesian analysis for Gaussian case 3 Approximate Bayesian analysis for general case
34 Assumptions and general method Drop the Gaussianity assumption Replace with a more general distribution (e.g. α-stable) Seek joint inference about d and α
35 Assumptions and general method Drop the Gaussianity assumption Replace with a more general distribution (e.g. α-stable) Seek joint inference about d and α Initially (for simplicity) we assume no short memory, i.e. we assume a (0, d, 0) model
36 Assumptions and general method Drop the Gaussianity assumption Replace with a more general distribution (e.g. α-stable) Seek joint inference about d and α Initially (for simplicity) we assume no short memory, i.e. we assume a (0, d, 0) model Infinite variance means that auto-covariance approach is no longer sound
37 Assumptions and general method Drop the Gaussianity assumption Replace with a more general distribution (e.g. α-stable) Seek joint inference about d and α Initially (for simplicity) we assume no short memory, i.e. we assume a (0, d, 0) model Infinite variance means that auto-covariance approach is no longer sound Lack of closed form for α-stable density implies lack of closed form for likelihood
38 Solution Approximate the long memory process as a very high order AR process
39 Solution Approximate the long memory process as a very high order AR process Construct the likelihood sequentially and evaluate using specialised efficient methods
40 Solution Approximate the long memory process as a very high order AR process Construct the likelihood sequentially and evaluate using specialised efficient methods f (x 1,..., x t H) = f (x t x t 1,..., x 1, H)f (x t 1,..., x 1 H) where H is the finite recent history of the process x 0, x 1,..., x n
41 Solution Approximate the long memory process as a very high order AR process Construct the likelihood sequentially and evaluate using specialised efficient methods f (x 1,..., x t H) = f (x t x t 1,..., x 1, H)f (x t 1,..., x 1 H) where H is the finite recent history of the process x 0, x 1,..., x n Use auxiliary variables to integrate out the (unknown) history H In practice, setting H = x,..., x suffices, providing enormous computational saving.
42 Example: Pure symmetric α-stable long memory (1 B) 0.15 X t = ε t, α = 1.5
43 Example: Pure symmetric α-stable long memory Density estimate of π(d) d
44 Example: Pure symmetric α-stable long memory Density estimate of π(α) α
45 Example: Pure symmetric α-stable long memory α d
46 Example: Pure symmetric α-stable long memory α d Good estimation of all parameters The posteriors of d and α are independent
47 Dependence of posterior variance on n
48 Dependence of posterior variance on n σ d 5e 04 1e 03 2e 03 5e 03 1e 02 2e 02 5e 02 1e n σ d n 1/2
49 Example: Pure asymmetric α-stable long memory (1 B) 0.1 X t = ε t, α = 1.5 β = 0.5
50 Example: Pure asymmetric α-stable long memory Density estimate of π(β) β
51 Example: Pure asymmetric α-stable long memory Density estimate of π(β) β Good estimation of all other parameters
52 References Green, P. J. (1995). Reversible jump Markov chain Monte Carlo computation and Bayesian model determination. Biometrika, 82(4), Samorodnitsky, G. (2006). Long range dependence. Foundations and Trends in Stochastic Systems, 1(3),
11. Time series and dynamic linear models
11. Time series and dynamic linear models Objective To introduce the Bayesian approach to the modeling and forecasting of time series. Recommended reading West, M. and Harrison, J. (1997). models, (2 nd
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct
More informationTutorial on Markov Chain Monte Carlo
Tutorial on Markov Chain Monte Carlo Kenneth M. Hanson Los Alamos National Laboratory Presented at the 29 th International Workshop on Bayesian Inference and Maximum Entropy Methods in Science and Technology,
More informationDirichlet Processes A gentle tutorial
Dirichlet Processes A gentle tutorial SELECT Lab Meeting October 14, 2008 Khalid El-Arini Motivation We are given a data set, and are told that it was generated from a mixture of Gaussian distributions.
More informationInference on Phase-type Models via MCMC
Inference on Phase-type Models via MCMC with application to networks of repairable redundant systems Louis JM Aslett and Simon P Wilson Trinity College Dublin 28 th June 202 Toy Example : Redundant Repairable
More informationBayesian Time Series Analysis
Bayesian Time Series Analysis Mark Steel, University of Warwick Abstract This article describes the use of Bayesian methods in the statistical analysis of time series. The use of Markov chain Monte Carlo
More informationMonte Carlo-based statistical methods (MASM11/FMS091)
Monte Carlo-based statistical methods (MASM11/FMS091) Jimmy Olsson Centre for Mathematical Sciences Lund University, Sweden Lecture 5 Sequential Monte Carlo methods I February 5, 2013 J. Olsson Monte Carlo-based
More informationLecture 4: Seasonal Time Series, Trend Analysis & Component Model Bus 41910, Time Series Analysis, Mr. R. Tsay
Lecture 4: Seasonal Time Series, Trend Analysis & Component Model Bus 41910, Time Series Analysis, Mr. R. Tsay Business cycle plays an important role in economics. In time series analysis, business cycle
More informationProbabilistic Models for Big Data. Alex Davies and Roger Frigola University of Cambridge 13th February 2014
Probabilistic Models for Big Data Alex Davies and Roger Frigola University of Cambridge 13th February 2014 The State of Big Data Why probabilistic models for Big Data? 1. If you don t have to worry about
More informationChapter 10 Introduction to Time Series Analysis
Chapter 1 Introduction to Time Series Analysis A time series is a collection of observations made sequentially in time. Examples are daily mortality counts, particulate air pollution measurements, and
More informationMonte Carlo and Empirical Methods for Stochastic Inference (MASM11/FMS091)
Monte Carlo and Empirical Methods for Stochastic Inference (MASM11/FMS091) Magnus Wiktorsson Centre for Mathematical Sciences Lund University, Sweden Lecture 5 Sequential Monte Carlo methods I February
More informationIntroduction to Monte Carlo. Astro 542 Princeton University Shirley Ho
Introduction to Monte Carlo Astro 542 Princeton University Shirley Ho Agenda Monte Carlo -- definition, examples Sampling Methods (Rejection, Metropolis, Metropolis-Hasting, Exact Sampling) Markov Chains
More informationLinear Models for Classification
Linear Models for Classification Sumeet Agarwal, EEL709 (Most figures from Bishop, PRML) Approaches to classification Discriminant function: Directly assigns each data point x to a particular class Ci
More informationBayesian Statistics: Indian Buffet Process
Bayesian Statistics: Indian Buffet Process Ilker Yildirim Department of Brain and Cognitive Sciences University of Rochester Rochester, NY 14627 August 2012 Reference: Most of the material in this note
More informationIntroduction to Markov Chain Monte Carlo
1 Introduction to Markov Chain Monte Carlo Charles J. Geyer 1.1 History Despite a few notable uses of simulation of random processes in the pre-computer era (Hammersley and Handscomb, 1964, Section 1.2;
More informationIncorporating cost in Bayesian Variable Selection, with application to cost-effective measurement of quality of health care.
Incorporating cost in Bayesian Variable Selection, with application to cost-effective measurement of quality of health care University of Florida 10th Annual Winter Workshop: Bayesian Model Selection and
More informationParallelization Strategies for Multicore Data Analysis
Parallelization Strategies for Multicore Data Analysis Wei-Chen Chen 1 Russell Zaretzki 2 1 University of Tennessee, Dept of EEB 2 University of Tennessee, Dept. Statistics, Operations, and Management
More informationReal-Time Monitoring of Complex Industrial Processes with Particle Filters
Real-Time Monitoring of Complex Industrial Processes with Particle Filters Rubén Morales-Menéndez Dept. of Mechatronics and Automation ITESM campus Monterrey Monterrey, NL México rmm@itesm.mx Nando de
More informationPS 271B: Quantitative Methods II. Lecture Notes
PS 271B: Quantitative Methods II Lecture Notes Langche Zeng zeng@ucsd.edu The Empirical Research Process; Fundamental Methodological Issues 2 Theory; Data; Models/model selection; Estimation; Inference.
More informationHow To Solve A Sequential Mca Problem
Monte Carlo-based statistical methods (MASM11/FMS091) Jimmy Olsson Centre for Mathematical Sciences Lund University, Sweden Lecture 6 Sequential Monte Carlo methods II February 3, 2012 Changes in HA1 Problem
More informationAusterity in MCMC Land: Cutting the Metropolis-Hastings Budget
Anoop Korattikara AKORATTI@UCI.EDU School of Information & Computer Sciences, University of California, Irvine, CA 92617, USA Yutian Chen YUTIAN.CHEN@ENG.CAM.EDU Department of Engineering, University of
More informationModel-based Synthesis. Tony O Hagan
Model-based Synthesis Tony O Hagan Stochastic models Synthesising evidence through a statistical model 2 Evidence Synthesis (Session 3), Helsinki, 28/10/11 Graphical modelling The kinds of models that
More informationAn Application of Inverse Reinforcement Learning to Medical Records of Diabetes Treatment
An Application of Inverse Reinforcement Learning to Medical Records of Diabetes Treatment Hideki Asoh 1, Masanori Shiro 1 Shotaro Akaho 1, Toshihiro Kamishima 1, Koiti Hasida 1, Eiji Aramaki 2, and Takahide
More informationImperfect Debugging in Software Reliability
Imperfect Debugging in Software Reliability Tevfik Aktekin and Toros Caglar University of New Hampshire Peter T. Paul College of Business and Economics Department of Decision Sciences and United Health
More information1 Short Introduction to Time Series
ECONOMICS 7344, Spring 202 Bent E. Sørensen January 24, 202 Short Introduction to Time Series A time series is a collection of stochastic variables x,.., x t,.., x T indexed by an integer value t. The
More informationEstimation and comparison of multiple change-point models
Journal of Econometrics 86 (1998) 221 241 Estimation and comparison of multiple change-point models Siddhartha Chib* John M. Olin School of Business, Washington University, 1 Brookings Drive, Campus Box
More informationChristfried Webers. Canberra February June 2015
c Statistical Group and College of Engineering and Computer Science Canberra February June (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 829 c Part VIII Linear Classification 2 Logistic
More informationHandling attrition and non-response in longitudinal data
Longitudinal and Life Course Studies 2009 Volume 1 Issue 1 Pp 63-72 Handling attrition and non-response in longitudinal data Harvey Goldstein University of Bristol Correspondence. Professor H. Goldstein
More informationTime Series Analysis
JUNE 2012 Time Series Analysis CONTENT A time series is a chronological sequence of observations on a particular variable. Usually the observations are taken at regular intervals (days, months, years),
More informationAnalysis and Computation for Finance Time Series - An Introduction
ECMM703 Analysis and Computation for Finance Time Series - An Introduction Alejandra González Harrison 161 Email: mag208@exeter.ac.uk Time Series - An Introduction A time series is a sequence of observations
More informationBayesian Phylogeny and Measures of Branch Support
Bayesian Phylogeny and Measures of Branch Support Bayesian Statistics Imagine we have a bag containing 100 dice of which we know that 90 are fair and 10 are biased. The
More informationFinancial TIme Series Analysis: Part II
Department of Mathematics and Statistics, University of Vaasa, Finland January 29 February 13, 2015 Feb 14, 2015 1 Univariate linear stochastic models: further topics Unobserved component model Signal
More informationDynamic Linear Models with R
Giovanni Petris, Sonia Petrone, Patrizia Campagnoli Dynamic Linear Models with R SPIN Springer s internal project number, if known Monograph August 10, 2007 Springer Berlin Heidelberg NewYork Hong Kong
More informationIntroduction to Time Series Analysis. Lecture 6.
Introduction to Time Series Analysis. Lecture 6. Peter Bartlett www.stat.berkeley.edu/ bartlett/courses/153-fall2010 Last lecture: 1. Causality 2. Invertibility 3. AR(p) models 4. ARMA(p,q) models 1 Introduction
More informationBayesian Statistics in One Hour. Patrick Lam
Bayesian Statistics in One Hour Patrick Lam Outline Introduction Bayesian Models Applications Missing Data Hierarchical Models Outline Introduction Bayesian Models Applications Missing Data Hierarchical
More informationMonte Carlo-based statistical methods (MASM11/FMS091)
Monte Carlo-based statistical methods (MASM11/FMS091) Magnus Wiktorsson Centre for Mathematical Sciences Lund University, Sweden Lecture 6 Sequential Monte Carlo methods II February 7, 2014 M. Wiktorsson
More information1 Simulating Brownian motion (BM) and geometric Brownian motion (GBM)
Copyright c 2013 by Karl Sigman 1 Simulating Brownian motion (BM) and geometric Brownian motion (GBM) For an introduction to how one can construct BM, see the Appendix at the end of these notes A stochastic
More informationGaussian Processes to Speed up Hamiltonian Monte Carlo
Gaussian Processes to Speed up Hamiltonian Monte Carlo Matthieu Lê Murray, Iain http://videolectures.net/mlss09uk_murray_mcmc/ Rasmussen, Carl Edward. "Gaussian processes to speed up hybrid Monte Carlo
More informationMaster s thesis tutorial: part III
for the Autonomous Compliant Research group Tinne De Laet, Wilm Decré, Diederik Verscheure Katholieke Universiteit Leuven, Department of Mechanical Engineering, PMA Division 30 oktober 2006 Outline General
More informationThe Power of the KPSS Test for Cointegration when Residuals are Fractionally Integrated
The Power of the KPSS Test for Cointegration when Residuals are Fractionally Integrated Philipp Sibbertsen 1 Walter Krämer 2 Diskussionspapier 318 ISNN 0949-9962 Abstract: We show that the power of the
More informationModel Uncertainty in Classical Conditioning
Model Uncertainty in Classical Conditioning A. C. Courville* 1,3, N. D. Daw 2,3, G. J. Gordon 4, and D. S. Touretzky 2,3 1 Robotics Institute, 2 Computer Science Department, 3 Center for the Neural Basis
More informationAppendix 1: Time series analysis of peak-rate years and synchrony testing.
Appendix 1: Time series analysis of peak-rate years and synchrony testing. Overview The raw data are accessible at Figshare ( Time series of global resources, DOI 10.6084/m9.figshare.929619), sources are
More informationIntroduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization
Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization Wolfram Burgard, Maren Bennewitz, Diego Tipaldi, Luciano Spinello 1 Motivation Recall: Discrete filter Discretize
More informationBayesX - Software for Bayesian Inference in Structured Additive Regression
BayesX - Software for Bayesian Inference in Structured Additive Regression Thomas Kneib Faculty of Mathematics and Economics, University of Ulm Department of Statistics, Ludwig-Maximilians-University Munich
More informationEchantillonnage exact de distributions de Gibbs d énergies sous-modulaires. Exact sampling of Gibbs distributions with submodular energies
Echantillonnage exact de distributions de Gibbs d énergies sous-modulaires Exact sampling of Gibbs distributions with submodular energies Marc Sigelle 1 and Jérôme Darbon 2 1 Institut TELECOM TELECOM ParisTech
More informationStatistics Graduate Courses
Statistics Graduate Courses STAT 7002--Topics in Statistics-Biological/Physical/Mathematics (cr.arr.).organized study of selected topics. Subjects and earnable credit may vary from semester to semester.
More informationSampling via Moment Sharing: A New Framework for Distributed Bayesian Inference for Big Data
Sampling via Moment Sharing: A New Framework for Distributed Bayesian Inference for Big Data (Oxford) in collaboration with: Minjie Xu, Jun Zhu, Bo Zhang (Tsinghua) Balaji Lakshminarayanan (Gatsby) Bayesian
More informationAn Introduction to Using WinBUGS for Cost-Effectiveness Analyses in Health Economics
Slide 1 An Introduction to Using WinBUGS for Cost-Effectiveness Analyses in Health Economics Dr. Christian Asseburg Centre for Health Economics Part 1 Slide 2 Talk overview Foundations of Bayesian statistics
More informationIntroduction to Markov Chain Monte Carlo
Introduction to Markov Chain Monte Carlo Monte Carlo: sample from a distribution to estimate the distribution to compute max, mean Markov Chain Monte Carlo: sampling using local information Generic problem
More informationBasics of Statistical Machine Learning
CS761 Spring 2013 Advanced Machine Learning Basics of Statistical Machine Learning Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu Modern machine learning is rooted in statistics. You will find many familiar
More informationEnhancing Parallel Pattern Search Optimization with a Gaussian Process Oracle
Enhancing Parallel Pattern Search Optimization with a Gaussian Process Oracle Genetha A. Gray, Monica Martinez-Canales Computational Sciences & Mathematics Research Department, Sandia National Laboratories,
More informationComputing with Finite and Infinite Networks
Computing with Finite and Infinite Networks Ole Winther Theoretical Physics, Lund University Sölvegatan 14 A, S-223 62 Lund, Sweden winther@nimis.thep.lu.se Abstract Using statistical mechanics results,
More informationService courses for graduate students in degree programs other than the MS or PhD programs in Biostatistics.
Course Catalog In order to be assured that all prerequisites are met, students must acquire a permission number from the education coordinator prior to enrolling in any Biostatistics course. Courses are
More informationStephane Crepey. Financial Modeling. A Backward Stochastic Differential Equations Perspective. 4y Springer
Stephane Crepey Financial Modeling A Backward Stochastic Differential Equations Perspective 4y Springer Part I An Introductory Course in Stochastic Processes 1 Some Classes of Discrete-Time Stochastic
More informationGenerating Valid 4 4 Correlation Matrices
Applied Mathematics E-Notes, 7(2007), 53-59 c ISSN 1607-2510 Available free at mirror sites of http://www.math.nthu.edu.tw/ amen/ Generating Valid 4 4 Correlation Matrices Mark Budden, Paul Hadavas, Lorrie
More informationBig Data, Statistics, and the Internet
Big Data, Statistics, and the Internet Steven L. Scott April, 4 Steve Scott (Google) Big Data, Statistics, and the Internet April, 4 / 39 Summary Big data live on more than one machine. Computing takes
More informationChenfeng Xiong (corresponding), University of Maryland, College Park (cxiong@umd.edu)
Paper Author (s) Chenfeng Xiong (corresponding), University of Maryland, College Park (cxiong@umd.edu) Lei Zhang, University of Maryland, College Park (lei@umd.edu) Paper Title & Number Dynamic Travel
More informationA Tutorial on Particle Filters for Online Nonlinear/Non-Gaussian Bayesian Tracking
174 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 50, NO. 2, FEBRUARY 2002 A Tutorial on Particle Filters for Online Nonlinear/Non-Gaussian Bayesian Tracking M. Sanjeev Arulampalam, Simon Maskell, Neil
More informationSample Size Designs to Assess Controls
Sample Size Designs to Assess Controls B. Ricky Rambharat, PhD, PStat Lead Statistician Office of the Comptroller of the Currency U.S. Department of the Treasury Washington, DC FCSM Research Conference
More informationNetwork Security A Decision and Game-Theoretic Approach
Network Security A Decision and Game-Theoretic Approach Tansu Alpcan Deutsche Telekom Laboratories, Technical University of Berlin, Germany and Tamer Ba ar University of Illinois at Urbana-Champaign, USA
More informationOptimising and Adapting the Metropolis Algorithm
Optimising and Adapting the Metropolis Algorithm by Jeffrey S. Rosenthal 1 (February 2013; revised March 2013 and May 2013 and June 2013) 1 Introduction Many modern scientific questions involve high dimensional
More informationA Bootstrap Metropolis-Hastings Algorithm for Bayesian Analysis of Big Data
A Bootstrap Metropolis-Hastings Algorithm for Bayesian Analysis of Big Data Faming Liang University of Florida August 9, 2015 Abstract MCMC methods have proven to be a very powerful tool for analyzing
More informationMarkov Chain Monte Carlo Simulation Made Simple
Markov Chain Monte Carlo Simulation Made Simple Alastair Smith Department of Politics New York University April2,2003 1 Markov Chain Monte Carlo (MCMC) simualtion is a powerful technique to perform numerical
More informationSpatial Statistics Chapter 3 Basics of areal data and areal data modeling
Spatial Statistics Chapter 3 Basics of areal data and areal data modeling Recall areal data also known as lattice data are data Y (s), s D where D is a discrete index set. This usually corresponds to data
More informationDealing with large datasets
Dealing with large datasets (by throwing away most of the data) Alan Heavens Institute for Astronomy, University of Edinburgh with Ben Panter, Rob Tweedie, Mark Bastin, Will Hossack, Keith McKellar, Trevor
More informationData Science Center Eindhoven. Big Data: Challenges and Opportunities for Mathematicians. Alessandro Di Bucchianico
Data Science Center Eindhoven Big Data: Challenges and Opportunities for Mathematicians Alessandro Di Bucchianico Dutch Mathematical Congress April 15, 2015 Contents 1. Big Data terminology 2. Various
More informationSection 5. Stan for Big Data. Bob Carpenter. Columbia University
Section 5. Stan for Big Data Bob Carpenter Columbia University Part I Overview Scaling and Evaluation data size (bytes) 1e18 1e15 1e12 1e9 1e6 Big Model and Big Data approach state of the art big model
More informationTraffic Safety Facts. Research Note. Time Series Analysis and Forecast of Crash Fatalities during Six Holiday Periods Cejun Liu* and Chou-Lin Chen
Traffic Safety Facts Research Note March 2004 DOT HS 809 718 Time Series Analysis and Forecast of Crash Fatalities during Six Holiday Periods Cejun Liu* and Chou-Lin Chen Summary This research note uses
More informationBayesian Machine Learning (ML): Modeling And Inference in Big Data. Zhuhua Cai Google, Rice University caizhua@gmail.com
Bayesian Machine Learning (ML): Modeling And Inference in Big Data Zhuhua Cai Google Rice University caizhua@gmail.com 1 Syllabus Bayesian ML Concepts (Today) Bayesian ML on MapReduce (Next morning) Bayesian
More informationReferences. Importance Sampling. Jessi Cisewski (CMU) Carnegie Mellon University. June 2014
Jessi Cisewski Carnegie Mellon University June 2014 Outline 1 Recall: Monte Carlo integration 2 3 Examples of (a) Monte Carlo, Monaco (b) Monte Carlo Casino Some content and examples from Wasserman (2004)
More informationBayesian analysis of aggregate loss models.
Bayesian analysis of aggregate loss models. M. C. Ausín,J.M.Vilar,R.CaoandC.González-Fragueiro. Department of Mathematics, Universidade da Coruña Abstract This paper describes a Bayesian approach to make
More informationINDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition)
INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition) Abstract Indirect inference is a simulation-based method for estimating the parameters of economic models. Its
More informationTime Series Analysis 1. Lecture 8: Time Series Analysis. Time Series Analysis MIT 18.S096. Dr. Kempthorne. Fall 2013 MIT 18.S096
Lecture 8: Time Series Analysis MIT 18.S096 Dr. Kempthorne Fall 2013 MIT 18.S096 Time Series Analysis 1 Outline Time Series Analysis 1 Time Series Analysis MIT 18.S096 Time Series Analysis 2 A stochastic
More information> plot(exp.btgpllm, main = "treed GP LLM,", proj = c(1)) > plot(exp.btgpllm, main = "treed GP LLM,", proj = c(2)) quantile diff (error)
> plot(exp.btgpllm, main = "treed GP LLM,", proj = c(1)) > plot(exp.btgpllm, main = "treed GP LLM,", proj = c(2)) 0.4 0.2 0.0 0.2 0.4 treed GP LLM, mean treed GP LLM, 0.00 0.05 0.10 0.15 0.20 x1 x1 0.4
More informationSelection Sampling from Large Data sets for Targeted Inference in Mixture Modeling
Selection Sampling from Large Data sets for Targeted Inference in Mixture Modeling Ioanna Manolopoulou, Cliburn Chan and Mike West December 29, 2009 Abstract One of the challenges of Markov chain Monte
More informationPREDICTIVE DISTRIBUTIONS OF OUTSTANDING LIABILITIES IN GENERAL INSURANCE
PREDICTIVE DISTRIBUTIONS OF OUTSTANDING LIABILITIES IN GENERAL INSURANCE BY P.D. ENGLAND AND R.J. VERRALL ABSTRACT This paper extends the methods introduced in England & Verrall (00), and shows how predictive
More informationProblem of Missing Data
VASA Mission of VA Statisticians Association (VASA) Promote & disseminate statistical methodological research relevant to VA studies; Facilitate communication & collaboration among VA-affiliated statisticians;
More informationHierarchical Bayesian Modeling of the HIV Response to Therapy
Hierarchical Bayesian Modeling of the HIV Response to Therapy Shane T. Jensen Department of Statistics, The Wharton School, University of Pennsylvania March 23, 2010 Joint Work with Alex Braunstein and
More informationChapter 9: Univariate Time Series Analysis
Chapter 9: Univariate Time Series Analysis In the last chapter we discussed models with only lags of explanatory variables. These can be misleading if: 1. The dependent variable Y t depends on lags of
More informationEE 570: Location and Navigation
EE 570: Location and Navigation On-Line Bayesian Tracking Aly El-Osery 1 Stephen Bruder 2 1 Electrical Engineering Department, New Mexico Tech Socorro, New Mexico, USA 2 Electrical and Computer Engineering
More informationDirichlet Process Gaussian Mixture Models: Choice of the Base Distribution
Görür D, Rasmussen CE. Dirichlet process Gaussian mixture models: Choice of the base distribution. JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY 5(4): 615 66 July 010/DOI 10.1007/s11390-010-1051-1 Dirichlet
More informationMAN-BITES-DOG BUSINESS CYCLES ONLINE APPENDIX
MAN-BITES-DOG BUSINESS CYCLES ONLINE APPENDIX KRISTOFFER P. NIMARK The next section derives the equilibrium expressions for the beauty contest model from Section 3 of the main paper. This is followed by
More informationMonte Carlo Methods and Models in Finance and Insurance
Chapman & Hall/CRC FINANCIAL MATHEMATICS SERIES Monte Carlo Methods and Models in Finance and Insurance Ralf Korn Elke Korn Gerald Kroisandt f r oc) CRC Press \ V^ J Taylor & Francis Croup ^^"^ Boca Raton
More informationThe CUSUM algorithm a small review. Pierre Granjon
The CUSUM algorithm a small review Pierre Granjon June, 1 Contents 1 The CUSUM algorithm 1.1 Algorithm............................... 1.1.1 The problem......................... 1.1. The different steps......................
More informationBig data challenges for physics in the next decades
Big data challenges for physics in the next decades David W. Hogg Center for Cosmology and Particle Physics, New York University 2012 November 09 punchlines Huge data sets create new opportunities. they
More informationProgram description for the Master s Degree Program in Mathematics and Finance
Program description for the Master s Degree Program in Mathematics and Finance : English: Master s Degree in Mathematics and Finance Norwegian, bokmål: Master i matematikk og finans Norwegian, nynorsk:
More informationMaster s Theory Exam Spring 2006
Spring 2006 This exam contains 7 questions. You should attempt them all. Each question is divided into parts to help lead you through the material. You should attempt to complete as much of each problem
More informationBias in the Estimation of Mean Reversion in Continuous-Time Lévy Processes
Bias in the Estimation of Mean Reversion in Continuous-Time Lévy Processes Yong Bao a, Aman Ullah b, Yun Wang c, and Jun Yu d a Purdue University, IN, USA b University of California, Riverside, CA, USA
More informationUnivariate Time Series Analysis; ARIMA Models
Econometrics 2 Spring 25 Univariate Time Series Analysis; ARIMA Models Heino Bohn Nielsen of4 Outline of the Lecture () Introduction to univariate time series analysis. (2) Stationarity. (3) Characterizing
More informationMarkov Chain Monte Carlo and Numerical Differential Equations
Markov Chain Monte Carlo and Numerical Differential Equations J.M. Sanz-Serna 1 Introduction This contribution presents a hopefully readable introduction to Markov Chain Monte Carlo methods with particular
More informationParameter estimation for nonlinear models: Numerical approaches to solving the inverse problem. Lecture 12 04/08/2008. Sven Zenker
Parameter estimation for nonlinear models: Numerical approaches to solving the inverse problem Lecture 12 04/08/2008 Sven Zenker Assignment no. 8 Correct setup of likelihood function One fixed set of observation
More informationThe Monte Carlo Framework, Examples from Finance and Generating Correlated Random Variables
Monte Carlo Simulation: IEOR E4703 Fall 2004 c 2004 by Martin Haugh The Monte Carlo Framework, Examples from Finance and Generating Correlated Random Variables 1 The Monte Carlo Framework Suppose we wish
More informationAdvanced Signal Processing and Digital Noise Reduction
Advanced Signal Processing and Digital Noise Reduction Saeed V. Vaseghi Queen's University of Belfast UK WILEY HTEUBNER A Partnership between John Wiley & Sons and B. G. Teubner Publishers Chichester New
More informationBayesian Methods for the Social and Behavioral Sciences
Bayesian Methods for the Social and Behavioral Sciences Jeff Gill Harvard University 2007 ICPSR First Session: June 25-July 20, 9-11 AM. Email: jgill@iq.harvard.edu TA: Yu-Sung Su (ys463@columbia.edu).
More informationArnold Zellner. Booth School of Business. University of Chicago. 5807 South Woodlawn Avenue Chicago, IL 60637. arnold.zellner@chicagobooth.
H.G.B. Alexander Research Foundation Graduate School of Business University of Chicago Comments on Harold Jeffreys Theory of Probability Revisited, co-authored by C.P. Robert, N. Chopin and J. Rousseau.
More informationAnalysis of a Production/Inventory System with Multiple Retailers
Analysis of a Production/Inventory System with Multiple Retailers Ann M. Noblesse 1, Robert N. Boute 1,2, Marc R. Lambrecht 1, Benny Van Houdt 3 1 Research Center for Operations Management, University
More informationHT2015: SC4 Statistical Data Mining and Machine Learning
HT2015: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Bayesian Nonparametrics Parametric vs Nonparametric
More informationComputational Statistics for Big Data
Lancaster University Computational Statistics for Big Data Author: 1 Supervisors: Paul Fearnhead 1 Emily Fox 2 1 Lancaster University 2 The University of Washington September 1, 2015 Abstract The amount
More informationThe Chinese Restaurant Process
COS 597C: Bayesian nonparametrics Lecturer: David Blei Lecture # 1 Scribes: Peter Frazier, Indraneel Mukherjee September 21, 2007 In this first lecture, we begin by introducing the Chinese Restaurant Process.
More informationThe HB. How Bayesian methods have changed the face of marketing research. Summer 2004
The HB How Bayesian methods have changed the face of marketing research. 20 Summer 2004 Reprinted with permission from Marketing Research, Summer 2004, published by the American Marketing Association.
More information