Supervised Image Segmentation across Scanner Protocols: A Transfer Learning Approach
|
|
- Marilyn White
- 7 years ago
- Views:
Transcription
1 Supervised Image Segmentation across Scanner Protocols: A Transfer Learning Approach Annegreet van Opbroek 1, M. Arfan Ikram 2, Meike W. Vernooij 2 and Marleen de Bruijne 1,3 1 Biomedical Imaging Group Rotterdam, Departments of Medical Informatics and Radiology, Erasmus MC - University Medical Center Rotterdam, the Netherlands 2 Departments of Epidemiology and Radiology, Erasmus MC - University Medical Center Rotterdam, the Netherlands 3 Department of Computer Science, University of Copenhagen, Denmark Abstract. Supervised classification techniques are among the most powerful methods used for automatic segmentation of medical images. A disadvantage of these methods is that they require a representative training set and thus encounter problems when the training data is acquired e.g. with a different scanner protocol than the target segmentation data. We therefore propose a framework for supervised biomedical image segmentation across different scanner protocols, by means of transfer learning. We establish a transfer learning algorithm for classification, which can exploit a large amount of labeled samples from different sources in addition to a small amount of samples from the target source. The algorithm iteratively re-weights the contribution of training samples from these different sources based on classification by a weighted SVM classifier. We evaluate this technique by performing tissue classification on MRI brain data from four substantially different scanning protocols. For a small number of labeled samples from a single image obtained with the same protocol, the proposed transfer learning method outperforms classification on all available training data as well as classification based on the labeled target samples only. The classification errors for these cases can be reduced with up to 40 percent compared to traditional classification techniques. 1 Introduction Supervised classification techniques are commonly used in automatic segmentation of biomedical images. A major drawback of these methods is that they require a sufficiently large training set from a similar distribution as the images to be segmented. This means that in practice these techniques often cannot be applied to data obtained with a different scanning protocol, scanner, or image modality, without establishing a new, usually manually annotated dataset. Common methods to cope with differences between training and test distributions are based on exploiting similarities between training and test data, e.g. by embedding the physics of the image acquisition process in the segmentation
2 framework [1], and using prior tissue probability maps to identify new training samples from target data [2]. Another approach is to use unsupervised clustering methods on the target source [3 5]. In this article we present a new approach to the problem of learning automatic classification across scanners, which can reduce the effort of re-training, by transferring knowledge from different scanners. Our method relies on a relatively new area of machine learning, called Transfer Learning [6]. Transfer learning copes with cases where distributions, feature spaces and/or tasks differ between training and test data, as opposed to traditional machine learning techniques where these are assumed to be the same between training and testing. As a proof of concept we investigate whether transfer learning can improve between-scanner segmentation performance in a basic voxelwise classification segmentation framework. Hereto we establish a transfer learning method that makes use of a large amount of different-distribution training samples, which come from different sources than the target data, in addition to a relatively small amount of same-distribution samples from the target distribution. Our method relies on iteratively weighting these different-distribution samples according to classification-outcome of a weighted support vector machine (SVM) classifier on all available training data. This way, the suitable different-distribution samples are selected, which help regularize the classification and thus reduce the problems related to the small sample size of the same-distribution data. 2 A Transfer Learning Approach to Classification We make use of a small amount of labeled data from the target source, which we will call the same-distribution training data, denoted by T s = {x s i,ys i }Ns i=1, where N s represents the number of same-distribution training samples x s i, with corresponding labels yi s. Apart from T s we also have a large amount of training data from other sources, which may have different distributions. This differentdistribution training data is denoted by T d = {x d i,yd i }N d i=1, where N d is the total number of different-distribution training samples x d i, with labels yd i (typically N d N s ), so that there is a total training set T = T d T s. Our algorithm iteratively calculates a weighted SVM classifier c t from all available training data, which is then used to determine a new weight vector w t+1 for all training samples. Samples from T d that contradict labeled same-distribution data may be misclassified by the trained classifier. In the next round these samples will receive a lower weight, and will thus have less influence on the decision boundary. The weighting of the different-distribution training samples is achieved by multiplication with β 1 δ(ct(xd i ),yd i ), where δ(c( t x d i ),yd i ) denotes the Kronecker delta. The value for β is taken as β = 1/(1 + 2lnN d /N it ), as determined for the TrAdaBoost algorithm [7], which is also based on iteratively reweighting same-distribution training samples. Here N it denotes the total number of iterations performed. Thus, when more iterations are performed the reduction of weights in one iteration diminishes. An initial weight vector w 1 for respec-
3 tively the different- and same-distribution samples gives each of the differentdistribution training samples a weight R N d and the same-distribution training 1 samples a weight N s, so that R (R 0) denotes the ratio between the total weight of the T d samples and the total weight of the T s samples. The algorithm is summarized in Table 1. Table 1. Our transfer learning algorithm in pseudo code Input: Output: T, N s, N d, N it,r classifier for test data c Nit set β = 1/(1+ { 2lnN d /N it) R set wi 1 for i = 1,2,...,N N d = d 1 for i = N d +1,...,N d +N s N s For t = 1,2,...,N it Normalize w t = w t N d +Ns j=1 w j t Calculate Classifier c t from T and w t, Update w t+1 i = w t iβ 1 δ(ct(xd i ),yd i ), for i = 1,2,...,N d end 2.1 Weighted Support Vector Machine Classification For classification we use a weighted support vector machine, as provided by LIB- SVM [8]. In this weighted SVM every trainingsample x i is givenaweight w t i 0 that describes the importance of the sample, so that a training sample with a high weight is more important to classify correctly. The objective function for the optimal classifier c t (x) = V x+v 0 by weighted SVM reads min V 1 2 V V +C N i=1 wt i ξ i (1) s.t. V T x i +v 0 1 ξ i for y i = +1 V T x i +v 0 1+ξ i for y i = 1 ξ i 0 i = 1,2,...,N. Also, as in regular SVM the trade-off parameter C can be determined with cross-validation. 3 Experiment: MRI Brain Tissue Segmentation We consider the application of MRI brain tissue segmentation by voxelwise classification on images acquired with a certain scanner, for which no manual labels
4 are yet available. Since manual segmentation is time consuming only a very small amount of voxels in a single image is labeled, to train a classification scheme for the remainder of images. In addition to these same-distribution training samples, a larger amount of manually labeled samples from different patients made with a variety of scanners, are already available from other studies. We evaluate whether our transfer learning algorithm can outperform traditional classification. Data Description We use annotated MRI brain data from four different sources, which display a large amount of variation in intensity and tissue contrast: 1. 6 T1-weighted images from the Rotterdam Scan Study [9] made with a 1.5T GE scanner with mm 3 voxel size HASTE-Odd images (inversion time = 4400 ms, TR = 2800 ms, TE = 29 ms) from the Rotterdam Scan Study [9] made with a 1.5T Siemens scanner with mm 3 voxel size T1-weighted images from the Internet Brain Segmentation Repository (IBSR) [10] with mm 3 voxel size T1-weighted images from the IBSR, 10 from a 1.5T Siemens scanner, 10 from a 1.5T GE scanner, both with mm 3 voxel size Slices from one image of the four different sources are shown in Fig. 1. (a) Source 1 (b) Source 2 (c) Source 3 (d) Source 4 Fig. 1. Slices of images from the four different sources. For the Haste-Odd the images are inverted so that CSF has lowest intensity and WM the highest, as in the T1-weighted images. All 56 images are corrected for non-uniformity in intensity with the N3 method [11] within a mask, and then normalized so that the voxels between the 4th and the 96th intensity percentile are mapped between 0 and 1. For all images expert segmentations for white matter (WM), gray matter (GM), and cerebrospinal fluid (CSF) are available for training and testing. Some images also have expert segmentations for white matter lesions; these voxels are included in the WM class. Experimental Setup A set of cross-validation experiments is performed. All four sources are in turn used as the test source, while the other three sources are
5 used to extract different-distribution training data. As training and test samples voxels are randomly selected within the manually annotated brain mask. As different-distribution training data, samples per source are randomly selected from the different images. Same-distribution training data is sampled from a single image from the test source, where we start by adding one samedistribution sample from every class, and subsequently add more samples randomly distributed over the classes to generate learning curves to analyze the classification performance as a function of the number of same-distribution training samples. The learning curves are created by testing on randomly selected samples from each test source image that is not used for training, resulting in a total of to test samples, depending on the amount of images in the source. This gives a reliable estimation of the classification performance on whole images, but is computationaly less expensive. We repeat the experiment for every test source image as same-distribution training source, and determine average classification accuracy per image. In addition, to compare to brain tissue segmentation results reported in the literature, full image segmentation is performed at N s = 20. A total of four features are used for classification, consisting of the image intensity plus three spatial features which give the x, y, and z coordinates of the voxel as a fraction of the total dimensions of the brain. The features within each source are normalized to zero mean and unit standard deviation. From now on we will refer to the transfer learning algorithm with the weighted SVM classifier as a Transfer SVM. The Transfer SVM is compared to normal SVM classification both onallavailabletrainingdata(t s T d ) andonthesame-distributiontraining data (T s ) alone. In all classification schemes the SVM classifier is extended to a multiclass classifier by one-vs-one classification, which overall gives better results on the data than one-vs-rest classification. All SVM classifiers use a linear kernel. Parameter Selection The SVM parameter C for each of the four experiments is determined with cross-validation with a regular SVM in the three differentdistribution training sources. The parameter with the best performance in the three sources is selected. The total number of iterations of the transfer learning algorithm is set to N it = 20, which is sufficient for convergence in all cases. For each number of same-distribution training samples the parameter R that determines the balance between the initial weights of the different- and samedistribution samples in the Transfer SVM is determined with cross-validation with the Transfer SVM on the data from the three different-distribution sources. In turn each source is selected as target source to extract labeled same-distribution samples and test data, while the other two sources are used to extract differentdistribution training data. The best R is then determined by averaging between the three sources. The resulting R is around 5 for three same-distribution training samples, but falls off exponentially to R 1 for 200 same-distribution training samples.
6 Results In Fig.2the learningcurvesforeachofthe fourexperimentsareshown. In all experiments we see that for a small number of same-distribution training samples (between 3 and 40 in Fig. 2(b), and a broader range for the other three experiments) the Transfer SVM outperforms both the SVM on all training data T s + T d and the SVM on the labeled same-distribution data T s only. When more same-distribution training samples are available the SVM classifier on T s converges to the Transfer SVM and might even perform slightly better. For a very small amount of same-distribution training data the SVM classifier on all available training samples performs best in two of the four cases. In these cases the Transfer SVM seems to give too little weight to the different-distribution training samples, which results in a classifier that does not outperform the SVM on T s +T d. (a) Source 1 (b) Source 2 (c) Source 3 (d) Source 4 Fig. 2. Mean classification errors and 95%-confidence intervals for our transfer learning SVM, the conventional SVM on T d T s, and the SVM on T s, all with a linear kernel. The test source consist of (a) Source 1, (b) Source 2, (c) Source 3, (d) Source 4, the different-distribution training data comes from the three remaining sources. Fig. 3 shows an example of segmentations from Source 1 obtained by the three methods for 20 same-distribution training samples. The Transfer SVM classifier produces a good segmentation with a classification error of 6.7%. The
7 segmentation of the SVM T s + T d classifier undersegments WM and CSF and gives an error of 9.3%, while the SVM T s classifier oversegments the CSF, resulting in an error of 14.0%. According to the learning curve in Fig. 2(a) the number of labeled same-distribution training samples needed for the SVM T s classifier to produce a similar result lies around N s = 80. Thus, in this case transferring knowledge from the different-distribution training samples with a transfer classifier reduces the costs on labeling new same-distribution data with a factor 4. (a) (b) (c) (d) (e) Fig.3. Segmentation results on Source 1 for N s = 20. (a) Original T1 image, (b) manual segmentation, (c) SVMT s+t d classifier, (d) SVMT s classifier, and (e) Transfer SVM classifier. 4 Conclusion and Discussion We presented a transfer learning method to segment biomedical images from different scanning protocols. Our algorithm makes use of labeled image data from a variety of scanners, on top of a small amount of labeled training data from the target scanner. Experiments on MRI brain tissue segmentation show that our algorithm can reduce the amount of misclassified voxels in the image with up to 40%. To do this, the algorithm exploits already available training data from different scanners and weights them according to correspondence with the training data from the target scanner. With our algorithm just a few manually annotated samples in a single image obtained with a new protocol are sufficient to retrain the method and provide much better results than a classifier trained on different- or same-distribution data alone. Our transfer learning algorithm has proved to be capable of handling data from four drastically different sources, with different pulse sequence parameters and different slice thickness in different orientations. To allow direct comparison of different learning algorithms, we have in this work focused on segmentation based on voxelwise classification. This framework could be used as the basis for a more advanced brain tissue segmentation algorithm including e.g. regularization, atlas based priors, and modeling of partial volume effect. To give an indication of the performance of our segmentation
8 method in comparison with the current state of the art, we compare the performance reported for nine methods in [3] applied to the IBSR dataset with 20 subjects. For N s = 20 our method reports a mean Jaccard index on the 20 imagesof0.15forcsf, 0.68forGM, and 0.61for WM. The best ofthe 9algorithms in [3] reports coefficients of 0.10 for CSF, 0.68 for GM, and 0.69 for WM. Three methods in [3] report coefficients higher than 0.61 for WM, but our algorithm outperforms all methods based on CSF scores, and all but one on GM scores. Also, when N s is increased our method performs even better. Eventhough the focus ofthe experiments in this paperis onmri braintissue segmentation, the proposed method is general and we expect it to be useful in many other applications where classification is used, such as voxel classification in other areas, brain structure segmentation, and computer aided diagnosis. To conclude, the experiments give a strong indication that supervised image segmentation techniques can benefit from different-distribution training data, in a transfer learning setting. This way, the large amount of (publicly) available annotated data from previous studies can be used to segment images with relatively large differences in imaging protocols, which forms an important step towards application in a clinical setting. References 1. Fischl, B., Salat, D., van der Kouwe, A., Makris, N., Ségonne, F., Quinn, B., Dale, A.: Sequence-independent segmentation of magnetic resonance images. Neuroimage 23 (2004) S69 S84 2. Cocosco, C., Zijdenbos, A., Evans, A.: A fully automatic and robust brain MRI tissue classification method. Medical Image Analysis 7(4) (2003) Mayer, A., Greenspan, H.: An adaptive mean-shift framework for MRI brain segmentation. Medical Imaging, IEEE Transactions on 28(8) (2009) Grabowski, T., Frank, R., Szumski, N., Brown, C., Damasio, H.: Validation of partial tissue segmentation of single-channel magnetic resonance images of the brain. NeuroImage 12(6) (2000) Van Leemput, K., Maes, F., Vandermeulen, D., Suetens, P.: Automated modelbased tissue classification of MR images of the brain. Medical Imaging, IEEE Transactions on 18(10) (1999) Pan, S., Yang, Q.: A survey on transfer learning. Knowledge and Data Engineering, IEEE Transactions on 22(10) (2010) Dai, W., Yang, Q., Xue, G., Yu, Y.: Boosting for transfer learning. In: Proceedings of the 24th international conference on Machine learning, ACM (2007) Chang, C., Lin, C.: Libsvm: a library for support vector machines. ACM Transactions on Intelligent Systems and Technology (TIST) 2(3) (2011) Hofman, A., Breteler, M., Van Duijn, C., Janssen, H., Krestin, G., Kuipers, E., Stricker, B., Tiemeier, H., Uitterlinden, A., Vingerling, J., et al.: The Rotterdam Study: 2010 objectives and design update. European journal of epidemiology 24(9) (2009) Worth, A.: The Internet Brain Segmentation Repository (IBSR) 11. Sled, J., Zijdenbos, A., Evans, A.: A nonparametric method for automatic correction of intensity nonuniformity in MRI data. Medical Imaging, IEEE Transactions on 17(1) (1998) 87 97
2. MATERIALS AND METHODS
Difficulties of T1 brain MRI segmentation techniques M S. Atkins *a, K. Siu a, B. Law a, J. Orchard a, W. Rosenbaum a a School of Computing Science, Simon Fraser University ABSTRACT This paper looks at
More informationMEDIMAGE A Multimedia Database Management System for Alzheimer s Disease Patients
MEDIMAGE A Multimedia Database Management System for Alzheimer s Disease Patients Peter L. Stanchev 1, Farshad Fotouhi 2 1 Kettering University, Flint, Michigan, 48504 USA pstanche@kettering.edu http://www.kettering.edu/~pstanche
More informationMorphological analysis on structural MRI for the early diagnosis of neurodegenerative diseases. Marco Aiello On behalf of MAGIC-5 collaboration
Morphological analysis on structural MRI for the early diagnosis of neurodegenerative diseases Marco Aiello On behalf of MAGIC-5 collaboration Index Motivations of morphological analysis Segmentation of
More informationGE Medical Systems Training in Partnership. Module 8: IQ: Acquisition Time
Module 8: IQ: Acquisition Time IQ : Acquisition Time Objectives...Describe types of data acquisition modes....compute acquisition times for 2D and 3D scans. 2D Acquisitions The 2D mode acquires and reconstructs
More informationMathematical Models of Supervised Learning and their Application to Medical Diagnosis
Genomic, Proteomic and Transcriptomic Lab High Performance Computing and Networking Institute National Research Council, Italy Mathematical Models of Supervised Learning and their Application to Medical
More informationActive Learning SVM for Blogs recommendation
Active Learning SVM for Blogs recommendation Xin Guan Computer Science, George Mason University Ⅰ.Introduction In the DH Now website, they try to review a big amount of blogs and articles and find the
More information32-Channel Head Coil Imaging at 3T
32-Channel Head Coil Imaging at 3T Thomas Benner Athinoula A. Martinos Center for Biomedical Imaging, Department of Radiology, Massachusetts General Hospital and Harvard Medical School, Boston, MA, USA
More informationCOST AID ASL post- processing Workshop
COST AID ASL post- processing Workshop This workshop runs thought the post- processing of ASL data using tools from the FMRIB Software Library (www.fmrib.ox.ac.uk.uk/fsl), we will primarily focus on the
More informationRecognition. Sanja Fidler CSC420: Intro to Image Understanding 1 / 28
Recognition Topics that we will try to cover: Indexing for fast retrieval (we still owe this one) History of recognition techniques Object classification Bag-of-words Spatial pyramids Neural Networks Object
More informationThe Artificial Prediction Market
The Artificial Prediction Market Adrian Barbu Department of Statistics Florida State University Joint work with Nathan Lay, Siemens Corporate Research 1 Overview Main Contributions A mathematical theory
More informationColour Image Segmentation Technique for Screen Printing
60 R.U. Hewage and D.U.J. Sonnadara Department of Physics, University of Colombo, Sri Lanka ABSTRACT Screen-printing is an industry with a large number of applications ranging from printing mobile phone
More informationPredict Influencers in the Social Network
Predict Influencers in the Social Network Ruishan Liu, Yang Zhao and Liuyu Zhou Email: rliu2, yzhao2, lyzhou@stanford.edu Department of Electrical Engineering, Stanford University Abstract Given two persons
More informationPATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical
More informationMaschinelles Lernen mit MATLAB
Maschinelles Lernen mit MATLAB Jérémy Huard Applikationsingenieur The MathWorks GmbH 2015 The MathWorks, Inc. 1 Machine Learning is Everywhere Image Recognition Speech Recognition Stock Prediction Medical
More informationUpdate: MRI in Multiple sclerosis
Nyt indenfor MS ved MR Update: MRI in Multiple sclerosis Hartwig Roman Siebner Danish Research Centre for Magnetic Resonance (DRCMR) Copenhagen University Hospital Hvidovre Dansk Radiologisk Selskabs 10.
More informationMaking Sense of the Mayhem: Machine Learning and March Madness
Making Sense of the Mayhem: Machine Learning and March Madness Alex Tran and Adam Ginzberg Stanford University atran3@stanford.edu ginzberg@stanford.edu I. Introduction III. Model The goal of our research
More informationA CBIR System for Human Brain Magnetic Resonance Image Indexing
A CBIR System for Human Brain Magnetic Resonance Image Indexing Mina Rafi Nazari Electrical engineering department Sharif University of Technology International campus Tehran, Iran Emad Fatemizadeh Electrical
More informationOpen-Set Face Recognition-based Visitor Interface System
Open-Set Face Recognition-based Visitor Interface System Hazım K. Ekenel, Lorant Szasz-Toth, and Rainer Stiefelhagen Computer Science Department, Universität Karlsruhe (TH) Am Fasanengarten 5, Karlsruhe
More informationAzure Machine Learning, SQL Data Mining and R
Azure Machine Learning, SQL Data Mining and R Day-by-day Agenda Prerequisites No formal prerequisites. Basic knowledge of SQL Server Data Tools, Excel and any analytical experience helps. Best of all:
More informationPredict the Popularity of YouTube Videos Using Early View Data
000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050
More informationThe File-Card-Browser View for Breast DCE-MRI Data
The File-Card-Browser View for Breast DCE-MRI Data Sylvia Glaßer 1, Kathrin Scheil 1, Uta Preim 2, Bernhard Preim 1 1 Department of Simulation and Graphics, University of Magdeburg 2 Department of Radiology,
More informationBlog Post Extraction Using Title Finding
Blog Post Extraction Using Title Finding Linhai Song 1, 2, Xueqi Cheng 1, Yan Guo 1, Bo Wu 1, 2, Yu Wang 1, 2 1 Institute of Computing Technology, Chinese Academy of Sciences, Beijing 2 Graduate School
More informationInternational Journal of Computer Science Trends and Technology (IJCST) Volume 2 Issue 3, May-Jun 2014
RESEARCH ARTICLE OPEN ACCESS A Survey of Data Mining: Concepts with Applications and its Future Scope Dr. Zubair Khan 1, Ashish Kumar 2, Sunny Kumar 3 M.Tech Research Scholar 2. Department of Computer
More informationFUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT MINING SYSTEM
International Journal of Innovative Computing, Information and Control ICIC International c 0 ISSN 34-48 Volume 8, Number 8, August 0 pp. 4 FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT
More informationApplied Mathematical Sciences, Vol. 7, 2013, no. 112, 5591-5597 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2013.
Applied Mathematical Sciences, Vol. 7, 2013, no. 112, 5591-5597 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2013.38457 Accuracy Rate of Predictive Models in Credit Screening Anirut Suebsing
More informationEnd-to-End Sentiment Analysis of Twitter Data
End-to-End Sentiment Analysis of Twitter Data Apoor v Agarwal 1 Jasneet Singh Sabharwal 2 (1) Columbia University, NY, U.S.A. (2) Guru Gobind Singh Indraprastha University, New Delhi, India apoorv@cs.columbia.edu,
More informationADVANCED MACHINE LEARNING. Introduction
1 1 Introduction Lecturer: Prof. Aude Billard (aude.billard@epfl.ch) Teaching Assistants: Guillaume de Chambrier, Nadia Figueroa, Denys Lamotte, Nicola Sommer 2 2 Course Format Alternate between: Lectures
More informationPractical Data Science with Azure Machine Learning, SQL Data Mining, and R
Practical Data Science with Azure Machine Learning, SQL Data Mining, and R Overview This 4-day class is the first of the two data science courses taught by Rafal Lukawiecki. Some of the topics will be
More informationDiffusione e perfusione in risonanza magnetica. E. Pagani, M. Filippi
Diffusione e perfusione in risonanza magnetica E. Pagani, M. Filippi DW-MRI DIFFUSION-WEIGHTED MRI Principles Diffusion results from a microspic random motion known as Brownian motion THE RANDOM WALK How
More informationMachine Learning for Medical Image Analysis. A. Criminisi & the InnerEye team @ MSRC
Machine Learning for Medical Image Analysis A. Criminisi & the InnerEye team @ MSRC Medical image analysis the goal Automatic, semantic analysis and quantification of what observed in medical scans Brain
More informationServer Load Prediction
Server Load Prediction Suthee Chaidaroon (unsuthee@stanford.edu) Joon Yeong Kim (kim64@stanford.edu) Jonghan Seo (jonghan@stanford.edu) Abstract Estimating server load average is one of the methods that
More informationClassification of Bad Accounts in Credit Card Industry
Classification of Bad Accounts in Credit Card Industry Chengwei Yuan December 12, 2014 Introduction Risk management is critical for a credit card company to survive in such competing industry. In addition
More informationSupport Vector Machines for Dynamic Biometric Handwriting Classification
Support Vector Machines for Dynamic Biometric Handwriting Classification Tobias Scheidat, Marcus Leich, Mark Alexander, and Claus Vielhauer Abstract Biometric user authentication is a recent topic in the
More informationSuper-Resolution Reconstruction in MRI: Better Images Faster?
Super-Resolution Reconstruction in MRI: Better Images Faster? Esben Plenge 1,, Dirk H. J. Poot 1, Monique Bernsen 2, Gyula Kotek 2, Gavin Houston 2, Piotr Wielopolski 2, Louise van der Weerd 3, Wiro J.
More informationInternet Traffic Prediction by W-Boost: Classification and Regression
Internet Traffic Prediction by W-Boost: Classification and Regression Hanghang Tong 1, Chongrong Li 2, Jingrui He 1, and Yang Chen 1 1 Department of Automation, Tsinghua University, Beijing 100084, China
More informationLinear smoother. ŷ = S y. where s ij = s ij (x) e.g. s ij = diag(l i (x)) To go the other way, you need to diagonalize S
Linear smoother ŷ = S y where s ij = s ij (x) e.g. s ij = diag(l i (x)) To go the other way, you need to diagonalize S 2 Online Learning: LMS and Perceptrons Partially adapted from slides by Ryan Gabbard
More informationMRI SEQUENCES. 1 Gradient Echo Sequence
5 An MRI sequence is an ordered combination of RF and gradient pulses designed to acquire the data to form the image. In this chapter I will describe the basic gradient echo, spin echo and inversion recovery
More informationScalable Developments for Big Data Analytics in Remote Sensing
Scalable Developments for Big Data Analytics in Remote Sensing Federated Systems and Data Division Research Group High Productivity Data Processing Dr.-Ing. Morris Riedel et al. Research Group Leader,
More informationStatistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
More informationNorbert Schuff Professor of Radiology VA Medical Center and UCSF Norbert.schuff@ucsf.edu
Norbert Schuff Professor of Radiology Medical Center and UCSF Norbert.schuff@ucsf.edu Medical Imaging Informatics 2012, N.Schuff Course # 170.03 Slide 1/67 Overview Definitions Role of Segmentation Segmentation
More informationAnalysis Tools and Libraries for BigData
+ Analysis Tools and Libraries for BigData Lecture 02 Abhijit Bendale + Office Hours 2 n Terry Boult (Waiting to Confirm) n Abhijit Bendale (Tue 2:45 to 4:45 pm). Best if you email me in advance, but I
More informationImage Normalization for Illumination Compensation in Facial Images
Image Normalization for Illumination Compensation in Facial Images by Martin D. Levine, Maulin R. Gandhi, Jisnu Bhattacharyya Department of Electrical & Computer Engineering & Center for Intelligent Machines
More informationTowards better accuracy for Spam predictions
Towards better accuracy for Spam predictions Chengyan Zhao Department of Computer Science University of Toronto Toronto, Ontario, Canada M5S 2E4 czhao@cs.toronto.edu Abstract Spam identification is crucial
More informationIntroduction to Data Mining
Introduction to Data Mining Jay Urbain Credits: Nazli Goharian & David Grossman @ IIT Outline Introduction Data Pre-processing Data Mining Algorithms Naïve Bayes Decision Tree Neural Network Association
More informationSupport Vector Machine (SVM)
Support Vector Machine (SVM) CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
More informationReliability of MRI-derived measurements of human cerebral cortical thickness: The effects of field strength, scanner upgrade and manufacturer
www.elsevier.com/locate/ynimg NeuroImage 32 (2006) 180 194 Reliability of MRI-derived measurements of human cerebral cortical thickness: The effects of field strength, scanner upgrade and manufacturer
More informationII. RELATED WORK. Sentiment Mining
Sentiment Mining Using Ensemble Classification Models Matthew Whitehead and Larry Yaeger Indiana University School of Informatics 901 E. 10th St. Bloomington, IN 47408 {mewhiteh, larryy}@indiana.edu Abstract
More informationL25: Ensemble learning
L25: Ensemble learning Introduction Methods for constructing ensembles Combination strategies Stacked generalization Mixtures of experts Bagging Boosting CSCE 666 Pattern Analysis Ricardo Gutierrez-Osuna
More informationMultiple Kernel Learning on the Limit Order Book
JMLR: Workshop and Conference Proceedings 11 (2010) 167 174 Workshop on Applications of Pattern Analysis Multiple Kernel Learning on the Limit Order Book Tristan Fletcher Zakria Hussain John Shawe-Taylor
More informationDiscriminative Multimodal Biometric. Authentication Based on Quality Measures
Discriminative Multimodal Biometric Authentication Based on Quality Measures Julian Fierrez-Aguilar a,, Javier Ortega-Garcia a, Joaquin Gonzalez-Rodriguez a, Josef Bigun b a Escuela Politecnica Superior,
More informationClassifying Manipulation Primitives from Visual Data
Classifying Manipulation Primitives from Visual Data Sandy Huang and Dylan Hadfield-Menell Abstract One approach to learning from demonstrations in robotics is to make use of a classifier to predict if
More informationSupervised and unsupervised learning - 1
Chapter 3 Supervised and unsupervised learning - 1 3.1 Introduction The science of learning plays a key role in the field of statistics, data mining, artificial intelligence, intersecting with areas in
More informationMusic Mood Classification
Music Mood Classification CS 229 Project Report Jose Padial Ashish Goel Introduction The aim of the project was to develop a music mood classifier. There are many categories of mood into which songs may
More informationData Quality Mining: Employing Classifiers for Assuring consistent Datasets
Data Quality Mining: Employing Classifiers for Assuring consistent Datasets Fabian Grüning Carl von Ossietzky Universität Oldenburg, Germany, fabian.gruening@informatik.uni-oldenburg.de Abstract: Independent
More informationThe Effects of Start Prices on the Performance of the Certainty Equivalent Pricing Policy
BMI Paper The Effects of Start Prices on the Performance of the Certainty Equivalent Pricing Policy Faculty of Sciences VU University Amsterdam De Boelelaan 1081 1081 HV Amsterdam Netherlands Author: R.D.R.
More informationAward Number: W81XWH-10-1-0739
AD Award Number: W81XWH-10-1-0739 TITLE: Voxel-Wise Time-Series Analysis of Quantitative MRI in Relapsing-Remitting MS: Dynamic Imaging Metrics of Disease Activity Including Prelesional Changes PRINCIPAL
More informationNeovision2 Performance Evaluation Protocol
Neovision2 Performance Evaluation Protocol Version 3.0 4/16/2012 Public Release Prepared by Rajmadhan Ekambaram rajmadhan@mail.usf.edu Dmitry Goldgof, Ph.D. goldgof@cse.usf.edu Rangachar Kasturi, Ph.D.
More informationSearch Taxonomy. Web Search. Search Engine Optimization. Information Retrieval
Information Retrieval INFO 4300 / CS 4300! Retrieval models Older models» Boolean retrieval» Vector Space model Probabilistic Models» BM25» Language models Web search» Learning to Rank Search Taxonomy!
More informationSegmentation of Brain MR Images Through a Hidden Markov Random Field Model and the Expectation-Maximization Algorithm
IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. 20, NO. 1, JANUARY 2001 45 Segmentation of Brain MR Images Through a Hidden Markov Random Field Model and the Expectation-Maximization Algorithm Yongyue Zhang*,
More informationTHREE DIMENSIONAL REPRESENTATION OF AMINO ACID CHARAC- TERISTICS
THREE DIMENSIONAL REPRESENTATION OF AMINO ACID CHARAC- TERISTICS O.U. Sezerman 1, R. Islamaj 2, E. Alpaydin 2 1 Laborotory of Computational Biology, Sabancı University, Istanbul, Turkey. 2 Computer Engineering
More informationTHEORY, SIMULATION, AND COMPENSATION OF PHYSIOLOGICAL MOTION ARTIFACTS IN FUNCTIONAL MRI. Douglas C. Noll* and Walter Schneider
THEORY, SIMULATION, AND COMPENSATION OF PHYSIOLOGICAL MOTION ARTIFACTS IN FUNCTIONAL MRI Douglas C. Noll* and Walter Schneider Departments of *Radiology, *Electrical Engineering, and Psychology University
More informationA fast multi-class SVM learning method for huge databases
www.ijcsi.org 544 A fast multi-class SVM learning method for huge databases Djeffal Abdelhamid 1, Babahenini Mohamed Chaouki 2 and Taleb-Ahmed Abdelmalik 3 1,2 Computer science department, LESIA Laboratory,
More informationEmory University RT to Bachelor of Medical Science Degree Medical Imaging
Emory University RT to Bachelor of Medical Science Degree Medical Imaging Courses: All RT-BMSc students must complete all of the program s core courses and the courses specific to their selected minor
More information5 Factors Affecting the Signal-to-Noise Ratio
5 Factors Affecting the Signal-to-Noise Ratio 29 5 Factors Affecting the Signal-to-Noise Ratio In the preceding chapters we have learned how an MR signal is generated and how the collected signal is processed
More informationFeature Selection using Integer and Binary coded Genetic Algorithm to improve the performance of SVM Classifier
Feature Selection using Integer and Binary coded Genetic Algorithm to improve the performance of SVM Classifier D.Nithya a, *, V.Suganya b,1, R.Saranya Irudaya Mary c,1 Abstract - This paper presents,
More informationONLINE SUPPLEMENTARY DATA. Potential effect of skull thickening on the associations between cognition and brain atrophy in ageing
ONLINE SUPPLEMENTARY DATA Potential effect of skull thickening on the associations between cognition and brain atrophy in ageing Benjamin S. Aribisala 1,2,3, Natalie A. Royle 1,2,3, Maria C. Valdés Hernández
More informationTrading Strategies and the Cat Tournament Protocol
M A C H I N E L E A R N I N G P R O J E C T F I N A L R E P O R T F A L L 2 7 C S 6 8 9 CLASSIFICATION OF TRADING STRATEGIES IN ADAPTIVE MARKETS MARK GRUMAN MANJUNATH NARAYANA Abstract In the CAT Tournament,
More informationA Feature Selection Methodology for Steganalysis
A Feature Selection Methodology for Steganalysis Yoan Miche 1, Benoit Roue 2, Amaury Lendasse 1, Patrick Bas 12 1 Laboratory of Computer and Information Science Helsinki University of Technology P.O. Box
More informationComparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data
CMPE 59H Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data Term Project Report Fatma Güney, Kübra Kalkan 1/15/2013 Keywords: Non-linear
More informationThis unit will lay the groundwork for later units where the students will extend this knowledge to quadratic and exponential functions.
Algebra I Overview View unit yearlong overview here Many of the concepts presented in Algebra I are progressions of concepts that were introduced in grades 6 through 8. The content presented in this course
More informationExample: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.
Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C
More informationMRI DATA PROCESSING. Compiled by: Nicolas F. Lori and Carlos Ferreira. Introduction
MRI DATA PROCESSING Compiled by: Nicolas F. Lori and Carlos Ferreira Introduction Magnetic Resonance Imaging (MRI) is a clinical exam that is safe to the patient. Nevertheless, it s very important to attend
More informationData Mining. Nonlinear Classification
Data Mining Unit # 6 Sajjad Haider Fall 2014 1 Nonlinear Classification Classes may not be separable by a linear boundary Suppose we randomly generate a data set as follows: X has range between 0 to 15
More informationIntroduction to Machine Learning Lecture 1. Mehryar Mohri Courant Institute and Google Research mohri@cims.nyu.edu
Introduction to Machine Learning Lecture 1 Mehryar Mohri Courant Institute and Google Research mohri@cims.nyu.edu Introduction Logistics Prerequisites: basics concepts needed in probability and statistics
More informationWhitepapers on Imaging Infrastructure for Research Paper 1. General Workflow Considerations
Whitepapers on Imaging Infrastructure for Research Paper 1. General Workflow Considerations Bradley J Erickson, Tony Pan, Daniel J Marcus, CTSA Imaging Informatics Working Group Introduction The use of
More informationSVM Ensemble Model for Investment Prediction
19 SVM Ensemble Model for Investment Prediction Chandra J, Assistant Professor, Department of Computer Science, Christ University, Bangalore Siji T. Mathew, Research Scholar, Christ University, Dept of
More informationGetting Even More Out of Ensemble Selection
Getting Even More Out of Ensemble Selection Quan Sun Department of Computer Science The University of Waikato Hamilton, New Zealand qs12@cs.waikato.ac.nz ABSTRACT Ensemble Selection uses forward stepwise
More informationEarly defect identification of semiconductor processes using machine learning
STANFORD UNIVERISTY MACHINE LEARNING CS229 Early defect identification of semiconductor processes using machine learning Friday, December 16, 2011 Authors: Saul ROSA Anton VLADIMIROV Professor: Dr. Andrew
More informationA Sarsa based Autonomous Stock Trading Agent
A Sarsa based Autonomous Stock Trading Agent Achal Augustine The University of Texas at Austin Department of Computer Science Austin, TX 78712 USA achal@cs.utexas.edu Abstract This paper describes an autonomous
More informationMAXIMIZING RETURN ON DIRECT MARKETING CAMPAIGNS
MAXIMIZING RETURN ON DIRET MARKETING AMPAIGNS IN OMMERIAL BANKING S 229 Project: Final Report Oleksandra Onosova INTRODUTION Recent innovations in cloud computing and unified communications have made a
More informationDistributed forests for MapReduce-based machine learning
Distributed forests for MapReduce-based machine learning Ryoji Wakayama, Ryuei Murata, Akisato Kimura, Takayoshi Yamashita, Yuji Yamauchi, Hironobu Fujiyoshi Chubu University, Japan. NTT Communication
More informationCoding science news (intrinsic and extrinsic features)
Coding science news (intrinsic and extrinsic features) M I G U E L Á N G E L Q U I N T A N I L L A, C A R L O S G. F I G U E R O L A T A M A R G R O V E S 2 Science news in Spain The corpus of digital
More informationHYBRID PROBABILITY BASED ENSEMBLES FOR BANKRUPTCY PREDICTION
HYBRID PROBABILITY BASED ENSEMBLES FOR BANKRUPTCY PREDICTION Chihli Hung 1, Jing Hong Chen 2, Stefan Wermter 3, 1,2 Department of Management Information Systems, Chung Yuan Christian University, Taiwan
More informationThe End of Primary School Test Marleen van der Lubbe, Cito, The Netherlands
The End of Primary School Test Marleen van der Lubbe, Cito, The Netherlands The Dutch Education System Before explaining more about the End of Primary School Test (better known as Citotest), it is important
More informationIntelligent Tools For A Productive Radiologist Workflow: How Machine Learning Enriches Hanging Protocols
GE Healthcare Intelligent Tools For A Productive Radiologist Workflow: How Machine Learning Enriches Hanging Protocols Authors: Tianyi Wang Information Scientist Machine Learning Lab Software Science &
More informationLeveraging Ensemble Models in SAS Enterprise Miner
ABSTRACT Paper SAS133-2014 Leveraging Ensemble Models in SAS Enterprise Miner Miguel Maldonado, Jared Dean, Wendy Czika, and Susan Haller SAS Institute Inc. Ensemble models combine two or more models to
More informationFoundations of Machine Learning On-Line Learning. Mehryar Mohri Courant Institute and Google Research mohri@cims.nyu.edu
Foundations of Machine Learning On-Line Learning Mehryar Mohri Courant Institute and Google Research mohri@cims.nyu.edu Motivation PAC learning: distribution fixed over time (training and test). IID assumption.
More informationD-optimal plans in observational studies
D-optimal plans in observational studies Constanze Pumplün Stefan Rüping Katharina Morik Claus Weihs October 11, 2005 Abstract This paper investigates the use of Design of Experiments in observational
More informationEnhanced Boosted Trees Technique for Customer Churn Prediction Model
IOSR Journal of Engineering (IOSRJEN) ISSN (e): 2250-3021, ISSN (p): 2278-8719 Vol. 04, Issue 03 (March. 2014), V5 PP 41-45 www.iosrjen.org Enhanced Boosted Trees Technique for Customer Churn Prediction
More informationGE Medical Systems Training in Partnership. Module 12: Spin Echo
Module : Spin Echo Spin Echo Objectives Review the SE PSD. Review the concepts of T, T, and T*. Spin Echo PSD RF Gz Gy 90 80 Gx Spin Echo - SE Spin echo is a standard pulse sequence on Signa MRi/LX and
More informationDescriptive statistics Statistical inference statistical inference, statistical induction and inferential statistics
Descriptive statistics is the discipline of quantitatively describing the main features of a collection of data. Descriptive statistics are distinguished from inferential statistics (or inductive statistics),
More informationadvancing magnetic resonance imaging and medical data analysis
advancing magnetic resonance imaging and medical data analysis INFN - CSN5 CdS - Luglio 2015 Sezione di Genova 2015-2017 objectives Image reconstruction, artifacts and noise management O.1 Components RF
More informationAnalysis of Micro-Macro Transformations of Railway Networks
Konrad-Zuse-Zentrum für Informationstechnik Berlin Takustraße 7 D-14195 Berlin-Dahlem Germany MARCO BLANCO THOMAS SCHLECHTE Analysis of Micro-Macro Transformations of Railway Networks Zuse Institute Berlin
More informationA new Approach for Intrusion Detection in Computer Networks Using Data Mining Technique
A new Approach for Intrusion Detection in Computer Networks Using Data Mining Technique Aida Parbaleh 1, Dr. Heirsh Soltanpanah 2* 1 Department of Computer Engineering, Islamic Azad University, Sanandaj
More informationEnsemble Methods. Knowledge Discovery and Data Mining 2 (VU) (707.004) Roman Kern. KTI, TU Graz 2015-03-05
Ensemble Methods Knowledge Discovery and Data Mining 2 (VU) (707004) Roman Kern KTI, TU Graz 2015-03-05 Roman Kern (KTI, TU Graz) Ensemble Methods 2015-03-05 1 / 38 Outline 1 Introduction 2 Classification
More informationData Mining Practical Machine Learning Tools and Techniques
Ensemble learning Data Mining Practical Machine Learning Tools and Techniques Slides for Chapter 8 of Data Mining by I. H. Witten, E. Frank and M. A. Hall Combining multiple models Bagging The basic idea
More informationNeuroimaging module I: Modern neuroimaging methods of investigation of the human brain in health and disease
1 Neuroimaging module I: Modern neuroimaging methods of investigation of the human brain in health and disease The following contains a summary of the content of the neuroimaging module I on the postgraduate
More informationRecognizing Cats and Dogs with Shape and Appearance based Models. Group Member: Chu Wang, Landu Jiang
Recognizing Cats and Dogs with Shape and Appearance based Models Group Member: Chu Wang, Landu Jiang Abstract Recognizing cats and dogs from images is a challenging competition raised by Kaggle platform
More informationA Modified Fuzzy C-Means Algorithm for Bias Field Estimation and Segmentation of MRI Data
IEEE TRANSACTIONS ON MEDICAL IMAGING, VOL. 21, NO. 3, MARCH 2002 193 A Modified Fuzzy C-Means Algorithm for Bias Field Estimation and Segmentation of MRI Data Mohamed N. Ahmed, Member, IEEE, Sameh M. Yamany,
More informationNeural Network Add-in
Neural Network Add-in Version 1.5 Software User s Guide Contents Overview... 2 Getting Started... 2 Working with Datasets... 2 Open a Dataset... 3 Save a Dataset... 3 Data Pre-processing... 3 Lagging...
More information