Deep Learning For Text Processing



Similar documents
Sense Making in an IOT World: Sensor Data Analysis with Deep Learning

The GPU Accelerated Data Center. Marc Hamilton, August 27, 2015

Steven C.H. Hoi School of Information Systems Singapore Management University

THE WORLD LEADER IN VISUAL COMPUTING

REGULATIONS FOR THE DEGREE OF MASTER OF SCIENCE IN COMPUTER SCIENCE (MSc[CompSc])

Chapter 19 Cloud Computing for Multimedia Services

Applications of Deep Learning to the GEOINT mission. June 2015

Amazon EC2 Product Details Page 1 of 5

BIOINF 585 Fall 2015 Machine Learning for Systems Biology & Clinical Informatics

Data Mining: STATISTICA

OCIO HIGH RESOLUTION GRAPHICS DESKTOP VIRTUALIZATION POC. Chip Charnley Technical Expert Client Technologies Infrastructure Architecture

1. INTRODUCTION Graphics 2

REGULATIONS FOR THE DEGREE OF MASTER OF SCIENCE IN COMPUTER SCIENCE (MSc[CompSc])

REGULATIONS FOR THE DEGREE OF MASTER OF SCIENCE IN COMPUTER SCIENCE (MSc[CompSc])

Deep Learning Meets Heterogeneous Computing. Dr. Ren Wu Distinguished Scientist, IDL, Baidu

Course Development of Programming for General-Purpose Multicore Processors

Acoustical Surfaces, Inc.

GPU Programming in Computer Vision

Sentiment Analysis. D. Skrepetos 1. University of Waterloo. NLP Presenation, 06/17/2015

Medical Image Processing on the GPU. Past, Present and Future. Anders Eklund, PhD Virginia Tech Carilion Research Institute

E6895 Advanced Big Data Analytics Lecture 14:! NVIDIA GPU Examples and GPU on ios devices

Programming models for heterogeneous computing. Manuel Ujaldón Nvidia CUDA Fellow and A/Prof. Computer Architecture Department University of Malaga

Introduction to GPU Programming Languages

WROX Certified Big Data Analyst Program by AnalytixLabs and Wiley

NVIDIA GRID DASSAULT CATIA V5/V6 SCALABILITY GUIDE. NVIDIA Performance Engineering Labs PerfEngDoc-SG-DSC01v1 March 2016

MulticoreWare. Global Company, 250+ employees HQ = Sunnyvale, CA Other locations: US, China, India, Taiwan

Software Development Training Camp 1 (0-3) Prerequisite : Program development skill enhancement camp, at least 48 person-hours.

The Future Of Animation Is Games

DEEP LEARNING WITH GPUS

GPU Renderfarm with Integrated Asset Management & Production System (AMPS)

GPU System Architecture. Alan Gray EPCC The University of Edinburgh

COPYRIGHTED MATERIAL. Contents. List of Figures. Acknowledgments

CIKM 2015 Melbourne Australia Oct. 22, 2015 Building a Better Connected World with Data Mining and Artificial Intelligence Technologies

Computer Graphics Hardware An Overview

Towards Fast SQL Query Processing in DB2 BLU Using GPUs A Technology Demonstration. Sina Meraji sinamera@ca.ibm.com

Winter 2016 Course Timetable. Legend: TIME: M = Monday T = Tuesday W = Wednesday R = Thursday F = Friday BREATH: M = Methodology: RA = Research Area

HIGH PERFORMANCE BIG DATA ANALYTICS

MetaMorph Microscopy Automation & Image Analysis Software Super-Resolution Module

Deep Learning GPU-Based Hardware Platform

An Introduction to Data Mining

Applying Deep Learning to Car Data Logging (CDL) and Driver Assessor (DA) October 22-Oct-15

NVIDIA VIDEO ENCODER 5.0

PhD in Computer Science and Engineering Bologna, April Machine Learning. Marco Lippi. Marco Lippi Machine Learning 1 / 80

Increasing Flash Throughput for Big Data Applications (Data Management Track)

NVIDIA GeForce GTX 580 GPU Datasheet

The multilayer sentiment analysis model based on Random forest Wei Liu1, Jie Zhang2

CPSC 340: Machine Learning and Data Mining. Mark Schmidt University of British Columbia Fall 2015

Faculty of Science School of Mathematics and Statistics

What is Artificial Intelligence?

LCMON Network Traffic Analysis

Virtual Desktop VMware View Horizon

Data Mining. Nonlinear Classification

Application of Data Visualization

How To Build An Ark Processor With An Nvidia Gpu And An African Processor

Writing Applications for the GPU Using the RapidMind Development Platform

GPU-Based Deep Learning Inference:

Machine learning for algo trading

An Introduction to Deep Learning

ECBDL 14: Evolu/onary Computa/on for Big Data and Big Learning Workshop July 13 th, 2014 Big Data Compe//on

ipad, a revolutionary device - Apple

S4726 VIRTUALIZATION AN INTRO TO VIRTUALIZATION. Luke Wignall & Jared Cowart Senior Solution Architects GRID

Texture Cache Approximation on GPUs

NVIDIA AUTOMOTIVE. Driving Innovation

Secured Embedded Many-Core Accelerator for Big Data Processing

Benchmark Hadoop and Mars: MapReduce on cluster versus on GPU

Learning outcomes. Knowledge and understanding. Competence and skills

CUDA in the Cloud Enabling HPC Workloads in OpenStack With special thanks to Andrew Younge (Indiana Univ.) and Massimo Bernaschi (IAC-CNR)

Government of Russian Federation. Faculty of Computer Science School of Data Analysis and Artificial Intelligence

Configuring RemoteFX on Windows Server 2012 R2

Data Analytics at NICTA. Stephen Hardy National ICT Australia (NICTA)

Table of Contents. P a g e 2

Taking Inverse Graphics Seriously

Pool Play Brackets 3 TEAM POOL PLAY FORMAT. Game 1 1 vs. 2* Game 2 2 vs. 3* Game 3 3 vs. 1* Example 3 Team Pool Play 4 TEAM POOL PLAY FORMAT

Computer Literacy. Hardware & Software Classification

Machine Learning Business Intelligence, Culturomics and Life Sciences

Bachelor Degree in Informatics Engineering Master courses

Learning to Process Natural Language in Big Data Environment

NVIDIA GRID OVERVIEW SERVER POWERED BY NVIDIA GRID. WHY GPUs FOR VIRTUAL DESKTOPS AND APPLICATIONS? WHAT IS A VIRTUAL DESKTOP?

ACCELERATING SELECT WHERE AND SELECT JOIN QUERIES ON A GPU

An Early Attempt at Applying Deep Reinforcement Learning to the Game 2048

Curriculum Vitae Anders Eklund

Machine Learning. CUNY Graduate Center, Spring Professor Liang Huang.

STORAGE HIGH SPEED INTERCONNECTS HIGH PERFORMANCE COMPUTING VISUALISATION GPU COMPUTING

Graphical Processing Units to Accelerate Orthorectification, Atmospheric Correction and Transformations for Big Data

Machine Learning for Data Science (CS4786) Lecture 1

~ Greetings from WSU CAPPLab ~

OpenPOWER Outlook AXEL KOEHLER SR. SOLUTION ARCHITECT HPC

Accelerating I/O- Intensive Applications in IT Infrastructure with Innodisk FlexiArray Flash Appliance. Alex Ho, Product Manager Innodisk Corporation

Trends in High-Performance Computing for Power Grid Applications

HETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK

The High Performance Internet of Things: using GVirtuS for gluing cloud computing and ubiquitous connected devices

Transcription:

Deep Learning For Text Processing Jeffrey A. Bilmes Professor Departments of Electrical Engineering & Computer Science and Engineering University of Washington, Seattle http://melodi.ee.washington.edu/~bilmes Monday, July 14th, 2014 J. Bilmes Deep Learning For Text Processing Microsoft Research Faculty Summit, 2014 page 2 / 6

Deep Neural Networks Aspects: Deep Generative Models and Deep Non-linear Mappings (classification & regression) Convolutional Deep Models and Aggregation methods (sum, max pooling, etc.) Non-Linear Learning of Feature Representations, Autoencoders Distributed Representations (efficiency of parameter utilization). Applications and Successes: Computer Vision (image, video, motion capture) Speech (acoustic modeling in speech, speech perception, continuous representations of words/phrases, etc.) and NLP (language modeling, sentiment classification, etc.) Information Retrieval, learning similarity, deep CCA, etc. J. Bilmes Deep Learning For Text Processing Microsoft Research Faculty Summit, 2014 page 3 / 6

Deep Neural Networks Aspects: Deep Generative Models and Deep Non-linear Mappings (classification & regression) Convolutional Deep Models and Aggregation methods (sum, max pooling, etc.) Non-Linear Learning of Feature Representations, Autoencoders Distributed Representations (efficiency of parameter utilization). Applications and Successes: Computer Vision (image, video, motion capture) Speech (acoustic modeling in speech, speech perception, continuous representations of words/phrases, etc.) and NLP (language modeling, sentiment classification, etc.) Information Retrieval, learning similarity, deep CCA, etc. Resources: Big data - big data is different data, big neural networks are different neural networks Big computation - CPU only computation is insufficient, while GPUs are feasible J. Bilmes Deep Learning For Text Processing Microsoft Research Faculty Summit, 2014 page 3 / 6

GPU vs. CPU Performance Over Time Source: GTC (GPU Tech. Conference) Keynote, NVIDIA CEO Jen-Hsun Huang, 3/25/2014 J. Bilmes Deep Learning For Text Processing Microsoft Research Faculty Summit, 2014 page 4 / 6

Deep Models and GPUs J. Bilmes Deep Learning For Text Processing Microsoft Research Faculty Summit, 2014 page 5 / 6

Questions Deep Learning needs big data to train, and is well-matched to GPUs (for matrix operations) to do so. J. Bilmes Deep Learning For Text Processing Microsoft Research Faculty Summit, 2014 page 6 / 6

Questions Deep Learning needs big data to train, and is well-matched to GPUs (for matrix operations) to do so. GPUs: commodity high-performance supercomputing (thanks to gaming industry). Hence, inexpensive and ubiquitous. J. Bilmes Deep Learning For Text Processing Microsoft Research Faculty Summit, 2014 page 6 / 6

Questions Deep Learning needs big data to train, and is well-matched to GPUs (for matrix operations) to do so. GPUs: commodity high-performance supercomputing (thanks to gaming industry). Hence, inexpensive and ubiquitous. Part of the success of deep models is due to wave of GPU performance increase, and that deep models computational needs are well-matched with what GPUs offer. J. Bilmes Deep Learning For Text Processing Microsoft Research Faculty Summit, 2014 page 6 / 6

Questions Deep Learning needs big data to train, and is well-matched to GPUs (for matrix operations) to do so. GPUs: commodity high-performance supercomputing (thanks to gaming industry). Hence, inexpensive and ubiquitous. Part of the success of deep models is due to wave of GPU performance increase, and that deep models computational needs are well-matched with what GPUs offer. Q: Could other machine learning methods (e.g., kernel machines, random forests, graphical models, etc.) have a correspondingly well matched, inexpensive, ubiquitous, and high performance hardware platform to run on? J. Bilmes Deep Learning For Text Processing Microsoft Research Faculty Summit, 2014 page 6 / 6

Questions Deep Learning needs big data to train, and is well-matched to GPUs (for matrix operations) to do so. GPUs: commodity high-performance supercomputing (thanks to gaming industry). Hence, inexpensive and ubiquitous. Part of the success of deep models is due to wave of GPU performance increase, and that deep models computational needs are well-matched with what GPUs offer. Q: Could other machine learning methods (e.g., kernel machines, random forests, graphical models, etc.) have a correspondingly well matched, inexpensive, ubiquitous, and high performance hardware platform to run on? Q: Or is the nature of such neural computations (matrix operations) such that it will always have an advantage due its regularity and hardware optimizability? J. Bilmes Deep Learning For Text Processing Microsoft Research Faculty Summit, 2014 page 6 / 6

Questions Deep Learning needs big data to train, and is well-matched to GPUs (for matrix operations) to do so. GPUs: commodity high-performance supercomputing (thanks to gaming industry). Hence, inexpensive and ubiquitous. Part of the success of deep models is due to wave of GPU performance increase, and that deep models computational needs are well-matched with what GPUs offer. Q: Could other machine learning methods (e.g., kernel machines, random forests, graphical models, etc.) have a correspondingly well matched, inexpensive, ubiquitous, and high performance hardware platform to run on? Q: Or is the nature of such neural computations (matrix operations) such that it will always have an advantage due its regularity and hardware optimizability? Q: If so, can text/nlp applications like parsing, SMT be made just as regular? J. Bilmes Deep Learning For Text Processing Microsoft Research Faculty Summit, 2014 page 6 / 6