Accurate is not always good: How Accuracy Metrics have hurt Recommender Systems
|
|
|
- Amy Brooks
- 10 years ago
- Views:
Transcription
1 Accurate is not always good: How Accuracy Metrics have hurt Recommender Systems Sean M. McNee John Riedl Joseph A. Konstan Copyright is held by the author/owner(s). CHI 2006, April 22 27, 2006, Montreal, Canada. ACM 1-xxxxxxxxxxxxxxxxxx. Abstract Recommender systems have shown great potential to help users find interesting and relevant items from within a large information space. Most research up to this point has focused on improving the accuracy of recommender systems. We believe that not only has this narrow focus been misguided, but has even been detrimental to the field. The recommendations that are most accurate according to the standard metrics are sometimes not the recommendations that are most useful to users. In this paper, we propose informal arguments that the recommender community should move beyond the conventional accuracy metrics and their associated experimental methodologies. We propose new user-centric directions for evaluating recommender systems. Keywords Recommender systems, personalization, collaborative filtering, metrics ACM Classification Keywords H3.3. Information Search and Retrieval: Information Filtering, Search Process, Retrieval Models
2 2 Introduction Imagine you are using a travel recommender system. Suppose all of the recommendations it gives to you are for places you have already traveled to? Even if the system was very good at ranking all of the places you have visited in order of preference, this still would be a poor recommender system. Would you use such a system? Unfortunately, this is exactly how we currently test our recommender systems. In the standard methodology, the travel recommender would be penalized for recommending new locations instead of places the users have already visited! Current accuracy metrics, such as MAE [Herlocker 1999], measure recommender algorithm performance by comparing the algorithm s prediction against a user s rating of an item. The most commonly used methodology with these metrics is the leave-n-out approach [Breese 1998] where a percentage of the dataset is withheld from the recommender and used as test data. In essence, we reward a travel recommender for recommending places a user has already visited, instead of rewarding it for finding new places for the user to visit. By focusing on this way of testing recommenders, are we really helping users find the items they are interested in? We claim there are many different aspects to the recommendation process which current accuracy metrics do not measure. In this paper, we will review three such aspects: the similarity of recommendation lists, recommendation serendipity, and the importance of user needs and expectations in a recommender. We will review how current methodologies fail for each aspect, and provide suggestions for improvement. Similarity More frequently than not, recommendation lists contain similar items on them. Going to Amazon.com for a book by Robert Heinlein, for example, will give you a recommendation list full of all of his other books. We have seen this behavior in algorithms as well. The Item-Item collaborative filtering algorithm can trap users in a similarity hole, only giving exceptionally similar recommendations (e.g. once a user rated one Star Trek movie she would only receive recommendations for more Star Trek movies) [Rashid 2001]. This problem is more noticeable when there is less data on which to base recommendations, such as for new users to a system. It is these times when a poor recommendation could convince a user to leave the recommender forever. Accuracy metrics cannot see this problem because they are designed to judge the accuracy of individual item predictions; they do not judge the contents of entire recommendation lists. Unfortunately, it is these lists that the users interact with. All recommendations are made in the context of the current recommendation list and the previous lists the user has already seen. The recommendation list should be judged for its usefulness as a complete entity, not just as a collection of individual items. One approach to solving this problem was proposed in [Ziegler 2005], with the introduction of the Intra-List Similarity Metric and the process of Topic Diversification for recommendation lists. Returned lists can be altered to either increase or decrease the diversity of items on that list. Results showed that these altered lists performed worse on accuracy measures than unchanged lists, but users preferred the altered lists.
3 3 Depending on the user s intentions, the makeup of items appearing on the list affected the user s satisfaction with the recommender more than the changes in the accuracy of the item on the list. Serendipity Serendipity in a recommender is the experience of receiving an unexpected and fortuitous item recommendation. There is a level of emotional response associated with serendipity that is difficult to capture in any metric. But even if we remove that component, the unexpectedness part of this concept the novelty of the received recommendations is still difficult to measure. The converse of this concept, the ratability of received recommendations, is quite easy to measure, however. And it is ratability that a leave-nout approach measures. When applied to a classification problem, a leave-n-out methodology works great: items belonging to a particular segment are withheld and the classifier is judged on how it classifies these items given what it knows of the dataset. Similarly, a leave-n-out methodology judges a recommender algorithm by how well it classifies, or recommends, withheld items to the users they were withheld from. We define the ratability of an item in a recommender to be the probability that this item will be the next item that the user will consume (and then rate) given what the system knows of the user s profile. From a machine learning point-of-view, the item with the highest ratability would be the next item a classifier would place into the user s profile. Thus, recommender algorithms which score well on accuracy metrics using a leave-n-out methodology can generate recommendation for items with high ratability. These algorithms are good at predicting what else will appear in that user s profile. Indeed, many machine learning algorithms score very well as recommenders based on a leave-n-out methodology [Billsus 1998, Breese 1998]. The implicit assumption is that a user is always interested in the items with the highest ratability. While this assumption is true in classification problems, we claim it often isn t true in recommenders. Users often judge recommendations that are for items they would not have thought of themselves. For instance, we once were helping an online music store with recommendations using a User-User Collaborative Filtering algorithm. The most common recommendation was for the Beatle s White Album. From an accuracy perspective these recommendations were dead-on: most users like that album very much. From a usefulness perspective, though, the recommendations were a complete failure: every user either already owned the White Album, or had specifically chosen not to own it. Even though it was highly ratable, White Album recommendations were almost never acted on by users, because they added almost no value. Our previous research [McNee 2002, Torres 2004] has given us one more piece to this puzzle: recommender algorithms generate qualitatively different recommendations lists from each other. When asked, users preferred lists from different recommender algorithms based on their current task. Users chose different words to describe the recommendations they received (e.g. User-based collaborative filtering was considered to generate novel recommendations).
4 4 This suggests that we need other ways to classify recommender algorithms. While a serendipity metric may be difficult to create without feedback from users, other metrics to judge a variety of algorithm aspects would provide a more detailed pictures of the differences between recommender algorithms. User Experiences and Expectations As we have shown in previous work, user satisfaction does not always correlate with high recommender accuracy [McNee 2002, Ziegler 2005]. There are many other factors important to users that need to be considered. New users have different needs from experienced users in a recommender. New users may benefit from an algorithm which generates highly ratable items, as they need to establish trust and rapport with a recommender before taking advantage of the recommendations it offers. Previous work shows that the choice of algorithm used for new users greatly affects the user s experience and the accuracy of the recommendations the system could generate for them [Rashid 2001]. Our previous work also suggested that differences in language and cultural background influenced user satisfaction [Torres 2004]. A recommender in a user s native language was greatly preferred to one in an alternate language, even if the items themselves recommended were in the alternate language (e.g. a Portuguese-based research paper recommender recommending papers written in English). Moving Forward Accuracy metrics have greatly helped the field of recommender systems; they have given us a way to compare algorithms and create robust experimental designs. We do not claim that we should stop using them. We just cannot use them alone to judge recommenders. Now, we need to think closely about the users of recommender systems. They don t care about using an algorithm that scored better on a metric, they want a meaningful recommendation. There a few ways we can do this. First, we need to judge the quality of recommendations as users see them: as recommendation lists. To do this, we need to create a variety of metrics which act on recommendation lists, not on items appearing in a list. There are already a few, such as the Intra-List Similarity metric, but we need more in order to understand other aspects of these lists. Second, we need to understand the differences between recommender algorithms and measure them in ways beyond their ratability. Users can tell the difference between recommender algorithms. For example, when we changed the algorithm running the MovieLens movie recommender, we received many s from users wondering why MovieLens had become so conservative with its recommendations. Both algorithms scored well on MAE measures, but were clearly different from each other. Finally, users return to recommenders over a period of time, growing from new users to experienced users. Each time they come to the system, they have some reason for coming: they have a purpose. We need to judge the recommendations we generate for each user based on whether or not we were able to meet their need. Until we acknowledge this relationship with
5 5 users, recommenders will continue to generate mismatched recommendations [Zaslow 2002]. In the end, recommenders exist to help users. We, as a community, have created many algorithms and approaches to studying recommender algorithms. It is now time to also study recommenders from a usercentric perspective to make them not only accurate and helpful, but also a pleasure to use. Acknowledgements We would like to thank our collaborators, Cai-Nicolas Ziegler and Roberto Torres, and all of GroupLens Research. This work was funded by the National Science Foundation, grants DGE , IIS , IIS IIS , and IIS References [1] Billsus, D., and Pazzani, M.J. Learning Collaborative Information Filters., In Proc. of ICML 1998, Morgan Kaufmann (1998), [2] Breese, J.S., Heckerman, D., and Kadie, C. Empirical Analysis of Predictive Algorithms for Collaborative Filtering. In Proc. of UAI 1998, Morgan Kaufmann (1998), [3] Herlocker, J.L., Konstan, J.A., Terveen, L.G., and Riedl, J.T. Evaluating Collaborative Filtering Recommender Systems. ACM TOIS 22, 1 (2004), [4] Herlocker, J.L., Konstan, J.A., Borchers, A., and Riedl, J. An algorithmic framework for performing collaborative filtering. In Proc. of ACM SIGIR 1999, ACM Press (1999), [5] McNee, S.M., Albert, I., Cosley, D., Gopalkrishnan, P., Lam, S.K., Rashid, A.M., Konstan, J.A. and Riedl, J. On the Recommending of Citations for Research Papers. In Proc. of ACM CSCW 2002, ACM Press (2002), [6] Rashid, A.M., Albert I., Cosley D., Lam, S.K., McNee, S.M., Konstan, J.A. and Riedl, J. Getting to know you: learning new user preferences in recommender systems. In Proc. of ACM IUI 2002, ACM Press (2002), [7] Torres, R., McNee, S.M., Abel, M., Konstan, J.A., and Riedl, J. Enhancing digital libraries with TechLens+. In Proc. of ACM/IEEE JCDL 2004, ACM Press (2004) [8] Zaslow, J. If TiVo Thinks You Are Gay, Here's How To Set It Straight --- Amazon.com Knows You, Too, Based on What You Buy; Why All the Cartoons? The Wall Street Journal, sect. A, p. 1, November 26, [9] Ziegler, C.N., McNee, S.M., Konstan, J.A., and Lausen, G., Improving Recommendation Lists through Topic Diversification. In Proc. of WWW 2005, ACM Press (2005),
Recommender Systems for Large-scale E-Commerce: Scalable Neighborhood Formation Using Clustering
Recommender Systems for Large-scale E-Commerce: Scalable Neighborhood Formation Using Clustering Badrul M Sarwar,GeorgeKarypis, Joseph Konstan, and John Riedl {sarwar, karypis, konstan, riedl}@csumnedu
4, 2, 2014 ISSN: 2277 128X
Volume 4, Issue 2, February 2014 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Recommendation
The Wisdom of Networks: Matching Recommender Systems and Social Network Theories
The Wisdom of Networks: Matching Recommender Systems and Social Network Theories R. Dandi 1 Abstract. This paper aims to analyzing the match between social network theories and recommender systems. Several
Improving Recommendation Lists Through Topic Diversification
Improving Recommendation Lists Through Topic Diversification Cai-Nicolas Ziegler 1 Sean M. McNee ABSTRACT 1 Institut für Informatik, Universität Freiburg Georges-Köhler-Allee, Gebäude Nr. 51 79110 Freiburg
arxiv:1506.04135v1 [cs.ir] 12 Jun 2015
Reducing offline evaluation bias of collaborative filtering algorithms Arnaud de Myttenaere 1,2, Boris Golden 1, Bénédicte Le Grand 3 & Fabrice Rossi 2 arxiv:1506.04135v1 [cs.ir] 12 Jun 2015 1 - Viadeo
Recommender Systems in E-Commerce
Recommender Systems in E-Commerce J. Ben Schafer, Joseph Konstan, John Riedl GroupLens Research Project Department of Computer Science and Engineering University of Minnesota Minneapolis, MN 55455 1-612-625-4002
Study on Recommender Systems for Business-To-Business Electronic Commerce
Study on Recommender Systems for Business-To-Business Electronic Commerce Xinrui Zhang Shanghai University of Science and Technology, Shanghai, China, 200093 Phone: 86-21-65681343, Email: [email protected]
Contact Recommendations from Aggegrated On-Line Activity
Contact Recommendations from Aggegrated On-Line Activity Abigail Gertner, Justin Richer, and Thomas Bartee The MITRE Corporation 202 Burlington Road, Bedford, MA 01730 {gertner,jricher,tbartee}@mitre.org
BUILDING A PREDICTIVE MODEL AN EXAMPLE OF A PRODUCT RECOMMENDATION ENGINE
BUILDING A PREDICTIVE MODEL AN EXAMPLE OF A PRODUCT RECOMMENDATION ENGINE Alex Lin Senior Architect Intelligent Mining [email protected] Outline Predictive modeling methodology k-nearest Neighbor
A NURSING CARE PLAN RECOMMENDER SYSTEM USING A DATA MINING APPROACH
Proceedings of the 3 rd INFORMS Workshop on Data Mining and Health Informatics (DM-HI 8) J. Li, D. Aleman, R. Sikora, eds. A NURSING CARE PLAN RECOMMENDER SYSTEM USING A DATA MINING APPROACH Lian Duan
Comparing Recommendations Made by Online Systems and Friends
Comparing Recommendations Made by Online Systems and Friends Rashmi Sinha and Kirsten Swearingen SIMS, University of California Berkeley, CA 94720 {sinha, kirstens}@sims.berkeley.edu Abstract: The quality
A Web Recommender System for Recommending, Predicting and Personalizing Music Playlists
A Web Recommender System for Recommending, Predicting and Personalizing Music Playlists Zeina Chedrawy 1, Syed Sibte Raza Abidi 1 1 Faculty of Computer Science, Dalhousie University, Halifax, Canada {chedrawy,
Automated Collaborative Filtering Applications for Online Recruitment Services
Automated Collaborative Filtering Applications for Online Recruitment Services Rachael Rafter, Keith Bradley, Barry Smyth Smart Media Institute, Department of Computer Science, University College Dublin,
Achieve Better Ranking Accuracy Using CloudRank Framework for Cloud Services
Achieve Better Ranking Accuracy Using CloudRank Framework for Cloud Services Ms. M. Subha #1, Mr. K. Saravanan *2 # Student, * Assistant Professor Department of Computer Science and Engineering Regional
Web Mining Seminar CSE 450. Spring 2008 MWF 11:10 12:00pm Maginnes 113
CSE 450 Web Mining Seminar Spring 2008 MWF 11:10 12:00pm Maginnes 113 Instructor: Dr. Brian D. Davison Dept. of Computer Science & Engineering Lehigh University [email protected] http://www.cse.lehigh.edu/~brian/course/webmining/
Utility of Distrust in Online Recommender Systems
Utility of in Online Recommender Systems Capstone Project Report Uma Nalluri Computing & Software Systems Institute of Technology Univ. of Washington, Tacoma [email protected] Committee: nkur Teredesai
Application of Dimensionality Reduction in Recommender System -- A Case Study
Application of Dimensionality Reduction in Recommender System -- A Case Study Badrul M. Sarwar, George Karypis, Joseph A. Konstan, John T. Riedl GroupLens Research Group / Army HPC Research Center Department
The Need for Training in Big Data: Experiences and Case Studies
The Need for Training in Big Data: Experiences and Case Studies Guy Lebanon Amazon Background and Disclaimer All opinions are mine; other perspectives are legitimate. Based on my experience as a professor
International Journal of Engineering Research-Online A Peer Reviewed International Journal Articles available online http://www.ijoer.
REVIEW ARTICLE ISSN: 2321-7758 UPS EFFICIENT SEARCH ENGINE BASED ON WEB-SNIPPET HIERARCHICAL CLUSTERING MS.MANISHA DESHMUKH, PROF. UMESH KULKARNI Department of Computer Engineering, ARMIET, Department
PolyLens: A Recommender System for Groups of Users
PolyLens: A Recommender System for Groups of Users Mark O Connor, Dan Cosley, Joseph A. Konstan, and John Riedl Department of Computer Science and Engineering, University of Minnesota, Minneapolis, MN
Recommending Social Events from Mobile Phone Location Data
Recommending Social Events from Mobile Phone Location Data Daniele Quercia Neal Lathia Francesco Calabrese Giusy Di Lorenzo Jon Crowcroft University of Cambridge Massachusetts Institute of Technology University
The Wisdom of the Few
The Wisdom of the Few A Collaborative Filtering Approach Based on Expert Opinions from the Web Xavier Amatriain Telefonica Research Via Augusta, 177 Barcelona 08021, Spain [email protected] Haewoon Kwak KAIST
Integrating social networks in an IPTV recommender system
Integrating social networks in an IPTV recommender system Cerdeira, Frederico [email protected] Almeida, Pedro [email protected] Abreu, Jorge [email protected] Almeida, Pedro L. [email protected] Cardoso,
How To Write A Summary Of A Review
PRODUCT REVIEW RANKING SUMMARIZATION N.P.Vadivukkarasi, Research Scholar, Department of Computer Science, Kongu Arts and Science College, Erode. Dr. B. Jayanthi M.C.A., M.Phil., Ph.D., Associate Professor,
Recommender Systems: Content-based, Knowledge-based, Hybrid. Radek Pelánek
Recommender Systems: Content-based, Knowledge-based, Hybrid Radek Pelánek 2015 Today lecture, basic principles: content-based knowledge-based hybrid, choice of approach,... critiquing, explanations,...
The Application of Data-Mining to Recommender Systems
The Application of Data-Mining to Recommender Systems J. Ben Schafer, Ph.D. University of Northern Iowa INTRODUCTION In a world where the number of choices can be overwhelming, recommender systems help
I like it... I like it not: Evaluating User Ratings Noise in Recommender Systems
I like it... I like it not: Evaluating User Ratings Noise in Recommender Systems Xavier Amatriain, Josep M. Pujol, and Nuria Oliver Telefonica Research Recent growing interest in predicting and influencing
CAS-ICT at TREC 2005 SPAM Track: Using Non-Textual Information to Improve Spam Filtering Performance
CAS-ICT at TREC 2005 SPAM Track: Using Non-Textual Information to Improve Spam Filtering Performance Shen Wang, Bin Wang and Hao Lang, Xueqi Cheng Institute of Computing Technology, Chinese Academy of
Integrating Collaborative Filtering and Sentiment Analysis: A Rating Inference Approach
Integrating Collaborative Filtering and Sentiment Analysis: A Rating Inference Approach Cane Wing-ki Leung and Stephen Chi-fai Chan and Fu-lai Chung 1 Abstract. We describe a rating inference approach
Content-Boosted Collaborative Filtering for Improved Recommendations
Proceedings of the Eighteenth National Conference on Artificial Intelligence(AAAI-2002), pp. 187-192, Edmonton, Canada, July 2002 Content-Boosted Collaborative Filtering for Improved Recommendations Prem
Recommendation Tool Using Collaborative Filtering
Recommendation Tool Using Collaborative Filtering Aditya Mandhare 1, Soniya Nemade 2, M.Kiruthika 3 Student, Computer Engineering Department, FCRIT, Vashi, India 1 Student, Computer Engineering Department,
User-centric vs. System-centric Evaluation of Recommender Systems
User-centric vs. System-centric Evaluation of Recommender Systems Paolo Cremonesi 1, Franca Garzotto 1, Roberto Turrin 2 1 Politecnico di Milano, Milano, Italy 2 ContentWise, Milano, Italy [paolo.cremonesi,
PRODUCTS AND SERVICES RECOMMENDATION SYSTEMS IN E-COMMERCE. RECOMMENDATION METHODS, ALGORITHMS, AND MEASURES OF THEIR EFFECTIVENESS
INFORMATYKA EKONOMICZNA BUSINESS INFORMATICS 1(31) 2014 ISSN 1507-3858 Mirosława Lasek, Dominik Kosieradzki University of Warsaw PRODUCTS AND SERVICES RECOMMENDATION SYSTEMS IN E-COMMERCE. RECOMMENDATION
Overview. Introduction. Recommender Systems & Slope One Recommender. Distributed Slope One on Mahout and Hadoop. Experimental Setup and Analyses
Slope One Recommender on Hadoop YONG ZHENG Center for Web Intelligence DePaul University Nov 15, 2012 Overview Introduction Recommender Systems & Slope One Recommender Distributed Slope One on Mahout and
International Journal of Innovative Research in Computer and Communication Engineering
Achieve Ranking Accuracy Using Cloudrank Framework for Cloud Services R.Yuvarani 1, M.Sivalakshmi 2 M.E, Department of CSE, Syed Ammal Engineering College, Ramanathapuram, India ABSTRACT: Building high
Frequency Matters. The keys to optimizing email send frequency
The keys to optimizing email send frequency Email send frequency requires a delicate balance. Send too little and you miss out on sales opportunities and end up leaving money on the table. Send too much
Sales Analysis of E-Commerce Websites using Data Mining Techniques
Sales Analysis of E-Commerce Websites using Data Mining Techniques Anurag Bejju Department of Computer Science Birla Institute of Technology & Science, Pilani, Dubai Campus, P.O.Box 345055, Dubai International
Scalable Collaborative Filtering with Jointly Derived Neighborhood Interpolation Weights
Seventh IEEE International Conference on Data Mining Scalable Collaborative Filtering with Jointly Derived Neighborhood Interpolation Weights Robert M. Bell and Yehuda Koren AT&T Labs Research 180 Park
Comparing IPL2 and Yahoo! Answers: A Case Study of Digital Reference and Community Based Question Answering
Comparing and : A Case Study of Digital Reference and Community Based Answering Dan Wu 1 and Daqing He 1 School of Information Management, Wuhan University School of Information Sciences, University of
II. RELATED WORK. Sentiment Mining
Sentiment Mining Using Ensemble Classification Models Matthew Whitehead and Larry Yaeger Indiana University School of Informatics 901 E. 10th St. Bloomington, IN 47408 {mewhiteh, larryy}@indiana.edu Abstract
Crowdfunding Support Tools: Predicting Success & Failure
Crowdfunding Support Tools: Predicting Success & Failure Michael D. Greenberg Bryan Pardo [email protected] [email protected] Karthic Hariharan [email protected] tern.edu Elizabeth
Personalization using Hybrid Data Mining Approaches in E-business Applications
Personalization using Hybrid Data Mining Approaches in E-business Applications Olena Parkhomenko, Chintan Patel, Yugyung Lee School of Computing and Engineering University of Missouri Kansas City {ophwf,
Big & Personal: data and models behind Netflix recommendations
Big & Personal: data and models behind Netflix recommendations Xavier Amatriain Netflix [email protected] ABSTRACT Since the Netflix $1 million Prize, announced in 2006, our company has been known to
Efficient Algorithm for Predicting QOS in Cloud Services
Efficient Algorithm for Predicting QOS in Cloud Services Sangeeta R. Alagi 1, Srinu Dharavath 2 1 ME Computer Engineering, G. S. MOZE College of Engineering Balewadi, Pune, Maharashtra India 2 Assistant
A New Recommendation Method for the User Clustering-Based Recommendation System
ISSN 1392 124X (print), ISSN 2335 884X (online) INFORMATION TECHNOLOGY AND CONTROL, 2015, T. 44, Nr. 1 A New Recommendation Method for the User Clustering-Based Recommendation System Aurimas Rapečka, Gintautas
Recommender Systems. Jan Boehmer, Yumi Jung, and Rick Wash. Michigan State University. Email: [email protected]. Word count: 4,744
Recommender Systems Jan Boehmer, Yumi Jung, and Rick Wash Michigan State University Email: [email protected] Word count: 4,744 Abstract Recommender systems are a vital part of today s information society
Importance of Online Product Reviews from a Consumer s Perspective
Advances in Economics and Business 1(1): 1-5, 2013 DOI: 10.13189/aeb.2013.010101 http://www.hrpub.org Importance of Online Product Reviews from a Consumer s Perspective Georg Lackermair 1,2, Daniel Kailer
Introducing diversity among the models of multi-label classification ensemble
Introducing diversity among the models of multi-label classification ensemble Lena Chekina, Lior Rokach and Bracha Shapira Ben-Gurion University of the Negev Dept. of Information Systems Engineering and
VACA: A Tool for Qualitative Video Analysis
VACA: A Tool for Qualitative Video Analysis Brandon Burr Stanford University 353 Serra Mall, Room 160 Stanford, CA 94305 USA [email protected] Abstract In experimental research the job of analyzing data
Pairwise Preference Regression for Cold-start Recommendation
Pairwise Preference Regression for Cold-start Recommendation ABSTRACT Seung-Taek Park Samsung Advanced Institute of Technology Mt. 14-1, Nongseo-dong, Giheung-gu Yongin-si, Gyunggi-do 446-712, South Korea
Quantitative Analysis of Desirability in User Experience
Quantitative Analysis of Desirability in User Experience Sisira Adikari School of Information Systems and Accounting University of Canberra Canberra, Australia ACT 2617 Email: [email protected]
! E6893 Big Data Analytics Lecture 5:! Big Data Analytics Algorithms -- II
! E6893 Big Data Analytics Lecture 5:! Big Data Analytics Algorithms -- II Ching-Yung Lin, Ph.D. Adjunct Professor, Dept. of Electrical Engineering and Computer Science Mgr., Dept. of Network Science and
Employee Surveys: Four Do s and Don ts. Alec Levenson
Employee Surveys: Four Do s and Don ts Alec Levenson Center for Effective Organizations University of Southern California 3415 S. Figueroa Street, DCC 200 Los Angeles, CA 90089 USA Phone: 1-213-740-9814
MHI3000 Big Data Analytics for Health Care Final Project Report
MHI3000 Big Data Analytics for Health Care Final Project Report Zhongtian Fred Qiu (1002274530) http://gallery.azureml.net/details/81ddb2ab137046d4925584b5095ec7aa 1. Data pre-processing The data given
Principles of Dat Da a t Mining Pham Tho Hoan [email protected] [email protected]. n
Principles of Data Mining Pham Tho Hoan [email protected] References [1] David Hand, Heikki Mannila and Padhraic Smyth, Principles of Data Mining, MIT press, 2002 [2] Jiawei Han and Micheline Kamber,
The Optimality of Naive Bayes
The Optimality of Naive Bayes Harry Zhang Faculty of Computer Science University of New Brunswick Fredericton, New Brunswick, Canada email: hzhang@unbca E3B 5A3 Abstract Naive Bayes is one of the most
MIRACLE at VideoCLEF 2008: Classification of Multilingual Speech Transcripts
MIRACLE at VideoCLEF 2008: Classification of Multilingual Speech Transcripts Julio Villena-Román 1,3, Sara Lana-Serrano 2,3 1 Universidad Carlos III de Madrid 2 Universidad Politécnica de Madrid 3 DAEDALUS
Towards Inferring Web Page Relevance An Eye-Tracking Study
Towards Inferring Web Page Relevance An Eye-Tracking Study 1, [email protected] Yinglong Zhang 1, [email protected] 1 The University of Texas at Austin Abstract We present initial results from a project,
Leveraging the engagement of games to change energy behavior
Leveraging the engagement of games to change energy behavior Byron Reeves, Ph.D. Stanford University Department of Communication Stanford, CA 94305 USA [email protected] James J. Cummings Stanford University
Prediction of Stock Performance Using Analytical Techniques
136 JOURNAL OF EMERGING TECHNOLOGIES IN WEB INTELLIGENCE, VOL. 5, NO. 2, MAY 2013 Prediction of Stock Performance Using Analytical Techniques Carol Hargreaves Institute of Systems Science National University
Effects of Online Recommendations on Consumers Willingness to Pay
Effects of Online Recommendations on Consumers Willingness to Pay Gediminas Adomavicius University of Minnesota Minneapolis, MN [email protected] Jesse Bockstedt University of Arizona Tucson, AZ [email protected]
Evaluating and Predicting Answer Quality in Community QA
Evaluating and Predicting Answer Quality in Community QA Chirag Shah Jefferey Pomerantz School of Communication & Information (SC&I) School of Information & Library Science (SILS) Rutgers, The State University
Understanding the popularity of reporters and assignees in the Github
Understanding the popularity of reporters and assignees in the Github Joicy Xavier, Autran Macedo, Marcelo de A. Maia Computer Science Department Federal University of Uberlândia Uberlândia, Minas Gerais,
Reading with Mobile Phone & Large Display
Reading with Mobile Phone & Large Display Roswitha Gostner Computing Department Lancaster University, [email protected] Hans Gellersen Computing Department Lancaster University Lancaster, LA1 4WY, UK
A Workbench for Comparing Collaborative- and Content-Based Algorithms for Recommendations
A Workbench for Comparing Collaborative- and Content-Based Algorithms for Recommendations Master Thesis Pat Kläy from Bösingen University of Fribourg March 2015 Prof. Dr. Andreas Meier, Information Systems,
A Rule-Based Short Query Intent Identification System
A Rule-Based Short Query Intent Identification System Arijit De 1, Sunil Kumar Kopparapu 2 TCS Innovation Labs-Mumbai Tata Consultancy Services Pokhran Road No. 2, Thane West, Maharashtra 461, India 1
