Modeling Learner Engagement in MOOCs using Probabilistic Soft Logic
|
|
|
- Osborn Walters
- 10 years ago
- Views:
Transcription
1 Modeling Learner Engagement in MOOCs using Probabilistic Soft Logic Arti Ramesh, Dan Goldwasser, Bert Huang, Hal Daumé III, Lise Getoor Department of Computer Science, University of Maryland, MD, USA {artir, bert, hal, Abstract Massive open online courses (MOOCs) attract a large number of student registrations, but recent studies have shown that only a small fraction of these students complete their courses. Student dropouts are thus a major deterrent for the growth and success of MOOCs. We believe that understanding student engagement as a course progresses is essential for minimizing dropout rates. Formally defining student engagement in an online setting is challenging. In this paper, we leverage activity (such as posting in discussion forums, timely submission of assignments, etc.), linguistic features from forum content and structural features from forum interaction to identify two different forms of student engagement (passive and active) in MOOCs. We use probabilistic soft logic (PSL) to model student engagement by capturing domain knowledge about student interactions and performance. We test our models on MOOC data from Coursera and demonstrate that modeling engagement is helpful in predicting student performance. 1 Introduction Massive open online courses (MOOCs) often attract up to hundreds of thousands of registrants, but only a small fraction of these successfully complete their courses. Even of those students who declare at the start of a course an intent to complete, 75% do not (according to a recent Coursera study [1]). Maintaining and cultivating student engagement is a prerequisite for MOOCs to have broad educational impact. Unlike regular courses in which students engage with class materials in a structured and monitored way, and instructors directly observe student behavior and obtain feedback, the distant nature and the sheer size of an online course require new approaches for providing student feedback and guiding instructor intervention. MOOCs provide a tantalizing opportunity for analyzing large-scale online interaction and behavioral data to improve student engagement, outcomes, and overall experience. To date, this opportunity is purely speculative: little work has truly exploited content (language), structure (social interactions), and outcome data. One significant technical challenge is that to do so requires the ability to combine language analysis of forum posts with graph analysis over very large networks of entities (students, instructors, topics, assignments, quizzes, etc.) to perform predictive modeling. We follow the observation that quantifying and measuring engagement is key to understanding learner participation in the course. In MOOCs particularly, there are different notions of student engagement. Learners often engage in different aspects of the course throughout its duration. For example, some students engage in the social aspects of the online community by posting in forums, asking and answering questions; while others only watch lectures and take quizzes without interacting with the community. 1
2 Unlike classroom courses where engagement can be observed in person, it is challenging to recognize and measure engagement in an online environment. Some indications of learner engagement include online activity by the learner on the course website, interactions with other learners/staff on the discussion forums, completion of quizzes/assignments and language used by the learner in posts. These differences make the problem of measuring learner engagement difficult. In this work, we use probabilistic soft logic (PSL), a recently introduced formalism for rich, graphical models over structured domains, which supports efficient inference and parameter estimation [1]. PSL provides an easy means to represent and combine the behavioral, linguistic and structural features in a concise manner. We propose a PSL model that uses learner behavioral features to distinguish between forms of engagement learners display in the course passive or active and use the above model to predict whether learners earned a statement of accomplishment from the course (learner performance). Our model reasons about types of learner engagement by formulating it as a latent variable that underlies learners behavior. Our empirical evaluation on a Coursera course dataset indicates that modeling engagement improves learner performance prediction. We also explore potential ways in which engagement predictions can be used to reason about other aspects of online course participation by observing the forum content posted by engaged and disengaged learners. 2 Related Work Student engagement is known to be a significant factor in success of student learning [2], but there is still limited work studying learner engagement in MOOCs. Our work is closest to that of Kizilcec et al., who attempt to understand learner disengagement in MOOCs [4]. In their work (based on three online MOOC courses), the authors identify four prototypical trajectories of engagement and describe the set of features for comparing the different trajectories. Their work clusters user trajectories and identifies them as engaged/disengaged based on cluster membership. Our work differs from the above work in that we view types of engagement as a latent variables and learn to differentiate among the engagement types from data. We use learner performance scores to train the model and use this model to predict whether the learner earned a statement of accomplishment in the online course. We model engagement more explicitly and show that it helps in predicting learner performance. Prior work [2, 3] has studied the relationship between student engagement and academic performance for traditional classroom courses; they identify several metrics for user engagement (such as student-faculty interaction, level of academic challenge). Carini et al. demonstrate quantitatively that though most engagement metrics are positively correlated to performance, the relationships in many cases can be weak [3]. Our work borrows ideas from Kuh et al. [2] and Carini et al. [3] and adapts these to an online setting. MOOCs are fundamentally different from traditional courses in, e.g., the number of students enrolled, student-faculty interactions, methods of assessment, and lack of personal interaction. In this work, we identify metrics for learner engagement tailored for online courses and analyze how these engagement metrics relate to performance. 3 Problem Statement Decreasing the student dropout rate is arguably the biggest challenge faced by online education providers conducting online courses. Recent studies [2] show that only about 5 percent of students who register for the course eventually end up completing it and obtaining a grade in it. This proportion is also consistent with the data we used (collected from the Surviving Disruptive Technologies course, described in Section 6), in which only 5.11% of the students completed the course. In this work we take a step towards helping educators in this goal using data driven methods. We analyze the learners online behavior in an attempt to identify how they engage with the course materials, and how this engagement can help predict whether they receive a statement of accomplishment in the course. We follow the intuition that the level and type of student engagement within the course provides good indication, and we distinguish between different types of engagement and how they relate to different activity patterns. For example, we find that learners who do not take quizzes during lectures still follow forum discussion, indicating a passive form of engagement. For these users engagement is typically manifested as viewing and voting forum activities. In this paper, we 2
3 formalize this intuition, and construct a probabilistic model capturing the latent engagement level of students. 4 Learner Engagement in MOOCs Studies of student engagement in classroom settings are based upon factors such as attendance, participation in discussions and grades. Many of these factors are not directly observable in an online course, posing difficulties in assessing learner engagement. So it is essential to interpret online behavioral patterns for inferring engagement in MOOCs. Since student responses are not directly observed in an online course, engagement cannot be easily determined, but rather it can be inferred by interpreting behavioral patterns which indicate student s level and type of involvement. In order to model learner engagement we first identify the relevant online behavioral activities of learners 1) Posting on discussion forums, 2) Subscribing/viewing/voting on content posted by others, 3) Following course material, and 4) Completing assessments associated with the course. Learners posting actively in the discussion forums can act as a good indicator of learner engagement, since forums are the primarily observable means of interaction in a MOOC setting. Learners can use the forum posts to convey the satisfaction or dissatisfaction with course and possible provide indications of their interest in the class and motivation to complete it. A finer analysis of the content of the posts is essential to conclude whether the learner is engaged. In order to use this information, in addition to the behavioral indicators above, we include linguistic indicators describing the posts content that could provide indication of the learner s engagement/disengagement. We use an automated tool (OpinionFinder [6]) to annotate the form posts with two types of labels: subjectivity of content in posts and sentiment polarity of post content. Figure 1 demonstrates the importance of capturing this information. The figure depicts the variation of subjective expressions in forum content with the reaction these posts attract (measured in votes). The increasing trend of votes with increase in subjective expressions is indicative of the fact that posts containing subjective expressions invite more attention and trigger engagement. Based on these two types of signals, we categorize learner engagement into the following types: Active Engagement Assigned to learners who show explicit signs of engagement by posting on the discussion forums, submitting quizzes and assessments. These signs require an active involvement from the learner. Passive Engagement Assigned to learners who show more implicit signs of engagement by viewing, subscribing or voting on posts/comments on discussion forums and views lectures. These users typically do not make an active effort to participate. Disengaged Learners that show signs of getting disengaged from the course, either by posting text that indicate their disengagement or show a significant decrease in posting, viewing, voting, and assessment submitting activity. 2.5 Subjective Expressions Votes Figure 1: Subjective expressions vs. votes in Disruptive Technologies course We are interested in modeling engagement and reasoning about which form(s) of engagement are strong indicators of performance. 3
4 5 Modeling Learner Engagement in PSL In this section, we provide a brief description of PSL. Additional details about PSL are available in [1]. We also discuss the various features (behavioral, linguistic and structural), which learners exhibit online while attending these courses. We then describe the PSL models to model student behavior which capture these features. We begin with a simple model which predicts performance using only behavioral, linguistic, structural and temporal features and then increase its complexity by including forms of engagement and disengagement as a latent feature. 5.1 Probabilistic Soft Logic Probabilistic soft logic (PSL) [1] is a framework for collective, probabilistic reasoning in relational domains. PSL uses syntax based on first-order logic as a templating language for continuous graphical models over random variables representing soft truth values. Like other statistical relational learning methods [5], PSL uses weighted rules to model the dependencies in a domain. However, one distinguishing aspect is that PSL uses continuous variables to represent truth values, relaxing Boolean truth values to the interval [0,1]. Triangular norms, which are continuous relaxations of logical connectives AND and OR, are used to combine the atoms in the first-order clauses. As a result of the soft formulation and the triangular norms, the underlying probabilistic model is a hinge-loss Markov random field (HL-MRF) [7]. Inference in HL-MRFs is a convex optimization problem, which makes working with PSL very efficient in comparison to relational modeling tools that use discrete representations. HL-MRFs admit various learning algorithms for fully-supervised training data, and are amenable to point-estimate hard expectation maximization for partially-supervised data with latent variables [9]. In our model, we exploit this to represent student engagement as a latent variable. 5.2 Features Our PSL model is built on a foundation of observable features from the data. In the following, we detail the various features we collect from the data. Following the first-order logic based syntax of PSL, we describe these features as logical predicates. Behavioral Features Our behavioral features are attributes that the learner exhibits while on the MOOC website. In our models, we consider two types of behavioral features: aggregate and nonaggregate. The aggregate features give the overall behavioral metrics, while the non-aggregate features are at the instance level. The predicates postactivity(user) and voteactivity(user) capture how active the user is in the forums. These are calculated for each user by assessing whether the user posts more than the average number of posts generated by all users. Predicate reputation(user) represents whether the overall reputation of a user is above average. The aggregate behavioral predicates take Boolean values. Predicate posts(user, POST) and votes(user, POST) capture an instance-level log of users posting and voting on the discussion forums. The predicates posts and votes take value 1 if the USER posts or votes on POST. Predicate upvote(post) is true if the post has positive votes and false otherwise, and predicate downvote(post) is true if a post has been downvoted. Linguistic Features The subjectivity and polarity scores of the posts are from OpinionFinder [6], represented by subjective(post) and polarity(post) respectively. Both these predicates are calculated by normalizing the number of subjective/objective tags and positive and negative polarity tags marked by opinion finder. These predicates take values [0,1]. Temporal Features The course is divided into the following time periods: one time period before the course starts and three time periods during the course. The time period splits during the course are constructed according to the deadlines in the course one before the first assignment, second and third being before and after the midterm respectively. The temporal features lastquiz, lastlecture, lastpost, lastview and lastvote capture the time-period in which each last interaction of the user occurred. These features measure to what lengths the user participated in different aspects of the course. 4
5 postactivity(u) reputation(u) performance(u) voteactivity(u) reputation[(u) performance(u) posts(u, P ) positive(p ) performance(u) votes(u, P ) positive(p ) performance(u) posts(u, P ) upvote(p ) performance(u) Table 1: Rules for simple PSL model. postactivity(u) reputation(u) eactive(u) voteactivity(u) reputation(u) epassive(u) posts(u, P ) positive(p ) eactive(u) votes(u, P ) positive(p ) epassive(u) posts(u, P ) upvote(p ) eactive(u) posts(u 1, P 1 ) posts(u 2, P 2 ) eactive(u 1 ) samethread(p 1, P 2 ) eactive(u 2 ) eactive(u) epassive(p ) performance(u) Table 2: Rules for PSL model with latent variables, latent variables highlighted in bold. Structural Features We also include structural features induced by forum structure. These are given by samethread(post 1, POST 2) and sameforum(thread 1, THREAD 2) to capture posts in the same thread and threads in the same forum respectively. Including these relationships allows for our model to capture structural phenomena in interactions and how it can affect student engagement and performance. 5.3 PSL Models We now construct two different PSL models for predicting learner performance in a MOOC setting 1) a simple model that directly infers learner performance from observable features and 2) a latent variable model that infers student engagement as a hidden variable to predict learner performance. In our simple PSL model, we model performance of learners by using the observable behavioral features exhibited by the learner, linguistic features corresponding to the content of the post and structural features derived from forum interaction. Meaningful combinations of one or more observable behavioral features described in Section 5.2 are used to predict performance. Table 1 gives a subset of rules used in this model (U and P in tables 1 and 2 refer to USER and POST respectively). As can be seen, the observable features directly imply performance of the learner in the simple model. We can enhance this type of model by adding latent variables that have semantic meaning but can not be directly measured from the data. We treat learner engagement as a latent variable and associate the various observed features to one or more forms of engagement and use the engagement variables to predict learner performance. The latent variables in this model are represented by predicates engagement active, engagement passive and disengagement. In this model, some of the observable features postactivity, voteactivity, viewactivity are used to classify learners into one or more forms of engagement or disengagement. Then the engagement predicates engagement active, engagement passive and disengagement conjuncted with features that were not used in classifying user engagement like reputation are then used to predict learner performance. For example, in Table 2, conjunction of postactivity and reputation implies engagement active; conjunction of voteactivity and reputation implies engagement passive; while engagement active and engagement passive implies performance. Note that engagement active and engagement passive are represented by eactive and epassive in the table. The weights of the model are learned by performing hard expectationmaximization with performance as a target variable. Learning a model with latent engagement gives information about which forms of engagement are good indicators of learner performance. The resulting model results in better predictive performance and can provide more insight into MOOC user behavior than a simpler model. 6 Experimental Results We evaluate our models using data from a Coursera course, Surviving Disruptive Technologies, taught by Professor Hank Lucas. The seven-week course had 1665 users participating in the forums and 826 users completing the course with a nonzero score. We conduct experiments that help us answer the following questions about learner engagement: (1) How effective are our models in predicting performance? (2) What are the key factors influencing 5
6 learner performance in an online setting? To address the first question, we evaluate the performance of our models in predicting learner performance. We compute the area under ROC curve (AUC) for our evaluations. We use 10-fold crossvalidation in our experiments, leaving out 10% of the data for testing and rest for training phase, where the model weights are learned. We observe that the PSL model with the latent engagement variables performs better at predicting learner performance and also provides more understanding of reasons why the user did not perform. Table 3 shows the AUC values for the PSL models discussed in Section 5.3. AUC-PR Pos. AUC-PR Neg. AUC-ROC Kendall Simple PSL model PSL model with latent variables Table 3: Performance of simple PSL model and PSL model with latent variables. To address the second question, we examined weights our model learns for the different engagement variables in predicting performance. We see that the rules involving predicates postactivity, voteactivity, viewactivity and reputation gain a high weight in the simple learned model. In our second model, we observe that rules involving engagement active predicts performance, while rules containing disengagement gain a higher weight in predicting performance. In addition to predicting performance, the engagement variables are helpful in interpreting the different facets of learner participation in the course. Of particular interest is the content of the posts made by learners and how it corresponds to the values predicted by the our model. Table 4 shows some examples of posts made by users and the engagement and performance scores predicted by the model. It is interesting to note that engaged learners post content with negative sentiment on the course, which may be perceived as participating in the course, for example, the first user in the Table 4. While disengaged learners may post similar content with negative sentiment, a change in activity will help discern them from the engaged ones, e.g., the third user in Table 4. Engaged learner (positive sentiment) Engaged learner (negative sentiment) Disengaged Learner (negative sentiment) Disengaged Learner (negative sentiment) performance = performance = performance = 0.5 disengagement = performance = Auditor performance = 0.0 Prof. Lucas, Thank you for a great course! And thank you Coursera! I have also received a 9, the most disappointing thing is that I have only received good or passing comments from my peers, 3 of 5 did not post any comment about my work. I agree completely. I used a lot of time on my assignment and got 7.5. think the evaluation criteria were wrong, it shouldn t be rated on whether you have 3 or 4 innovations in your description but on a subjective measure (which is also flawed). Generally you should pass if it s obvious that you have followed the course and that you have tried to use the theories that the course has(to a certain degree ofc.). The grades I received are ridiculous! One pointed out that the good point in my assignment was my fluency in English!!! Oh God!Another one told me that I used general knowledge for asking the questions. Yes, I ve tried to use general knowledge for applying the theory so that it is useful (and not just a theory!).i ve re-read my assignment and I still can t believe in my grade. Is it really fair?. This has been an otherwise fantastic course. Too bad the potential for success is so heavily weighted on two assignments. Table 4: Relevant forum content posted by users assigned different engagement labels by our model. 7 Discussion In this work, we take a first step towards understanding student engagement and its impact on student performance using data-driven methods. We formalize, using PSL, our intuition that student engagement can be modeled as a complex interaction of behavioral, linguistic and social cues, and we show that our model can construct an interpretation for the latent engagement types from data, without supervision, based on their impact on final performance. In addition to our empirical results, we provide some qualitative results of possible uses of the learned engagement-classes. While not comprehensive, it demonstrates the direction of our future work facilitating instructors intervention when needed, based on identifying user groups and the sentiment emerging from their forum content. 6
7 Acknowledgements This work is supported by National Science Foundation under Grant No. CCF Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation. References [1] Broecheler, M., Mihalkova, L. & Getoor, L., Probabilistic Soft Logic. In Conference on Uncertainty in Artificial Intelligence, [2] Kuh, G. D., What we re learning about student engagement from NSSE: Benchmarks for effective educational practices. Change: The Magazine of Higher Learning, [3] Carini, R. M., Kuh, G. D. & Klein, S. P., Student engagement and student learning: Testing the linkages. Research in Higher Education, 47(1), 1-32, [4] Kizilcec, R. F., Piech, C. & Schneider, E., Deconstructing disengagement: analyzing learner subpopulations in massive open online courses. In International Conference on Learning Analytics and Knowledge, [5] Getoor, L. & Tasker, B., Introduction to Statistical Relational Learning. The MIT Press, [6] Wilson, T., Hoffmann, P., Somasundaran, S., Kessler, J., Wiebe, J., Choi, Y., Cardie, C., Riloff, E., & Patwardhan, S., OpinionFinder: A System for Subjectivity Analysis. In Proceedings of Human Language Technologies Conference/Conference on Empirical Methods in Natural Language Processing (HLT/EMNLP 2005), Companion Volume (software demonstration), [7] Bach, S. H., Huang, B., London, B. & Getoor, L., Hinge-loss Markov Random Fields: Convex Inference for Structured Prediction. In Uncertainty in Artificial Intelligence, [8] Bach, S. H., Broecheler, M., Getoor, L. & O Leary, D. P., Scaling MPE Inference for Constrained Continuous Markov Random Fields. In Advances in Neural Information Processing Systems (NIPS) pages , [9] Bach, S. H., Huang, B. & Getoor, L, Learning Latent Groups with Hinge-loss Markov Random Fields. In Inferning: ICML Workshop on Interactions between Inference and Learning,
Learning Latent Engagement Patterns of Students in Online Courses
Learning Latent Engagement Patterns of Students in Online Courses 1 Arti Ramesh, 1 Dan Goldwasser, 1 Bert Huang, 1 Hal Daumé III, 2 Lise Getoor 1 University of Maryland, College Park 2 University of California,
Big Data Science. Prof. Lise Getoor University of Maryland, College Park. http://www.cs.umd.edu/~getoor. October 17, 2013
Big Data Science Prof Lise Getoor University of Maryland, College Park October 17, 2013 http://wwwcsumdedu/~getoor BIG Data is not flat 2004-2013 lonnitaylor Data is multi-modal, multi-relational, spatio-temporal,
arxiv:1407.5238v1 [cs.cy] 20 Jul 2014
Towards Feature Engineering at Scale for Data from Massive Open Online Courses arxiv:1407.5238v1 [cs.cy] 20 Jul 2014 Kalyan Veeramachaneni Massachusetts Institute of Technology, Cambridge, MA 02139 USA
MOOCdb: Developing Data Standards for MOOC Data Science
MOOCdb: Developing Data Standards for MOOC Data Science Kalyan Veeramachaneni, Franck Dernoncourt, Colin Taylor, Zachary Pardos, and Una-May O Reilly Massachusetts Institute of Technology, USA. {kalyan,francky,colin
Tensor Methods for Machine Learning, Computer Vision, and Computer Graphics
Tensor Methods for Machine Learning, Computer Vision, and Computer Graphics Part I: Factorizations and Statistical Modeling/Inference Amnon Shashua School of Computer Science & Eng. The Hebrew University
Learning is a very general term denoting the way in which agents:
What is learning? Learning is a very general term denoting the way in which agents: Acquire and organize knowledge (by building, modifying and organizing internal representations of some external reality);
Data Mining Applications in Higher Education
Executive report Data Mining Applications in Higher Education Jing Luan, PhD Chief Planning and Research Officer, Cabrillo College Founder, Knowledge Discovery Laboratories Table of contents Introduction..............................................................2
CHAPTER 1 INTRODUCTION
1 CHAPTER 1 INTRODUCTION Exploration is a process of discovery. In the database exploration process, an analyst executes a sequence of transformations over a collection of data structures to discover useful
Basic Data Analysis. Stephen Turnbull Business Administration and Public Policy Lecture 12: June 22, 2012. Abstract. Review session.
June 23, 2012 1 review session Basic Data Analysis Stephen Turnbull Business Administration and Public Policy Lecture 12: June 22, 2012 Review session. Abstract Quantitative methods in business Accounting
René F. Kizilcec. Curriculum Vitae
Department of Communication Stanford University, Stanford, CA 94305 [email protected] http://rene.kizilcec.com/ 650-798-4606 Education Ph.D., Communication, Stanford University, Stanford, CA, Expected
Instructional Design Principles in the Development of LearnFlex
Instructional Design Principles in the Development of LearnFlex A White Paper by Dr. Gary Woodill, Ed.D. Chief Learning Officer, Operitel Corporation [email protected] Dr. Karen Anderson, PhD. Senior
Cell Phone based Activity Detection using Markov Logic Network
Cell Phone based Activity Detection using Markov Logic Network Somdeb Sarkhel [email protected] 1 Introduction Mobile devices are becoming increasingly sophisticated and the latest generation of smart
Measurement and Metrics Fundamentals. SE 350 Software Process & Product Quality
Measurement and Metrics Fundamentals Lecture Objectives Provide some basic concepts of metrics Quality attribute metrics and measurements Reliability, validity, error Correlation and causation Discuss
WWC Single Study Review A review of the design and summary of findings for an individual study
What Works Clearinghouse WWC Single Study Review A review of the design and summary of findings for an individual study U.S. DEPARTMENT OF EDUCATION July 2015 WWC Review of the Report Interactive Online
Healthcare data analytics. Da-Wei Wang Institute of Information Science [email protected]
Healthcare data analytics Da-Wei Wang Institute of Information Science [email protected] Outline Data Science Enabling technologies Grand goals Issues Google flu trend Privacy Conclusion Analytics
Identifying At-Risk Students Using Machine Learning Techniques: A Case Study with IS 100
Identifying At-Risk Students Using Machine Learning Techniques: A Case Study with IS 100 Erkan Er Abstract In this paper, a model for predicting students performance levels is proposed which employs three
Introduction to Longitudinal Data Analysis
Introduction to Longitudinal Data Analysis Longitudinal Data Analysis Workshop Section 1 University of Georgia: Institute for Interdisciplinary Research in Education and Human Development Section 1: Introduction
Modeling Problem Solving in Massive Open Online Courses. Fang Han
Modeling Problem Solving in Massive Open Online Courses by Fang Han Submitted to the Department of Electrical Engineering and Computer Science in partial fulfillment of the requirements for the degree
Forum Posting Habits and Attainment in a Dual-Mode MOOC
International Journal for Cross-Disciplinary Subjects in Education (IJCDSE), Special Issue Volume 5 Issue, 05 Forum Posting Habits and Attainment in a Dual-Mode MOOC D. F. O. Onah, J. E. Sinclair, R. Boyatt
Preliminary Syllabus for the course of Data Science for Business Analytics
Preliminary Syllabus for the course of Data Science for Business Analytics Miguel Godinho de Matos 1,2 and Pedro Ferreira 2,3 1 Cato_lica-Lisbon, School of Business and Economics 2 Heinz College, Carnegie
Cluster Analysis for Evaluating Trading Strategies 1
CONTRIBUTORS Jeff Bacidore Managing Director, Head of Algorithmic Trading, ITG, Inc. [email protected] +1.212.588.4327 Kathryn Berkow Quantitative Analyst, Algorithmic Trading, ITG, Inc. [email protected]
LAGUARDIA COMMUNITY COLLEGE CITY UNIVERSITY OF NEW YORK DEPARTMENT OF MATHEMATICS, ENGINEERING, AND COMPUTER SCIENCE
LAGUARDIA COMMUNITY COLLEGE CITY UNIVERSITY OF NEW YORK DEPARTMENT OF MATHEMATICS, ENGINEERING, AND COMPUTER SCIENCE MAT 119 STATISTICS AND ELEMENTARY ALGEBRA 5 Lecture Hours, 2 Lab Hours, 3 Credits Pre-
Social Media Mining. Data Mining Essentials
Introduction Data production rate has been increased dramatically (Big Data) and we are able store much more data than before E.g., purchase data, social media data, mobile phone data Businesses and customers
INTRUSION PREVENTION AND EXPERT SYSTEMS
INTRUSION PREVENTION AND EXPERT SYSTEMS By Avi Chesla [email protected] Introduction Over the past few years, the market has developed new expectations from the security industry, especially from the intrusion
RELATIONSHIPS BETWEEN SELF DIRECTION, EFFORT AND PERFORMANCE IN AN ONLINE INFORMATION TECHNOLOGY COURSE
RELATIONSHIPS BETWEEN SELF DIRECTION, EFFORT AND PERFORMANCE IN AN ONLINE INFORMATION TECHNOLOGY COURSE William D. Armitage, Naomi Boyer, Sarah Langevin, Alessio Gaspar University of South Florida Polytechnic
Software User Experience and Likelihood to Recommend: Linking UX and NPS
Software User Experience and Likelihood to Recommend: Linking UX and NPS Erin Bradner User Research Manager Autodesk Inc. One Market St San Francisco, CA USA [email protected] Jeff Sauro Founder
Introduction to Machine Learning and Data Mining. Prof. Dr. Igor Trajkovski [email protected]
Introduction to Machine Learning and Data Mining Prof. Dr. Igor Trakovski [email protected] Neural Networks 2 Neural Networks Analogy to biological neural systems, the most robust learning systems
Welcome to the Era of Big Data and Predictive Analytics in Higher Education
Welcome to the Era of Big Data and Predictive Analytics in Higher Education Ellen Wagner WICHE Cooperative for Educational Technologies Joel Hartman University of Central Florida The Focus of this Session
Learning Analytics: Targeting Instruction, Curricula and Student Support
Learning Analytics: Targeting Instruction, Curricula and Student Support Craig Bach Office of the Provost, Drexel University Philadelphia, PA 19104, USA ABSTRACT For several decades, major industries have
Unsupervised Modeling for Understanding MOOC Discussion Forums: A Learning Analytics Approach
Unsupervised Modeling for Understanding MOOC Discussion Forums: A Learning Analytics Approach ABSTRACT Aysu Ezen-Can Department of Computer Science [email protected] Shaun Kellogg Friday Institute for Educational
THE BACKWARDS CLASSROOM: USING PEER INSTRUCTION TO FLIP THE CLASSROOM
The lecture is a process whereby the lecture notes of the instructor get transferred to the notebooks of the students without passing through the brains of either. -Eric Mazur THE BACKWARDS CLASSROOM:
Criticality of Schedule Constraints Classification and Identification Qui T. Nguyen 1 and David K. H. Chua 2
Criticality of Schedule Constraints Classification and Identification Qui T. Nguyen 1 and David K. H. Chua 2 Abstract In construction scheduling, constraints among activities are vital as they govern the
Bayesian networks - Time-series models - Apache Spark & Scala
Bayesian networks - Time-series models - Apache Spark & Scala Dr John Sandiford, CTO Bayes Server Data Science London Meetup - November 2014 1 Contents Introduction Bayesian networks Latent variables Anomaly
STATISTICA. Clustering Techniques. Case Study: Defining Clusters of Shopping Center Patrons. and
Clustering Techniques and STATISTICA Case Study: Defining Clusters of Shopping Center Patrons STATISTICA Solutions for Business Intelligence, Data Mining, Quality Control, and Web-based Analytics Table
DIGITS CENTER FOR DIGITAL INNOVATION, TECHNOLOGY, AND STRATEGY THOUGHT LEADERSHIP FOR THE DIGITAL AGE
DIGITS CENTER FOR DIGITAL INNOVATION, TECHNOLOGY, AND STRATEGY THOUGHT LEADERSHIP FOR THE DIGITAL AGE INTRODUCTION RESEARCH IN PRACTICE PAPER SERIES, FALL 2011. BUSINESS INTELLIGENCE AND PREDICTIVE ANALYTICS
Automated Text Analytics. Testing Manual Processing against Automated Listening
Automated Text Analytics Testing Manual Processing against Automated Listening Contents Executive Summary... 3 Why Is Manual Analysis Inaccurate and Automated Text Analysis on Target?... 3 Text Analytics
2015 Workshops for Professors
SAS Education Grow with us Offered by the SAS Global Academic Program Supporting teaching, learning and research in higher education 2015 Workshops for Professors 1 Workshops for Professors As the market
CEPA Working Paper No. 15-09
CEPA Working Paper No. 15-09 Persistence Patterns in Massive Open Online Courses (MOOCs) AUTHORS Brent J. Evans Vanderbilt University Rachel B. Baker University of California, Irvine Thomas Dee Stanford
What is Data Science? Data, Databases, and the Extraction of Knowledge Renée T., @becomingdatasci, November 2014
What is Data Science? { Data, Databases, and the Extraction of Knowledge Renée T., @becomingdatasci, November 2014 Let s start with: What is Data? http://upload.wikimedia.org/wikipedia/commons/f/f0/darpa
CEB is the leading member-based advisory company. By combining the best practices of thousands of member companies with the advanced research
CEB is the leading member-based advisory company. By combining the best practices of thousands of member companies with the advanced research methodologies and human capital analytics, CEB equips senior
Weakly Supervised Models of Aspect-Sentiment for Online Course Discussion Forums
Weakly Supervised Models of Aspect-Sentiment for Online Course Discussion Forums Arti Ramesh, 1 Shachi H. Kumar, 2 James Foulds, 2 Lise Getoor 2 1 University of Maryland, College Park 2 University of California,
A fundamental goal of education is to equip people with the knowledge and skills that
AUTHORS Candace Thille, Ed.D. Abstract The article addresses the question of how the assessment process with large scale data derived from online learning environments will be different from the assessment
FE670 Algorithmic Trading Strategies. Stevens Institute of Technology
FE670 Algorithmic Trading Strategies Lecture 1. An Overview of Trading and Markets Steve Yang Stevens Institute of Technology 08/29/2012 Outline 1 Logistics 2 Topics 3 Policies 4 Exams & Grades 5 Mathematical
Teacher Evaluation. Missouri s Educator Evaluation System
Teacher Evaluation Missouri s Educator Evaluation System Teacher Evaluation Protocol Introduction Missouri s Educator Evaluation System was created and refined by hundreds of educators across the state.
SECURITY METRICS: MEASUREMENTS TO SUPPORT THE CONTINUED DEVELOPMENT OF INFORMATION SECURITY TECHNOLOGY
SECURITY METRICS: MEASUREMENTS TO SUPPORT THE CONTINUED DEVELOPMENT OF INFORMATION SECURITY TECHNOLOGY Shirley Radack, Editor Computer Security Division Information Technology Laboratory National Institute
Massive Open Online Courses (MOOCs): An Assessment Based on Personal Experience
Massive Open Online Courses (MOOCs): An Assessment Based on Personal Experience By Martina Mangelsdorf, Managing Director and founder of GAIA Insights Trends come and go in education as in any other domain.
Predicting Student Retention in Massive Open Online Courses using Hidden Markov Models
Predicting Student Retention in Massive Open Online Courses using Hidden Markov Models Girish Balakrishnan Electrical Engineering and Computer Sciences University of California at Berkeley Technical Report
Feature Factory: A Crowd Sourced Approach to Variable Discovery From Linked Data
Feature Factory: A Crowd Sourced Approach to Variable Discovery From Linked Data Kiarash Adl Advisor: Kalyan Veeramachaneni, Any Scale Learning for All Computer Science and Artificial Intelligence Laboratory
AMS 5 Statistics. Instructor: Bruno Mendes [email protected], Office 141 Baskin Engineering. July 11, 2008
AMS 5 Statistics Instructor: Bruno Mendes [email protected], Office 141 Baskin Engineering July 11, 2008 Course contents and objectives Our main goal is to help a student develop a feeling for experimental
Online Course Syllabus Page 1
Page 1 STATS 8: Introduction to Biological Statistics Summer Session Class Meeting Information This course meets online for 5 weeks of instruction during Summer Session. Instructor Information Instructor:
Sentiment Analysis on Big Data
SPAN White Paper!? Sentiment Analysis on Big Data Machine Learning Approach Several sources on the web provide deep insight about people s opinions on the products and services of various companies. Social
Mimicking human fake review detection on Trustpilot
Mimicking human fake review detection on Trustpilot [DTU Compute, special course, 2015] Ulf Aslak Jensen Master student, DTU Copenhagen, Denmark Ole Winther Associate professor, DTU Copenhagen, Denmark
CS 2750 Machine Learning. Lecture 1. Machine Learning. http://www.cs.pitt.edu/~milos/courses/cs2750/ CS 2750 Machine Learning.
Lecture Machine Learning Milos Hauskrecht [email protected] 539 Sennott Square, x5 http://www.cs.pitt.edu/~milos/courses/cs75/ Administration Instructor: Milos Hauskrecht [email protected] 539 Sennott
M.S. in Education Assessment in the Major Report 2010. By Dr. Renee Chandler, Program Director Submitted: October 2011
M.S. in Education Assessment in the Major Report 2010 By Dr. Renee Chandler, Program Director Submitted: October 2011 Table of Contents 1. Outcomes of Previous Assessments... 2 2. Questions To Be Answered
How To Write The English Language Learner Can Do Booklet
WORLD-CLASS INSTRUCTIONAL DESIGN AND ASSESSMENT The English Language Learner CAN DO Booklet Grades 9-12 Includes: Performance Definitions CAN DO Descriptors For use in conjunction with the WIDA English
Virtual Teaching in Higher Education: The New Intellectual Superhighway or Just Another Traffic Jam?
Virtual Teaching in Higher Education: The New Intellectual Superhighway or Just Another Traffic Jam? Jerald G. Schutte California State University, Northridge email - [email protected] Abstract An experimental
Designing e-assessment in Massive Online Courses
20115th International Conference on Distance Learning and Education IPCSIT vol.12 (2011) (2011) IACSIT Press, Singapore Designing e-assessment in Massive Online Courses K. P. Hewagamage 1, Hakim Usoof
Learning and Academic Analytics in the Realize it System
Learning and Academic Analytics in the Realize it System Colm Howlin CCKF Limited Dublin, Ireland [email protected] Danny Lynch CCKF Limited Dublin, Ireland [email protected] Abstract: Analytics
CI6227: Data Mining. Lesson 11b: Ensemble Learning. Data Analytics Department, Institute for Infocomm Research, A*STAR, Singapore.
CI6227: Data Mining Lesson 11b: Ensemble Learning Sinno Jialin PAN Data Analytics Department, Institute for Infocomm Research, A*STAR, Singapore Acknowledgements: slides are adapted from the lecture notes
Anatomy of Cyber Threats, Vulnerabilities, and Attacks
Anatomy of Cyber Threats, Vulnerabilities, and Attacks ACTIONABLE THREAT INTELLIGENCE FROM ONTOLOGY-BASED ANALYTICS 1 Anatomy of Cyber Threats, Vulnerabilities, and Attacks Copyright 2015 Recorded Future,
The Predictive Data Mining Revolution in Scorecards:
January 13, 2013 StatSoft White Paper The Predictive Data Mining Revolution in Scorecards: Accurate Risk Scoring via Ensemble Models Summary Predictive modeling methods, based on machine learning algorithms
Robust Sentiment Detection on Twitter from Biased and Noisy Data
Robust Sentiment Detection on Twitter from Biased and Noisy Data Luciano Barbosa AT&T Labs - Research [email protected] Junlan Feng AT&T Labs - Research [email protected] Abstract In this
Single Level Drill Down Interactive Visualization Technique for Descriptive Data Mining Results
, pp.33-40 http://dx.doi.org/10.14257/ijgdc.2014.7.4.04 Single Level Drill Down Interactive Visualization Technique for Descriptive Data Mining Results Muzammil Khan, Fida Hussain and Imran Khan Department
REFLECTING ON EXPERIENCES OF THE TEACHER INDUCTION SCHEME
REFLECTING ON EXPERIENCES OF THE TEACHER INDUCTION SCHEME September 2005 Myra A Pearson, Depute Registrar (Education) Dr Dean Robson, Professional Officer First Published 2005 The General Teaching Council
How To Find Influence Between Two Concepts In A Network
2014 UKSim-AMSS 16th International Conference on Computer Modelling and Simulation Influence Discovery in Semantic Networks: An Initial Approach Marcello Trovati and Ovidiu Bagdasar School of Computing
A FUZZY LOGIC APPROACH FOR SALES FORECASTING
A FUZZY LOGIC APPROACH FOR SALES FORECASTING ABSTRACT Sales forecasting proved to be very important in marketing where managers need to learn from historical data. Many methods have become available for
Maximize Social Media Effectiveness with Data Science. An Insurance Industry White Paper from Saama Technologies, Inc.
Maximize Social Media Effectiveness with Data Science An Insurance Industry White Paper from Saama Technologies, Inc. February 2014 Table of Contents Executive Summary 1 Social Media for Insurance 2 Effective
Texas Success Initiative (TSI) Assessment. Interpreting Your Score
Texas Success Initiative (TSI) Assessment Interpreting Your Score 1 Congratulations on taking the TSI Assessment! The TSI Assessment measures your strengths and weaknesses in mathematics and statistics,
Using Artificial Intelligence to Manage Big Data for Litigation
FEBRUARY 3 5, 2015 / THE HILTON NEW YORK Using Artificial Intelligence to Manage Big Data for Litigation Understanding Artificial Intelligence to Make better decisions Improve the process Allay the fear
WHITE PAPER. Social media analytics in the insurance industry
WHITE PAPER Social media analytics in the insurance industry Introduction Insurance is a high involvement product, as it is an expense. Consumers obtain information about insurance from advertisements,
Standards for Quality Online Teaching
Standard 1 Standards for Quality Online Teaching The teacher plans, designs and incorporates strategies to encourage active learning, interaction, participation and collaboration in the online environment.
GRADUATE STUDENT SATISFACTION WITH AN ONLINE DISCRETE MATHEMATICS COURSE *
GRADUATE STUDENT SATISFACTION WITH AN ONLINE DISCRETE MATHEMATICS COURSE * Amber Settle DePaul University 243 S. Wabash Avenue Chicago, IL 60604 (312) 362-5324 [email protected] Chad Settle University
Data Mining - Evaluation of Classifiers
Data Mining - Evaluation of Classifiers Lecturer: JERZY STEFANOWSKI Institute of Computing Sciences Poznan University of Technology Poznan, Poland Lecture 4 SE Master Course 2008/2009 revised for 2010
UNIVERSITY OF SOUTHERN CALIFORNIA Marshall School of Business BUAD 425 Data Analysis for Decision Making (Fall 2013) Syllabus
UNIVERSITY OF SOUTHERN CALIFORNIA Marshall School of Business BUAD 425 Data Analysis for Decision Making (Fall 2013) Contact Information Syllabus Professor: Dr. Abbass Sharif Office: BRI 400-E Office Hours:
Cognitive and Organizational Challenges of Big Data in Cyber Defense
Cognitive and Organizational Challenges of Big Data in Cyber Defense Nathan Bos & John Gersh Johns Hopkins University Applied Laboratory [email protected], [email protected] The cognitive and organizational
Teaching Portfolio. Teaching Philosophy
Teaching Portfolio Teaching Philosophy Over the course of my education, I have had the privilege of interacting with some truly excellent teachers, who have shaped my knowledge, reasoning, and technical
Course Outline Department of Computing Science Faculty of Science. COMP 3710-3 Applied Artificial Intelligence (3,1,0) Fall 2015
Course Outline Department of Computing Science Faculty of Science COMP 710 - Applied Artificial Intelligence (,1,0) Fall 2015 Instructor: Office: Phone/Voice Mail: E-Mail: Course Description : Students
Pentaho Data Mining Last Modified on January 22, 2007
Pentaho Data Mining Copyright 2007 Pentaho Corporation. Redistribution permitted. All trademarks are the property of their respective owners. For the latest information, please visit our web site at www.pentaho.org
COLUMBIA UNIVERSITY IN THE CITY OF NEW YORK DEPARTMENT OF INDUSTRIAL ENGINEERING AND OPERATIONS RESEARCH
Course: IEOR 4575 Business Analytics for Operations Research Lectures MW 2:40-3:55PM Instructor Prof. Guillermo Gallego Office Hours Tuesdays: 3-4pm Office: CEPSR 822 (8 th floor) Textbooks and Learning
Transcript: What Is Progress Monitoring?
Transcript: What Is Progress Monitoring? Slide 1: Welcome to the webinar, What Is Progress Monitoring? This is one of 11 webinars developed by the National Center on Response to Intervention (NCRTI). This
Marketing Funnels integrated into your Customer Journey Maps: A Winning Combination
B2C Marketing Management Marketing Funnels integrated into your Customer Journey Maps: A Winning Combination On most websites the most common path is usually followed by less than five percent of visitors,
A Comparison of System Dynamics (SD) and Discrete Event Simulation (DES) Al Sweetser Overview.
A Comparison of System Dynamics (SD) and Discrete Event Simulation (DES) Al Sweetser Andersen Consultng 1600 K Street, N.W., Washington, DC 20006-2873 (202) 862-8080 (voice), (202) 785-4689 (fax) [email protected]
Data Isn't Everything
June 17, 2015 Innovate Forward Data Isn't Everything The Challenges of Big Data, Advanced Analytics, and Advance Computation Devices for Transportation Agencies. Using Data to Support Mission, Administration,
South Carolina College- and Career-Ready (SCCCR) Probability and Statistics
South Carolina College- and Career-Ready (SCCCR) Probability and Statistics South Carolina College- and Career-Ready Mathematical Process Standards The South Carolina College- and Career-Ready (SCCCR)
How To Perform An Ensemble Analysis
Charu C. Aggarwal IBM T J Watson Research Center Yorktown, NY 10598 Outlier Ensembles Keynote, Outlier Detection and Description Workshop, 2013 Based on the ACM SIGKDD Explorations Position Paper: Outlier
Supervised Learning (Big Data Analytics)
Supervised Learning (Big Data Analytics) Vibhav Gogate Department of Computer Science The University of Texas at Dallas Practical advice Goal of Big Data Analytics Uncover patterns in Data. Can be used
Social Media Analytics
Social Media Analytics Raghu Krishnapuram and Jitendra Ajmera IBM Research - India 2011 IBM Corporation Convergence of Social and Analytic Technologies Transform the Way the World Operates Socially Synergistic
