A STAB at Making Sense of VAST Data
|
|
|
- Jonas Bryan
- 9 years ago
- Views:
Transcription
1 Adams, S. & Goel, A. (2007) A STAB at Making Sense of VAST Data. In Proc. AAAI-2007 Workshop on Plan, Activity, and Intent Recognition, Vancouver, Canada, July 23, 2007, pp A STAB at Making Sense of VAST Data Summer Adams & Ashok K. Goel Artificial Intelligence Laboratory & Southeastern Regional Visual Analytics Center School of Interactive Computing, Georgia Institute of Technology Atlanta, GA 30332, USA {[email protected], [email protected]} Abstract We view sensemaking in threat analysis as abducing stories that explain the current data and make verifiable predictions about future data. We have developed a preliminary system, called STAB, that abduces multiple stories from the VAST dataset. STAB uses the TMKL knowledge representation language to represent skeletal story plots as plans with goals and states, and to organize the plans in goal-plan-subgoal abstraction hierarchies. STAB abduces competing story hypotheses by retrieving and instantiating plans matching the current evidence. Given the VAST data incrementally, STAB generates multiple story hypotheses, calculates their belief values, and generates predictions about future data. Background, Motivation and Goals Making sense of vast amounts of data in intelligence analysis (Heuer 1999; Krizan 1999; Thomas & Cook 2005) generally involves the tasks of recognizing and characterizing a threat based on some initial evidence about an event or activity, generating multiple explanatory hypotheses based on the evidence, collecting and assimilating additional data, evaluating the multiple explanatory hypotheses, and selecting the most plausible hypothesis. The sensemaking task is complex because of the constantly evolving, and often unreliable and conflicting, nature of data. The evolving nature of data implies a need for ongoing monitoring and continual generation and evaluation of hypotheses so that new evidence can be accounted for as it arrives and the most confident explanation can be produced at any given time. Pirolli & Card (2005) describe an information-processing model of threat analysis based on a cognitive task analysis of human analysts as they did their jobs. They have identified two main, overlapping loops in the analyst s problem solving, a foraging loop and a sensemaking loop. The foraging loop involves finding the right data sources; searching and filtering the information; and extracting the information. The sensemaking loop involves iterative development of a conceptualization (a hypothesis) from a Copyright 2007, Association for the Advancement of Artificial Intelligence ( All rights reserved. stored schema that best fits the evidence, and the presentation of the knowledge product that results from this conceptualization. As Bodnar (2005) notes, the information processing in this model is both bottom-up (from data to hypotheses) and top-down (verifiable predictions made by the instantiated hypotheses). This model, however, does not identify the content and the structure of the schemas in the sensemaking loop, or describe the process by which specific schemas are conceptualized as hypotheses. The goal of our work is to help answer these questions in a manner consistent with the current cognitive accounts of intelligence analysis (Heuer 1999; Krizan 1999; Pirolli & Card 2005). We view the task of sensemaking in threat analysis as that of abducing stories that explains the current data and makes verifiable predictions about future data. We adopt the general conceptual framework for abductive reasoning described in (Bylander et. al., 1991; Goel et. al. 1995; Josephson & Josephson 1995) and use it for story abduction. We have developed a preliminary system, called STAB (for STory ABduction), that abduces multiple, competing stories from the VAST 2006 dataset 1. Since real intelligence data is not available in the public domain, the National Visual Analytics Center (NVAC) at the Pacific Northwest National Laboratory (PNNL) generated the VAST dataset as a substitute. This synthetic dataset pertains to illegal and unethical activities, as well as normal and typical activities, in a fictitious town in the United States. STAB abduces multiple competing story hypotheses by retrieving and instantiating skeletal story plots that match the current evidence. STAB uses the TMKL knowledge representation language (Murdock & Goel 2003, 2007) to represent a skeletal story plot as a plan with goals and states, and to organize the plan in a goal-plan-subgoal abstraction hierarchy. This knowledge representation captures both intent and causality at multiple levels of abstraction. Given the VAST data incrementally, STAB generates multiple story hypotheses, calculates their belief values, and generates predictions about future data. 1 See
2 We view STAB as an automated assistant that may act as an external memory for human analysts. The external memory may support threat analysis by (a) focusing the attention of the analyst on specific hypotheses, and (b) keeping track of multiple, competing hypotheses. A Model of the VAST Dataset We begin with an introduction to the VAST-2006 dataset. The VAST dataset contains over a thousand news stories written in English, and a score of tables, maps and photographs. Figure 1 shows an example news story from the VAST dataset. We analyzed this dataset and screened the stories that indicated an illegal or unethical activity, which left about a hundred news stories out of the more than a thousand originally in the dataset. We then analyzed the remaining stories and manually extracted all the events that pertained to illegal/unethical activities; Table 1 shows a sample of such events that form the input to STAB. We also hand crafted representations for each of these events in terms of the knowledge states it produces; Table 2 shows the representations for a sample of events input to STAB. In addition, we examined the maps, photos and tables that are part of the VAST dataset and similarly extracted and represented the relevant input information. Figure 1. Example news story from the VAST dataset. TABLE 1 SAMPLE INPUTS FOR STAB Sample STAB Inputs stolen(money $40 Highway-Tire-Store) cured-disease(boynton-labs Philip-Boynton priondisease) named-after(lab Philip-Boynton Dean-USC) was-founded(boynton-labs) have-developed(boynton-labs prion-disease) announced-investigation(usfda Boynton-Labs) discontinued-investigation(usfda Boynton-Labs) Injected-mouse(Boynton-Labs prion-disease) Injected-cow(Boynton-Labs prion-disease) treatment-mouse(boynton-labs prion-disease) treatment-cow(boynton-labs prion-disease) TABLE 2 KNOWLEDGE STATE PRODUCED BY THE SAMPLE INPUT EVENTS Actions Resulting State Stolen Has-object Broken Is-broken Cured-disease Is-rich-and-famous Named-after Expert-involved Was-founded Is-open Have-developed Exists-new-disease Announced-investigation Is-investigating Discontinued-investigation Cancel-investigation Injected-cow Cow-is-infected Treatment-cow Cow-is-cured Abductive Reasoning, Plan Recognition, and Story Understanding Abduction is generally characterized as inference to the best explanation for a given set of data. The general task of abduction takes as input a set of data, and has the goal of giving as output the best explanation for the dataset. Abductive reasoning includes generation, criticism and possible acceptance of candidate explanatory hypotheses. Factors that make one explanation better than another include explanatory coverage, plausibility, and, in case of composite explanations, internal consistency and parsimony. AI research has led to a number of computational models of abduction starting with Pople s (1977) early work on medical diagnosis. These models range from normative models based on minimal set covering algorithms, e.g., (Reggia, Nau & Wang, 1983), logical reasoning from first principles, e.g., (Reiter 1987), and probabilistic inference in Bayesian networks, e.g., (Pearl 1987); to functional models based on task decomposition, e.g., (Josephson & Josephson 1994). The input to the abduction task in threat analysis is characterized by the following features (e.g., the VAST dataset): The amount of data is huge. Data comes from multiple sources and in multiple forms. Data from various sources may be unreliable and conflicting. Data arrives incrementally and is constantly evolving.
3 Data may pertain to multiple actors, where the actions of the various actors need not be coordinated. The actors may try to hide data about their actions, and may even introduce spurious data to hide their actions. Data may pertain to novel actors as well rare actions. The amount of useful evidence is a small fraction of the vast amount of data (the proverbial needle in the haystack problem). The desired output of the abduction task in threat analysis include the following: Explanations that causally relate specific sequences of actions into plans and stories. Since the data pertains to actions of multiple, uncoordinated actors, the output may contain multiple unconnected stories. Explanations that specify intent of the various actors in the stories. Ideally, the intent should be specified for specific subsequences of actions in addition to complete sequences. Belief values for the explanations that can help focus and prioritize subsequent processing of data. Explanations that can make verifiable predictions. Given the structure and complexity of the abduction task in threat analysis, we have adopted the general functional model of abductive reasoning described in (Bylander et. al., 1991; Goel et. al. 1995; Josephson & Josephson 1994). This functional mechanism for abduction has previously been used in a number of domains, such as medical data interpretation and diagnosis, scientific theory formation, speech recognition, and diagram interpretation (Josephson & Josephson 1994). This model characterizes the best composite explanation for a set of data based on three criteria (Bylander et. al, 1991 and Goel et. al (1995) provide more formal specification): 1. Coverage: One composite explanation is better than another if explains more of the observed data. 2. Belief: One composite explanation is better than another if it has a higher belief value. 3. Parsimony: One composite explanation is better than another if it is a subset of the other. The model then characterizes the abduction task as finding a composite explanation for the given set of data such that it satisfies four conditions: I. The composite explanation explains as much of the observed data as possible in the available time. II. The composite explanation is internally consistent. III. The composite explanation is parsimonious, i.e., no proper subset of the composite explanation can explain as much of the data. IV. The composite explanation is a confident explanation; an explanation is confident if its belief value is significantly higher than that of alternative explanations. Plan Recognition: Schmidt, Sridharan & Goodson (1978) identified plan recognition as an AI problem, and proposed a hypothesize-revise mechanism for addressing it. Cohen, Perrault & Allen (1982) distinguished between two kinds of plan recognition situations: keyhole, in which the observed actor acts as if unobserved, and intended, in which the observed actor performs actions to aid in the plan recognition. In threat analysis, we can distinguish additional types of plan recognition situations: covert, in which the observed actor tries to hide its actions, and multiple actor, in which the observed data contains actions of multiple actors. In multiple actor plan recognition, while the actions of some actors may be coordinated towards a shared goal, other actors may have their own independent goals. Further, the evidence of the actions of various actors in general is partially-ordered and interleaved. Charniak & McDermott (1985) identified plan recognition as an instance of the general abduction task. As with abduction, AI research since then has led to several normative models of plan recognition, including minimal set covering algorithms and logical reasoning from first principles, e.g., (Kautz 1991), and probabilistic inference in Bayesian networks, e.g. (Charniak & Goldman 1993). More recently, Goldman, Geib & Miller (1999) describe a model that also takes causal requirements of plan execution into account. At present, it is unclear whether or how these normative models of plan recognition might be used for threat analysis. For example, plan recognition in threat analysis takes partially-observed sequences of actions as its input, and desires a specification of actors goals as part of the output. In contrast, the model of Kautz (1991) requires as input a fully-observed sequence of actions, and does not explicitly represent goals. Charniak & Goldman s (1993) model requires conditional probability distributions between atomic events. While these conditional probability distributions may be available for domains in which large corpus of historical data is available, or constructed for domains which can be easily simulated (e.g., interactive games), it is unclear how we can obtain such conditional probability distributions for threat analysis. The VAST dataset, for example, contains many routine goals, plans and actions, such as purse snatching, burglaries, and vandalism. For these, we could try to obtain conditional probability distributions from investigative
4 Figure 2. The content and structure of a plan in STAB. agencies such as the local police. However, the VAST dataset also contains plans for removing a political opponent from an election by digging dirt on the opponent, and suspicions that a scientific laboratory may have released mad cow disease bacteria into a community; the conditional probabilities for such events are harder to find. Of course, we could handcraft and fine tune the conditional probabilities in STAB so that we get the right results for the VAST dataset, but that seems contrived and is unlikely to generalize. Therefore, at present we use heuristic methods for plan recognition in STAB, with the hope these methods will reveal the deep structure of the problem of threat analysis. Story Understanding: From the early work of Charniak (1977) and Cullingford (1981) to the recent work of Mueller (2004), there has been substantial research on story understanding. There also has been significant work on story generation, e.g., (Meehan 1981). STAB s task, however, is neither story understanding nor story generation. Traditional story understanding programs (such as SAM (Cullingford 1981)) take one story as input and give an understanding of that one story as output. Similarly story generation programs produce one story at a time. In contrast, STAB takes fragments from multiple news stories in the VAST dataset as input, and composes interpretations of the fragments into one or more stories as its output. A skeletal story represented as a plan in STAB is like a script (Schank & Abelson 1976). However, while a script specifies a sequence of actions, a story in STAB also specifies the goals of the action (sub-)sequences and the knowledge state(s) produced by each action. These knowledge states play an important role in the retrieval of stories relevant to a given event. Further, the goals of the action sequences in the stories specify their intent. Knowledge Representation STAB contains a library of generic, skeletal scripts relevant to the VAST domain. We found that seven major scripts appear to cover all the illegal/unethical activities in the VAST dataset. We handcrafted this library of scripts into STAB. Figure 2 illustrates a simple script in STAB s library, which is composed of several smaller scripts. The main pattern (in the middle of the figure) is to Rob a Store, which has several steps to it: Go to Store, Break into Store, Take Money. This pattern has the goal of Have Money, given the initial state of Not Have Money (top of figure). Each of the steps in this pattern can (potentially) be done using multiple methods. For example, the step of Break into Store can be done by Entering through a Window or Entering through a Door (bottom of figure). Each of these methods in turn is a process consisting of multiple steps. Figure 3 illustrates a more complex pattern of political conspiracy in which a political figure may get an opponent out of an electoral race by exposing dirt on him or having him assassinated. The story patterns are represented in the TMKL knowledge representation language (Murdock & Goel 2003, 2007). A task in TMKL represents a goal of an agent, and is specified by the knowledge states it takes as input, the knowledge states it gives as output, and relations (if any) between the input and output states. A task may be accomplished by multiple methods. A method specifies the decomposition of a task into multiple subtasks as well as the causal ordering of the subtasks for accomplishing the task, and is represented as a finite state machine. Thus, the TMKL representation of a plan pattern captures both intent and causality at multiple levels of abstraction.
5 Figure 3: The plan for a political conspiracy intended to remove an opponent from an electoral race. Activated nodes are denoted by a thick outline around yellow boxes and with bold text. Although we originally developed TMKL to capture an agent s self-model of its own knowledge, reasoning and architecture, it has since been used in several other applications. TIELT (Molineaux & Aha, 2005), an environment for evaluating learning in computer games, uses TMKL as its agent description language. The AHEAD system (Murdock, Aha & Breslow 2003) uses TMKL for representing hypotheses about asymmetric threats. In particular, AHEAD uses past cases of such threats to generate arguments for and against a given hypothesis. The explicit specification of the goals of subsequences of actions allows AHEAD to retrieve past cases relevant to specific subsequences. TMKL is more expressive than Hierarchical Task Networks (HTNs) (Erol, Hendler & Nau 1994), but HTN s implicitly provide support for features explicitly represented in TMKL. In a recent experiment, Hoang, Lee- Urban & Munoz-Avila (2005) encoded the same gameplaying agent in both TMKL and HTN. They found that TMKL provides constructs for looping, conditional execution, assignment functions with return values, and other features not found in HTN. They also found that since HTN implicitly provides support for the same features, translation from TMKL to HTN is always possible. Thus, while TMKL shares the good computational properties of HTNs, it makes reasoning easier and more perspicuous. Computational Process Figure 4 shows the high-level architecture of STAB. First, The Evidence Collector collects the input events in an Evidence File in chronological order. Next, the Story Matcher takes one input event at a time and uses its resulting knowledge state of the event with the task nodes in the TMKL representations of the scripts stored in the Story Library. The Story Matcher tags the matching tasks and passes the matching plans to a Working Memory. Then, the Story Matcher inspects the next input event in the Evidence File and repeats the above process. If the new input event results in the retrieval of a new script, then the corresponding plan is similarly stored in the Working Memory. If the newly retrieved plan is already in the Working Memory, then additional task nodes that match the new input are also tagged but only one plan instance is kept. Figures 5 & 6 illustrate the two script plans, Rob a Store and Commit Vandalism, respectively, whose task nodes match the input event Break(Window). The matching task nodes are shown with a thick outline around yellow boxes and with bold text. Note that when a leaf task node in a plan (e.g., Break(Window) in the Rob a Store plot) is activated, then the higher-level task nodes in the plan that provide the intentional contexts for the leaf node (Break into(store) & Rob(Store)) are also activated. STAB stores the multiple competing plans (Rob a Store and Commit Vandalism) in its Working Memory and assigns belief values to them. The belief value of a plan hypothesis depends on the proportion of the task nodes in a plan that are matched by the input evidence (higher the proportion, higher is the belief value) and the level of abstraction of the matched task nodes (higher the abstraction level, more is the weight of the node). Equation (1) represents the formula for calculating belief values where level is the depth of the task within the hierarchy of
6 Figure 4: High-Level Architecture of STAB. the plan and n is the maximum depth of the task hierarchy for the plan. As an example, the belief value for the Commit Vandalism plan (Fig. 6) is (100% / 1) + (50% / 2) = Note that only the sub-tree of the method with activated tasks is used in the belief calculation. Similarly, the belief value of the Rob a Store before the Take(Money) node is activated equals n level = 1 # activated tasks at level/total tasks at level level The plan hypotheses in the Working Memory generate expectations. Thus, the Rob a Store hypothesis generates expectations about the events Go to (Store), Enter (Building), and Take (Money), while the Commit Vandalism hypothesis generates expectation about only Kick In (Door). As additional data arrives as input in the Evidence File, STAB matches the data with the expectations generated by the candidate hypotheses. If, for example, the new data contains evidence about Take (1) (Money), then this node too in the Rob a Store story is tagged, and Equation 1 is used to update the belief value of the hypothesis to If the new data contains evidence that contradicts an expectation generated by a hypothesis, then the hypothesis is considered as refuted, and its belief value is reduced to 0. At the end, STAB generates a report which displays all current plan hypotheses (including refuted hypotheses, if any), the belief value of each hypothesis, and the evidence for and against each hypothesis. Since STAB continually monitors the Evidence File and updates its Working Memory, the user may at any point query STAB to inspect the current hypotheses and the related evidence. Evaluation Evaluation of STAB is challenging for two reasons. Firstly, since there is little intelligence data available in the public domain in a suitable form, we cannot directly test Figure 5: The activated nodes in the Rob a Store plan.
7 expectations generated by a hypothesis but also for evidence that may contradict the expectations. Figure 6: The activated nodes in the Commit Vandalism plan. STAB. Secondly, given the high cost of an analyst s time, evaluating STAB with a human intelligence analyst can be justified and made possible only after STAB is validated as a proof of concept system. Thus, our evaluation of STAB as a proof of concept system has taken two alternative threads. Firstly, we have demonstrated STAB to an expert who has done significant work with intelligence analysts. This expert found STAB s knowledge representations and information processing as plausible. However, the expert also raised serious concerns about the usability of STAB s preliminary graphical interface. Secondly, we have begun evaluating STAB for the new VAST-2007 dataset 2 recently released by PNNL. As with the VAST-2006 dataset, we handcrafted representations of events in the VAST-2007 dataset corresponding to illegal/unethical activities. When these events were given to STAB as input, we found that STAB invoked six plans (of the seven stored in its library), three plans with high belief values and the other three with relatively low belief values. Our preliminary analysis suggests that STAB thus makes the right plan hypotheses for the VAST-2007 dataset. Summary, Conclusions, and Future Work Psychological studies of intelligence analysis (Heuer 1999) indicate the three main errors made by human analysts in hypothesis generation: (1) Due to limitations of human memory, intelligence analysts may have difficulty keeping track of multiple explanations for a set of data over a long period of time. (2) Analysts may quickly decide on a single hypothesis for the data set and stick to it even as new data arrives. (3) Analysts may look for data that supports the hypothesis on which they are fixated, and not necessarily the data that may refute the hypothesis. STAB attempts to address these limitations. Firstly, there are no limitations on the size of STAB s library or its working memory. On the contrary, STAB offers a non-volatile memory of both generic scripts and specific plan hypotheses. Secondly, for each new additional input event, STAB examines all the scripts whose task nodes match the input. Thus, it is not fixated on any particular hypothesis. Thirdly, STAB explicitly looks not only for evidence that may confirm the 2 STAB provides preliminary, but verifiable, answers to two questions arising from Pirroli & Card s model of sensemaking: what is the content of the schemas, and what is process for developing a hypothesis? According to STAB, the schemas are scripts, where the scripts pertain the activities of interest to an analyst, e.g., illegal/unethical activities in the VAST domain. The scripts are represented as plans with goals and states. The TMKL language provides a scheme for representing intent and causality of sequences of events at multiple levels of abstraction. Scripts are retrieved by matching input events with the task nodes in the script representations. Belief value of a plan hypothesis depends on the proportion and level of the matched task nodes. The retrieved stories generate expectations about data, and the confirmation of these expectations results in updates to the belief values. In its present state of development, STAB has many limitations. Firstly, the input events to STAB are extracted from new stories and represented by hand. We are building a mechanism for automatic extraction of input events from news stories. Secondly, STAB uses a simple, heuristic method for calculating belief values. We are developing a more robust method for calculating belief values based on the principles of coverage and parsimony. Thirdly, STAB does not propagate the values of variables between the nodes in a plan pattern. We are augmenting TMKL to automate variable propagation. Finally, we are conducting usability tests with STAB s interface, and constructing a more usable graphical interface. Acknowledgements This research has benefited from contributions by Neha Sugandh and Tanvi Bhadbhade, and discussions with John Stasko. We thank Jean Scholtz (PNNL) for discussions on STAB s evaluation. We also thank Kristin Cook (PNNL) and anonymous reviewers for their comments on earlier drafts. This work was sponsored by NVAC under the auspices of the Southeastern Regional Visualization and Analytics Center. NVAC is a U.S. Department of Homeland Security Program led by PNNL. References E. Bodnar. (2005) Making Sense of Massive Data by Hypothesis Testing. In Proceedings of 2005 International Conference on Intelligence Analysis, May T. Bylander, D. Allemang, M. Tanner, J. Josephson. (1991) The Computational Complexity of Abduction. Artificial Intelligence, 49(1-3): 25-60, 1991.
8 E. Charniak. (1977) Ms. Malaprop, A Language Comprehension Program. In Proc. Fifth International Conference on Artificial Intelligence (IJCAI-77). E. Charniak and R. Goldman. (1993) A Bayesian Model of Plan Recognition. Artificial Intelligence, 64(1): 53-79, E. Charniak and D. McDermott. (1985) Introduction to Artificial Intelligence. Reading MA: Addison-Wesley, P. Cohen, C. Perrault & J. Allen. (1982) Beyond Question Answering. In Strategies for Natural Language Processing, W. Lehnert & M. Ringle (eds), Erlbaum, R. Cullingford. (1981) SAM and Micro SAM. In R. Schank, & C. Riesbeck (Eds.), Inside Computer Understanding: Five Programs Plus Miniatures. Hillsdale, NJ: Erlbaum. K. Erol, J. Hendler & D. Nau. (1994) HTN Planning: Complexity and Expressivity. In Proc. Twelfth National Conference on Artificial Intelligence (AAAI-94). A. Goel, J. Josephson, O. Fischer & P. Sadayappan. (1995) Practical Abduction: Characterization, Decomposition and Distribution. Journal of Experimental and Theoretical Artificial Intelligence, 7: , R. Goldman, C. Geib & C. Miller. (1999) A New Model of Plan Recognition. In Proc Conference on Uncertainty in Intelligence, Stockholm, R. J. Heuer. (1999) Psychology of Intelligence Analysis. Center for the Study of Intelligence, Central Intelligence Agency, H. Hoang, S. Lee-Urban & H. Muñoz-Avila. (2005) Hierarchical Plan Representations for Encoding Strategic Game AI. In Proc. Artificial Intelligence and Interactive Digital Entertainment Conference (AIIDE-05). J. Josephson & S. Josephson (editors). (1994) Abductive Inference: Computation, Philosophy, Technology. Cambridge Univerity Press, H. Kautz. (1991) A Formal Theory of Plan Recognition and Its Implementation. In Reasoning About Plans, J. Allen, H. Kautz, R. Pelavin & J. Tenenberg (eds), Morgan Kauffman, pp L. Krizan, (1999) Intelligence Essentials for Everyone, Joint Military Intelligence College, J. Meehan. (1981) TALE-SPIN and Micro TALE-SPIN. In R. Schank, & C. Riesbeck (Eds.), Inside Computer Understanding: Five Programs Plus Miniatures. Hillsdale, NJ: Erlbaum. M. Molineaux & D. Aha. (2005) TIELT: A Testbed for Gaming Environments. In Proc National Conference on AI (AAAI 2005), pp E. Mueller. (2004) Understanding Script-Based Stories using Commonsense Reasoning. Cognitive Systems Research, 5(4), J. Murdock, D. Aha & L. Breslow. (2003) Assessing Elaborated Hypotheses: An Interpretive Case-Based Reasoning Approach. In Proc. Fifth International Conference on Case-Based Reasoning. Trondheim, Norway, June J. Murdock & A. Goel. (2003) Localizing Planning with Functional Process Models. In Proc. Thirteenth International Conference on Automated Planning & Scheduling (ICAPS'03). Trento, Italy, June 9-13, J. Murdock & A. Goel. (2007) Meta-Case-Based Reasoning: Self-Improvement through Self-Understanding. To appear in Journal of Experimental & Theoretical Artificial Intelligence, J. Pearl. (1987) Distributed Revision of Composite Beliefs. Artificial Intelligence, 33: P. Pirolli and S. Card. (2005) The Sensemaking Process and Leverage Points for Analyst Technology as Identified Through Cognitive Task Analysis. In Proceedings of 2005 International Conference on Intelligence Analysis, May H. Pople. (1977) The Formation of Composite Hypotheses in Diagnosis: Problem Solving and Synthetic Reasoning. In Proc. Fifth International Conference on AI (IJCAI-77), pp R. Schank & R. Abelson. (1977) Scripts, Plans, Goals and Understanding. Erlbaum, C. Schmidt, N.S. Sridharan & J. Goodson. (1978) The Plan Recognition Problem: An Intersection of Psychology and Artificial Intelligence. Artificial Intelligence, 11(1-2), J. Thomas & K. Cook. (2005) Illuminating the Path, National Visual Analytics Center, Pacific Northwest National Laboratory, 2005.
Agenda. Interface Agents. Interface Agents
Agenda Marcelo G. Armentano Problem Overview Interface Agents Probabilistic approach Monitoring user actions Model of the application Model of user intentions Example Summary ISISTAN Research Institute
Integrating Cognitive Models Based on Different Computational Methods
Integrating Cognitive Models Based on Different Computational Methods Nicholas L. Cassimatis ([email protected]) Rensselaer Polytechnic Institute Department of Cognitive Science 110 8 th Street Troy, NY 12180
Information Visualization WS 2013/14 11 Visual Analytics
1 11.1 Definitions and Motivation Lot of research and papers in this emerging field: Visual Analytics: Scope and Challenges of Keim et al. Illuminating the path of Thomas and Cook 2 11.1 Definitions and
Another Look at Sensitivity of Bayesian Networks to Imprecise Probabilities
Another Look at Sensitivity of Bayesian Networks to Imprecise Probabilities Oscar Kipersztok Mathematics and Computing Technology Phantom Works, The Boeing Company P.O.Box 3707, MC: 7L-44 Seattle, WA 98124
A Bayesian Network Model for Diagnosis of Liver Disorders Agnieszka Onisko, M.S., 1,2 Marek J. Druzdzel, Ph.D., 1 and Hanna Wasyluk, M.D.,Ph.D.
Research Report CBMI-99-27, Center for Biomedical Informatics, University of Pittsburgh, September 1999 A Bayesian Network Model for Diagnosis of Liver Disorders Agnieszka Onisko, M.S., 1,2 Marek J. Druzdzel,
Improving Knowledge-Based System Performance by Reordering Rule Sequences
Improving Knowledge-Based System Performance by Reordering Rule Sequences Neli P. Zlatareva Department of Computer Science Central Connecticut State University 1615 Stanley Street New Britain, CT 06050
Application Design: Issues in Expert System Architecture. Harry C. Reinstein Janice S. Aikins
Application Design: Issues in Expert System Architecture Harry C. Reinstein Janice S. Aikins IBM Scientific Center 15 30 Page Mill Road P. 0. Box 10500 Palo Alto, Ca. 94 304 USA ABSTRACT We describe an
Data Validation During Diagnosis, A Step Beyond Traditional Sensor Validation.
From: AAAI-87 Proceedings. Copyright 1987, AAAI (www.aaai.org). All rights reserved. Data Validation During Diagnosis, A Step Beyond Traditional Sensor Validation. B. Chandrasekaran and W.F Punch III Laboratory
Masters in Information Technology
Computer - Information Technology MSc & MPhil - 2015/6 - July 2015 Masters in Information Technology Programme Requirements Taught Element, and PG Diploma in Information Technology: 120 credits: IS5101
KNOWLEDGE ORGANIZATION
KNOWLEDGE ORGANIZATION Gabi Reinmann Germany [email protected] Synonyms Information organization, information classification, knowledge representation, knowledge structuring Definition The term
Overview of the TACITUS Project
Overview of the TACITUS Project Jerry R. Hobbs Artificial Intelligence Center SRI International 1 Aims of the Project The specific aim of the TACITUS project is to develop interpretation processes for
In Proceedings of the Eleventh Conference on Biocybernetics and Biomedical Engineering, pages 842-846, Warsaw, Poland, December 2-4, 1999
In Proceedings of the Eleventh Conference on Biocybernetics and Biomedical Engineering, pages 842-846, Warsaw, Poland, December 2-4, 1999 A Bayesian Network Model for Diagnosis of Liver Disorders Agnieszka
Graduate Co-op Students Information Manual. Department of Computer Science. Faculty of Science. University of Regina
Graduate Co-op Students Information Manual Department of Computer Science Faculty of Science University of Regina 2014 1 Table of Contents 1. Department Description..3 2. Program Requirements and Procedures
Masters in Human Computer Interaction
Masters in Human Computer Interaction Programme Requirements Taught Element, and PG Diploma in Human Computer Interaction: 120 credits: IS5101 CS5001 CS5040 CS5041 CS5042 or CS5044 up to 30 credits from
Self Organizing Maps for Visualization of Categories
Self Organizing Maps for Visualization of Categories Julian Szymański 1 and Włodzisław Duch 2,3 1 Department of Computer Systems Architecture, Gdańsk University of Technology, Poland, [email protected]
Design Problem Solving: A Task Analysis 1
Design Problem Solving: A Task Analysis 1 Articles B. Chandrasekaran Design problem solving is a complex activity involving a number of subtasks and a number of alternative methods potentially available
Simulating the Structural Evolution of Software
Simulating the Structural Evolution of Software Benjamin Stopford 1, Steve Counsell 2 1 School of Computer Science and Information Systems, Birkbeck, University of London 2 School of Information Systems,
Masters in Computing and Information Technology
Masters in Computing and Information Technology Programme Requirements Taught Element, and PG Diploma in Computing and Information Technology: 120 credits: IS5101 CS5001 or CS5002 CS5003 up to 30 credits
The Science of Analytical Reasoning
It is not enough to have a good mind. The main thing is to use it well. Rene Descartes, Discourse on Method, 1637 2 The Science of Analytical Reasoning When we create a mental picture, speak of the mind
ICT Perspectives on Big Data: Well Sorted Materials
ICT Perspectives on Big Data: Well Sorted Materials 3 March 2015 Contents Introduction 1 Dendrogram 2 Tree Map 3 Heat Map 4 Raw Group Data 5 For an online, interactive version of the visualisations in
Reusable Knowledge-based Components for Building Software. Applications: A Knowledge Modelling Approach
Reusable Knowledge-based Components for Building Software Applications: A Knowledge Modelling Approach Martin Molina, Jose L. Sierra, Jose Cuena Department of Artificial Intelligence, Technical University
2 AIMS: an Agent-based Intelligent Tool for Informational Support
Aroyo, L. & Dicheva, D. (2000). Domain and user knowledge in a web-based courseware engineering course, knowledge-based software engineering. In T. Hruska, M. Hashimoto (Eds.) Joint Conference knowledge-based
School of Computer Science
School of Computer Science Computer Science - Honours Level - 2014/15 October 2014 General degree students wishing to enter 3000- level modules and non- graduating students wishing to enter 3000- level
Hierarchical Judgement Composition: Revisiting the Structural Credit Assignment Problem
Hierarchical Judgement Composition: Revisiting the Structural Credit Assignment Problem Joshua Jones, Ashok Goel College of Computing Georgia Institute of Technology Atlanta, USA 30332 {jkj, goel}@cc.gatech.edu
Listening to the Customer s Voice 1
Listening to the Customer s Voice 1 Karl E. Wiegers Process Impact 716-377-5110 www.processimpact.com Perhaps the greatest challenge facing the software developer is sharing the vision of the final product
Stabilization by Conceptual Duplication in Adaptive Resonance Theory
Stabilization by Conceptual Duplication in Adaptive Resonance Theory Louis Massey Royal Military College of Canada Department of Mathematics and Computer Science PO Box 17000 Station Forces Kingston, Ontario,
Masters in Advanced Computer Science
Masters in Advanced Computer Science Programme Requirements Taught Element, and PG Diploma in Advanced Computer Science: 120 credits: IS5101 CS5001 up to 30 credits from CS4100 - CS4450, subject to appropriate
Appendix B Data Quality Dimensions
Appendix B Data Quality Dimensions Purpose Dimensions of data quality are fundamental to understanding how to improve data. This appendix summarizes, in chronological order of publication, three foundational
Masters in Artificial Intelligence
Masters in Artificial Intelligence Programme Requirements Taught Element, and PG Diploma in Artificial Intelligence: 120 credits: IS5101 CS5001 CS5010 CS5011 CS4402 or CS5012 in total, up to 30 credits
Computational Scientific Discovery and Cognitive Science Theories
Computational Scientific Discovery and Cognitive Science Theories Peter D Sozou University of Liverpool and LSE Joint work with: Mark Addis Birmingham City University and LSE Fernand Gobet University of
Course Syllabus For Operations Management. Management Information Systems
For Operations Management and Management Information Systems Department School Year First Year First Year First Year Second year Second year Second year Third year Third year Third year Third year Third
Masters in Networks and Distributed Systems
Masters in Networks and Distributed Systems Programme Requirements Taught Element, and PG Diploma in Networks and Distributed Systems: 120 credits: IS5101 CS5001 CS5021 CS4103 or CS5023 in total, up to
Learning is a very general term denoting the way in which agents:
What is learning? Learning is a very general term denoting the way in which agents: Acquire and organize knowledge (by building, modifying and organizing internal representations of some external reality);
A Tale of Alderwood: Visualizing Relationships in a Diverse Data Collection
A Tale of Alderwood: Visualizing Relationships in a Diverse Data Collection Summer Adams Susan Gov Sheena Lewis Kanupriya Singhal College of Computing Georgia Institute of Technology Atlanta, GA ABSTRACT
Sanjeev Kumar. contribute
RESEARCH ISSUES IN DATAA MINING Sanjeev Kumar I.A.S.R.I., Library Avenue, Pusa, New Delhi-110012 [email protected] 1. Introduction The field of data mining and knowledgee discovery is emerging as a
What is Artificial Intelligence?
CSE 3401: Intro to Artificial Intelligence & Logic Programming Introduction Required Readings: Russell & Norvig Chapters 1 & 2. Lecture slides adapted from those of Fahiem Bacchus. 1 What is AI? What is
Improving your Data Warehouse s IQ
Improving your Data Warehouse s IQ Derek Strauss Gavroshe USA, Inc. Outline Data quality for second generation data warehouses DQ tool functionality categories and the data quality process Data model types
IAI : Expert Systems
IAI : Expert Systems John A. Bullinaria, 2005 1. What is an Expert System? 2. The Architecture of Expert Systems 3. Knowledge Acquisition 4. Representing the Knowledge 5. The Inference Engine 6. The Rete-Algorithm
CFSD 21 ST CENTURY SKILL RUBRIC CRITICAL & CREATIVE THINKING
Critical and creative thinking (higher order thinking) refer to a set of cognitive skills or strategies that increases the probability of a desired outcome. In an information- rich society, the quality
131-1. Adding New Level in KDD to Make the Web Usage Mining More Efficient. Abstract. 1. Introduction [1]. 1/10
1/10 131-1 Adding New Level in KDD to Make the Web Usage Mining More Efficient Mohammad Ala a AL_Hamami PHD Student, Lecturer m_ah_1@yahoocom Soukaena Hassan Hashem PHD Student, Lecturer soukaena_hassan@yahoocom
Analysis of Data Mining Concepts in Higher Education with Needs to Najran University
590 Analysis of Data Mining Concepts in Higher Education with Needs to Najran University Mohamed Hussain Tawarish 1, Farooqui Waseemuddin 2 Department of Computer Science, Najran Community College. Najran
CSC384 Intro to Artificial Intelligence
CSC384 Intro to Artificial Intelligence What is Artificial Intelligence? What is Intelligence? Are these Intelligent? CSC384, University of Toronto 3 What is Intelligence? Webster says: The capacity to
Chapter 28. Bayesian Networks
Chapter 28. Bayesian Networks The Quest for Artificial Intelligence, Nilsson, N. J., 2009. Lecture Notes on Artificial Intelligence, Spring 2012 Summarized by Kim, Byoung-Hee and Lim, Byoung-Kwon Biointelligence
MANAGING QUEUE STABILITY USING ART2 IN ACTIVE QUEUE MANAGEMENT FOR CONGESTION CONTROL
MANAGING QUEUE STABILITY USING ART2 IN ACTIVE QUEUE MANAGEMENT FOR CONGESTION CONTROL G. Maria Priscilla 1 and C. P. Sumathi 2 1 S.N.R. Sons College (Autonomous), Coimbatore, India 2 SDNB Vaishnav College
Contact Recommendations from Aggegrated On-Line Activity
Contact Recommendations from Aggegrated On-Line Activity Abigail Gertner, Justin Richer, and Thomas Bartee The MITRE Corporation 202 Burlington Road, Bedford, MA 01730 {gertner,jricher,tbartee}@mitre.org
Bachelor of Games and Virtual Worlds (Programming) Subject and Course Summaries
First Semester Development 1A On completion of this subject students will be able to apply basic programming and problem solving skills in a 3 rd generation object-oriented programming language (such as
A Cognitive Approach to Vision for a Mobile Robot
A Cognitive Approach to Vision for a Mobile Robot D. Paul Benjamin Christopher Funk Pace University, 1 Pace Plaza, New York, New York 10038, 212-346-1012 [email protected] Damian Lyons Fordham University,
A Bayesian Approach for on-line max auditing of Dynamic Statistical Databases
A Bayesian Approach for on-line max auditing of Dynamic Statistical Databases Gerardo Canfora Bice Cavallo University of Sannio, Benevento, Italy, {gerardo.canfora,bice.cavallo}@unisannio.it ABSTRACT In
Experiments in Web Page Classification for Semantic Web
Experiments in Web Page Classification for Semantic Web Asad Satti, Nick Cercone, Vlado Kešelj Faculty of Computer Science, Dalhousie University E-mail: {rashid,nick,vlado}@cs.dal.ca Abstract We address
Application of Adaptive Probing for Fault Diagnosis in Computer Networks 1
Application of Adaptive Probing for Fault Diagnosis in Computer Networks 1 Maitreya Natu Dept. of Computer and Information Sciences University of Delaware, Newark, DE, USA, 19716 Email: [email protected]
Email Spam Detection Using Customized SimHash Function
International Journal of Research Studies in Computer Science and Engineering (IJRSCSE) Volume 1, Issue 8, December 2014, PP 35-40 ISSN 2349-4840 (Print) & ISSN 2349-4859 (Online) www.arcjournals.org Email
ONTOLOGY-BASED APPROACH TO DEVELOPMENT OF ADJUSTABLE KNOWLEDGE INTERNET PORTAL FOR SUPPORT OF RESEARCH ACTIVITIY
ONTOLOGY-BASED APPROACH TO DEVELOPMENT OF ADJUSTABLE KNOWLEDGE INTERNET PORTAL FOR SUPPORT OF RESEARCH ACTIVITIY Yu. A. Zagorulko, O. I. Borovikova, S. V. Bulgakov, E. A. Sidorova 1 A.P.Ershov s Institute
Investment Analysis using the Portfolio Analysis Machine (PALMA 1 ) Tool by Richard A. Moynihan 21 July 2005
Investment Analysis using the Portfolio Analysis Machine (PALMA 1 ) Tool by Richard A. Moynihan 21 July 2005 Government Investment Analysis Guidance Current Government acquisition guidelines mandate the
LONG BEACH CITY COLLEGE MEMORANDUM
LONG BEACH CITY COLLEGE MEMORANDUM DATE: May 5, 2000 TO: Academic Senate Equivalency Committee FROM: John Hugunin Department Head for CBIS SUBJECT: Equivalency statement for Computer Science Instructor
Bayesian networks - Time-series models - Apache Spark & Scala
Bayesian networks - Time-series models - Apache Spark & Scala Dr John Sandiford, CTO Bayes Server Data Science London Meetup - November 2014 1 Contents Introduction Bayesian networks Latent variables Anomaly
Measurement Information Model
mcgarry02.qxd 9/7/01 1:27 PM Page 13 2 Information Model This chapter describes one of the fundamental measurement concepts of Practical Software, the Information Model. The Information Model provides
72. Ontology Driven Knowledge Discovery Process: a proposal to integrate Ontology Engineering and KDD
72. Ontology Driven Knowledge Discovery Process: a proposal to integrate Ontology Engineering and KDD Paulo Gottgtroy Auckland University of Technology [email protected] Abstract This paper is
Database Marketing, Business Intelligence and Knowledge Discovery
Database Marketing, Business Intelligence and Knowledge Discovery Note: Using material from Tan / Steinbach / Kumar (2005) Introduction to Data Mining,, Addison Wesley; and Cios / Pedrycz / Swiniarski
DEGREE PLAN INSTRUCTIONS FOR COMPUTER ENGINEERING
DEGREE PLAN INSTRUCTIONS FOR COMPUTER ENGINEERING Fall 2000 The instructions contained in this packet are to be used as a guide in preparing the Departmental Computer Science Degree Plan Form for the Bachelor's
ACH 1.1 : A Tool for Analyzing Competing Hypotheses Technical Description for Version 1.1
ACH 1.1 : A Tool for Analyzing Competing Hypotheses Technical Description for Version 1.1 By PARC AI 3 Team with Richards Heuer Lance Good, Jeff Shrager, Mark Stefik, Peter Pirolli, & Stuart Card ACH 1.1
A View Integration Approach to Dynamic Composition of Web Services
A View Integration Approach to Dynamic Composition of Web Services Snehal Thakkar, Craig A. Knoblock, and José Luis Ambite University of Southern California/ Information Sciences Institute 4676 Admiralty
2. Analysis, Design and Implementation
2. Subject/Topic/Focus: Software Production Process Summary: Software Crisis Software as a Product: From Individual Programs to Complete Application Systems Software Development: Goals, Tasks, Actors,
Visionet IT Modernization Empowering Change
Visionet IT Modernization A Visionet Systems White Paper September 2009 Visionet Systems Inc. 3 Cedar Brook Dr. Cranbury, NJ 08512 Tel: 609 360-0501 Table of Contents 1 Executive Summary... 4 2 Introduction...
A Conceptual Approach to Data Visualization for User Interface Design of Smart Grid Operation Tools
A Conceptual Approach to Data Visualization for User Interface Design of Smart Grid Operation Tools Dong-Joo Kang and Sunju Park Yonsei University [email protected], [email protected] Abstract
THE DEVELOPMENT OF AN EXPERT CAR FAILURE DIAGNOSIS SYSTEM WITH BAYESIAN APPROACH
Journal of Computer Science 9 (10): 1383-1388, 2013 ISSN: 1549-3636 2013 doi:10.3844/jcssp.2013.1383.1388 Published Online 9 (10) 2013 (http://www.thescipub.com/jcs.toc) THE DEVELOPMENT OF AN EXPERT CAR
Engineering Process Software Qualities Software Architectural Design
Engineering Process We need to understand the steps that take us from an idea to a product. What do we do? In what order do we do it? How do we know when we re finished each step? Production process Typical
Delivering the power of the world s most successful genomics platform
Delivering the power of the world s most successful genomics platform NextCODE Health is bringing the full power of the world s largest and most successful genomics platform to everyday clinical care NextCODE
Updating Action Domain Descriptions
Updating Action Domain Descriptions Thomas Eiter, Esra Erdem, Michael Fink, and Ján Senko Institute of Information Systems, Vienna University of Technology, Vienna, Austria Email: (eiter esra michael jan)@kr.tuwien.ac.at
Putting IBM Watson to Work In Healthcare
Martin S. Kohn, MD, MS, FACEP, FACPE Chief Medical Scientist, Care Delivery Systems IBM Research [email protected] Putting IBM Watson to Work In Healthcare 2 SB 1275 Medical data in an electronic or
LOG AND EVENT MANAGEMENT FOR SECURITY AND COMPLIANCE
PRODUCT BRIEF LOG AND EVENT MANAGEMENT FOR SECURITY AND COMPLIANCE The Tripwire VIA platform delivers system state intelligence, a continuous approach to security that provides leading indicators of breach
A Service Modeling Approach with Business-Level Reusability and Extensibility
A Service Modeling Approach with Business-Level Reusability and Extensibility Jianwu Wang 1,2, Jian Yu 1, Yanbo Han 1 1 Institute of Computing Technology, Chinese Academy of Sciences, 100080, Beijing,
Chapter 10 Practical Database Design Methodology and Use of UML Diagrams
Chapter 10 Practical Database Design Methodology and Use of UML Diagrams Copyright 2011 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Chapter 10 Outline The Role of Information Systems in
A Framework of Context-Sensitive Visualization for User-Centered Interactive Systems
Proceedings of 10 th International Conference on User Modeling, pp423-427 Edinburgh, UK, July 24-29, 2005. Springer-Verlag Berlin Heidelberg 2005 A Framework of Context-Sensitive Visualization for User-Centered
Important dimensions of knowledge Knowledge is a firm asset: Knowledge has different forms Knowledge has a location Knowledge is situational Wisdom:
Southern Company Electricity Generators uses Content Management System (CMS). Important dimensions of knowledge: Knowledge is a firm asset: Intangible. Creation of knowledge from data, information, requires
MONITORING AND DIAGNOSIS OF A MULTI-STAGE MANUFACTURING PROCESS USING BAYESIAN NETWORKS
MONITORING AND DIAGNOSIS OF A MULTI-STAGE MANUFACTURING PROCESS USING BAYESIAN NETWORKS Eric Wolbrecht Bruce D Ambrosio Bob Paasch Oregon State University, Corvallis, OR Doug Kirby Hewlett Packard, Corvallis,
INTERNATIONAL JOURNAL OF COMPUTER ENGINEERING & TECHNOLOGY (IJCET)
INTERNATIONAL JOURNAL OF COMPUTER ENGINEERING & TECHNOLOGY (IJCET) International Journal of Computer Engineering and Technology (IJCET), ISSN 0976 6367(Print), ISSN 0976 6367(Print) ISSN 0976 6375(Online)
Architecture Artifacts Vs Application Development Artifacts
Architecture Artifacts Vs Application Development Artifacts By John A. Zachman Copyright 2000 Zachman International All of a sudden, I have been encountering a lot of confusion between Enterprise Architecture
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
Software Engineering of NLP-based Computer-assisted Coding Applications
Software Engineering of NLP-based Computer-assisted Coding Applications 1 Software Engineering of NLP-based Computer-assisted Coding Applications by Mark Morsch, MS; Carol Stoyla, BS, CLA; Ronald Sheffer,
A STRATEGIC PLANNER FOR ROBOT EXCAVATION' by Humberto Romero-Lois, Research Assistant, Department of Civil Engineering
A STRATEGIC PLANNER FOR ROBOT EXCAVATION' by Humberto Romero-Lois, Research Assistant, Department of Civil Engineering Chris Hendrickson, Professor, Department of Civil Engineering, and Irving Oppenheim,
Using an Instructional Systems Development Model as a Framework for Research on Scale Up 1
Using an Instructional Systems Development Model as a Framework for Research on Scale Up 1 Michael R. Vitale East Carolina University Nancy R. Romance Florida Atlantic University Abstract This paper presents
How To Find Influence Between Two Concepts In A Network
2014 UKSim-AMSS 16th International Conference on Computer Modelling and Simulation Influence Discovery in Semantic Networks: An Initial Approach Marcello Trovati and Ovidiu Bagdasar School of Computing
Technology WHITE PAPER
Technology WHITE PAPER What We Do Neota Logic builds software with which the knowledge of experts can be delivered in an operationally useful form as applications embedded in business systems or consulted
Pitfalls and Best Practices in Role Engineering
Bay31 Role Designer in Practice Series Pitfalls and Best Practices in Role Engineering Abstract: Role Based Access Control (RBAC) and role management are a proven and efficient way to manage user permissions.
