Medical-Miner at TREC 2011 Medical Records Track
|
|
|
- Pauline Cook
- 10 years ago
- Views:
Transcription
1 Medical-Miner at TREC 2011 Medical Records Track 1 J.M. Córdoba, 1 M.J. Maña, 1 N.P. Cruz, 1 J. Mata, 2 F. Aparicio, 2 M. Buenaga, 3 D. Glez-Peña, 3 F. Fdez-Riverola 1 Universidad de Huelva 2 Universidad Europea de Madrid 3 Universidad de Vigo Abstract This paper presents work done at Medical Minner Project on the TREC-2011 Medical Records Track. The paper proposes four models for medical information retrieval based on Lucene index approach. Our retrieval engine used an Lucen Index scheme with traditional stopping and stemming, enhanced with entity recognition software on query terms. Our aim in this rst competition is to set a broader project that involves the develop of a congurable Apache Lucene-based framework that allows the rapid development of medical search facilities. Results around the track median have been achieved. In this exploratory track, we think that these results are a good beginning and encourage us for future developments. 1 Introduction Clinical databases have accumulated large quantities of information about patients and their medical conditions. Relationships and patterns within this data could provide new medical knowledge. Unfortunately, few methodologies have been developed and applied to discover this hidden knowledge. In the TREC framework, Medical Records track encourages to research on providing contentbased access to the free-text elds of electronic medical records. Related with this same research area, the Medical Minner initiative 1 works for provide new useful tools in this environment, enabling the achievement of a double benet: rst, to provide specic and relevant mechanisms to a professional environment (biomedical), and secondly, obtaining innovative techniques and signicant advances in research in the area of information retrieval, particularly on this domain, and based on a signicant integration eort. In this context, this paper presents the models developed for our rst participation in TREC, particularly in the Medical Records track
2 This paper is organised as follows. In Section 2 we describe some of the ways have been used to represent the documents elds in reports processing, and describe the way the index used in our later experiments is built automatically from document collection. In Section 3 we cover the overall architecture of a indexing application and where Lucene ts. The document retrieval models described in Section 4. This section, also covers the further development and elaboration of the models. Data and results tables are given in Section 5 with key results summarised. Finally, in Section 6, some conclusions are provided. 2 Document Collection Preprocessing Assuming the system has access to a large document collection as a knowledge resource for answering medical topics, this collection may need to be processed before querying, in order to transform it into a form which is appropriate for topic answering. Thus, in order to eective TREC medical visits retrieval, some issues need to be solved. First, reports need to be grouped into visits. Next, XML Reports must be adapted to be queried in a rapid and accurate way. Additionally, another type of additional processing can be performed. In our case, we used the ICD codes for better retrieval. 2.1 Visit Procesing As the Medical Records track will use the visit as the answer unit then is there need for additional processing in order to group the reports into visits. The Medical Records Track repository contains a simple ASCII table called the Report-to-Visit Mapping Key that species which reports belong to the same visit. This table is sorted per visit ID. Once the table is ordered, the visits reports can be processed sequentially. As a precaution, checking procedures have been implemented to control null visits, number of reports per visit, total visits, etc XML Parsing We need to quickly search a collection of XML documents, and, to do this, we need to create an index of terms keeping track of the context in which these terms appear. Our solution uses Apache Lucene and the Java SAX library in order to create an index of Lucene Document objects for the lowest level of granularity to search. In order to have a testing framework as complete as possible, all reports elds have been processed. 2.3 ICD9 Mining The International Statistical Classication of Diseases and Related Health Problems (most commonly known by the abbreviation ICD) is a medical classication that provides codes to classify diseases and a wide variety of signs, symptoms, 2
3 abnormal ndings, complaints, social circumstances, and external causes of injury or disease. The Collection reports includes elds with discharge diagnosis in ICD9 codes form. In our approach, codes are processed to include a textual description to gather aditional information. Thus, when a ICD9 code is detected in report processing by SAX the code description is searched in a database and automaticaly added to the index. 3 Indexing Model main component is a search engine based on Apache Lucene. Lucene is a powerful Java library that lets you easily add document retrieval to any application. In recent years Lucene has become exceptionally popular and is now the most widely used information retrieval library. What we end up with after running Lucene is a directory named index, which contains les used by Lucene to associate terms with documents. To accomplish this, a Lucene index was created with a specic analyzer model-dependent. An Analyzer takes a series of terms or tokens and creates the terms to be indexed. A unique kind of Lucene index has been used for all developed models, or in other words, all models share the same Lucene index. Documents and elds are Lucene's fundamental units of indexing and searching. A document is Lucene's atomic unit of indexing and searching. It is a container that holds one or more elds, which in turn contain the real content. Each eld has a name to identify it, a text or binary value, and a series of detailed options that describe what Lucene should do with the eld value when you add the document to the index. To index our collection sources, we must rst translate it into Lucene's documents and elds. The indexing module takes from preprocessed collection all visit reports into a single Lucene document. Lucene allows duplicate elds to be added to a Document. This can make updating documents easy because it allows to add elds of various reports into a single Lucene document, with a mixing of all visit reports in a single document Lucene. At the end, we have the visit reports sharing the same Lucene document and the elds of the various reports sharing the same Lucene eld. 4 Retrieval approaches This section presents the dierent models developed for evaluation. Among the models developed four have been selected for submission: Baseline 2, Medical Face 3, BWRTDDD 4 and BWRTDDD2 5. From the four topics samples review, we consider using some alternative models to the nal runs. Although discard some models, we believe interesting to include a review of these unused models that can have interesting for scientic community. 2 UHU1BL TREC run ID. 3 UHU2MFB TREC run ID. 4 UHU3BWRTDDD TREC run ID. 5 UHU4BWRTDDD2 3
4 Figure 1: Prototype architecture. 4
5 The retrieval module completes the visits retrieval process. The subsequent processes of assessment and results extraction have been implemented according to the TREC eval utilities. The overall process can be seen in Figure Baseline This model has been designed to be the simplest approximation to the task. The Baseline model is based on the method of the bag of words. In this model, topics text is represented as an unordered collection of words, disregarding grammar and even word order. Therefore, the model matches the words in the topic with the words contained in the index. The usefulness of the model is twofold: provides a basis results for comparing and, on the other hand, its code serves as a basis for implementing more complex models. In order to maintain the simplicity, the baseline model makes match the topics words with the words contained in the eld report text. To develop this model we used a very simple analyzer. The analyzer takes only the topic and provides a set of terms to be searched in the index. Our base analyzer discards stops words with little semantic value, such as "the", "a", "an","for",... Cutting down on the number of terms indexed can save time and space in an index, but it can also limit accuracy. 4.2 Medical Face MedicalMiner initiative brings together three research groups from Huelva, Vigo and Madrid European Universities. MedicalMiner initiative proposes to analyse, experiment and develop new text and data mining techniques in an interrelated way, in intelligent medical information systems. An intelligent information access system based on them will be developed, oering advanced functionalities able to interrelate medical information, mainly information (text and data) from clinical records and scientic documentation, making use of standard resources of the domain (e.g. UMLS, SNOMED, Gene Ontology). An open source platform will be developed integrating all the elements [3]. A main component of this platform is Medical Face 6. Medical Face is a Biomedical text annotation system using multiple ontologies. In its simplest form, Medical Face takes a input text and makes medical entity recognition over this text. For our approach, Medical Face has been used to develop an analyzer that extracts relevant entities in the medical domain from the topics. This relevant entities are used to search for relevant documents in the generated index during indexing process. In this model, only the report text eld is taken account. 4.3 BWRTDDD The BWRTDDD model is based on the method of the bag of words in a similar way of the baseline model. Unlike the baseline model this model takes into 6 5
6 account the report text and the modied discharge diagnosis elds. The modied discharge diagnosis eld is a modied version of the eld that takes account a textual description of ICD9 codes generated during preprocessing process. In addition, a specic analyzer has been developed for processing the items. Like our base analyzer, the BWRTDDD analyzer discards common English words with little semantic value, such as "the", "a", "an","for",... Additionaly, some irrelevant terms are removed from the topic (patient, for example). 4.4 Exploring Age and Gender In the rst exploration of the sample topics, we think that would be interesting to identify genre and age in the topics and reports. It is easy to detect the occurrence of words that help classify subjects by patient gender and/or age (like woman or elderly). While gender can be detected directly in both the topics and reports, age can be estimated from some labels in the topics (young, adult, elderly,...) and reports (**AGE[in XXs]yearold). With this simple approach we can match those relevant visits to a topic according to age and gender. Unfortunately, this approach has been unsuccessful with the test topics by the small number of examples that show some of these features (only the adult, children, female and women tags have been found by this model). 4.5 Explore Negation Another way that looked promising was the negation detection. Usually, narrative reports in medical records contain a wealth of information that may augment structured data for managing patient information and predicting trends in diseases. Pertinent negatives are evident in text but are dicult to automatic processing. In our preliminary study of the sample topics, we saw that one of the examples contained a negation. Our research group have experience in the development of new text mining techniques adapted to bilingual medical domain. Including basic operations of text categorization and information extraction, capable of addressing specic problems of special relevance in the domain, such as the treatment of negation. In [2] we present a machine learning system that identify the negation and speculation signals in biomedical texts, in particular, in the BioScope corpus. Using the [2] approach, a prototype to handle negation was developed. This model takes a topic to detect negation presence. If the topic has a negative part then the medical entities in the negative clausule are detected. A visit search are performed deleting those visits with presence of these negative entities. Unfortunately, no topic with negative part was given, therefore, no model run was sent. 6
7 bpref R-prec Best Median Worst UHU1BL UHU2MFB UHU3BWRTDDD UHU4BWRTDDD Table 1: Medical Minner Medical Records results at TREC Table 2: Color data indicating results best than median. 7
8 Figure 2: Dierences in bpref between the submitted Medical-Miner runs and median results per topic. 5 Results This section shows the results for our four runs presented to the track (table 1). All our models results are around the evaluation median. In this exploratory track, we think that these results are a good beginning and encourage us for future developments. In despite of initial results, more analysis work needs to be developed. A deeply per topic study must be make. Due to the dierent topic types, a harder work should be performed to classify the behavior of each model depending on the topic type. About this idea, we have obtained some initial conclusions from the results in some topics. A rst improve to the overall system must be made in key concepts detection. At present, our research group are working in a open source platform for medical information retrieval. By now, this platform uses Medline and Freebase as main source of medical concepts detection. Extreme cases (see gure 2), such as topic 104, show that using Medline and Freebase is not enough to detect key medical concepts such as osteopenia. In a similar way, our systems have dicult to process expressions like "anti-coagulant medications post-op" (topic 128) or "underwent minimally invasive abdominal surgery" (topic 131). On the other hand, this approximation work ne with usual concepts like Hepatitis C, HIV or coronary stent (topics 118 and 125). The use of new ways of keyword extraction based on medical terminology databases, like UMLS [1] and SNOMED [5], are in our future plans. About our models, the best results were obtained with the Baseline model. Despite being the best model, there are no major dierences with other models. 8
9 When inspecting in detail the topics we have seen that an important number of them (14-15 depending on model) have better results with the alternative models (HUH2MFB, UHU3BWRTDDD, UHU4BWRTDDD2), better even track median (see table 2). This performance in certain topics suggests that a better t in model parameters can give better results if a larger number of sample topics are available. 6 Conclusions We have presented a novel runs of looking at the problem of medical text retrieval in the Medical Records Track framework. We feel that our models will provide eective retrieval in the order of track median. In our rst TREC participation, we feel that track goal has been met reasonably well by the runs proposed. In the near future, we will make a more detailed study about results. We hope to get more facts about models performance and alternative ways to improve the models. Apart from the send runs, new ways to improve the medical text retrieval have been proposed. Mainly we have shown our interest in detection of expressions of negation. We think that the detection of negation must be taken into account in the track future. Our research group has gained experience on development of applied retrieval information systems [4]. We think This rst prototype developed to TREC is feasible to help medical domain experts and medical professionals to carry out patient record searches in a more accurate and eciently way. References [1] O. Bodenreider. The unied medical language system (umls): integrating biomedical terminology. Nucleic acids research, 32(suppl 1):D267D270, [2] N.P. Cruz, M.J. Maña, and J. Mata. Aprendizaje automático versus expresiones regulares en la detección de la negación y la especulación en biomedicina. Procesamiento de Lenguaje Natural, 45, [3] M. de Buenaga, F. Fdez-Riverola, M. Maña, E. Puertas, D. Glez-Peña, and J. Mata. Medical-miner: Integración de conocimiento textual explícito en técnicas de minería de datos para la creación de herramientas traslacionales en medicina. Procesamiento de Lenguaje Natural, 45, [4] M. Millán, A. Muñoz, M. de la Villa, and M.J. Maña. A biomedical information retrieval system based on clustering for mobile devices. Procesamiento de Lenguaje Natural, 45,
10 [5] M.Q. Stearns, C. Price, K.A. Spackman, and A.Y. Wang. Snomed clinical terms: overview of the development process and project status. In Proceedings of the AMIA Symposium, page 662. American Medical Informatics Association,
LABERINTO at ImageCLEF 2011 Medical Image Retrieval Task
LABERINTO at ImageCLEF 2011 Medical Image Retrieval Task Jacinto Mata, Mariano Crespo, Manuel J. Maña Dpto. de Tecnologías de la Información. Universidad de Huelva Ctra. Huelva - Palos de la Frontera s/n.
Building a Spanish MMTx by using Automatic Translation and Biomedical Ontologies
Building a Spanish MMTx by using Automatic Translation and Biomedical Ontologies Francisco Carrero 1, José Carlos Cortizo 1,2, José María Gómez 3 1 Universidad Europea de Madrid, C/Tajo s/n, Villaviciosa
Ontology-based Archetype Interoperability and Management
Ontology-based Archetype Interoperability and Management Catalina Martínez-Costa, Marcos Menárguez-Tortosa, J. T. Fernández-Breis Departamento de Informática y Sistemas, Facultad de Informática Universidad
Managing large sound databases using Mpeg7
Max Jacob 1 1 Institut de Recherche et Coordination Acoustique/Musique (IRCAM), place Igor Stravinsky 1, 75003, Paris, France Correspondence should be addressed to Max Jacob ([email protected]) ABSTRACT
TMUNSW: Identification of disorders and normalization to SNOMED-CT terminology in unstructured clinical notes
TMUNSW: Identification of disorders and normalization to SNOMED-CT terminology in unstructured clinical notes Jitendra Jonnagaddala a,b,c Siaw-Teng Liaw *,a Pradeep Ray b Manish Kumar c School of Public
Towards a Visually Enhanced Medical Search Engine
Towards a Visually Enhanced Medical Search Engine Lavish Lalwani 1,2, Guido Zuccon 1, Mohamed Sharaf 2, Anthony Nguyen 1 1 The Australian e-health Research Centre, Brisbane, Queensland, Australia; 2 The
Find the signal in the noise
Find the signal in the noise Electronic Health Records: The challenge The adoption of Electronic Health Records (EHRs) in the USA is rapidly increasing, due to the Health Information Technology and Clinical
Web Database Integration
Web Database Integration Wei Liu School of Information Renmin University of China Beijing, 100872, China [email protected] Xiaofeng Meng School of Information Renmin University of China Beijing, 100872,
MIRACLE at VideoCLEF 2008: Classification of Multilingual Speech Transcripts
MIRACLE at VideoCLEF 2008: Classification of Multilingual Speech Transcripts Julio Villena-Román 1,3, Sara Lana-Serrano 2,3 1 Universidad Carlos III de Madrid 2 Universidad Politécnica de Madrid 3 DAEDALUS
1. Introduction MINING AND TRACKING EVOLVING WEB USER TRENDS FROM LARGE WEB SERVER LOGS. Basheer Hawwash and Olfa Nasraoui
MINING AND TRACKING EVOLVING WEB USER TRENDS FROM LARGE WEB SERVER LOGS Basheer Hawwash and Olfa Nasraoui Knowledge Discovery and Web Mining Lab Dept. of Computer Engineering and Computer Science University
Software Architecture Document
Software Architecture Document Natural Language Processing Cell Version 1.0 Natural Language Processing Cell Software Architecture Document Version 1.0 1 1. Table of Contents 1. Table of Contents... 2
Technical Report. The KNIME Text Processing Feature:
Technical Report The KNIME Text Processing Feature: An Introduction Dr. Killian Thiel Dr. Michael Berthold [email protected] [email protected] Copyright 2012 by KNIME.com AG
Exercises Engenharia de Software (cod. 5386 & 6633 )
Exercises Engenharia de Software (cod. 5386 & 6633 ) Departamento de Informática Universidade da Beira Interior Ano lectivo 2010/2011 These exercises are taken from Software Engineering, 9th edition, Pearson
Identify Disorders in Health Records using Conditional Random Fields and Metamap
Identify Disorders in Health Records using Conditional Random Fields and Metamap AEHRC at ShARe/CLEF 2013 ehealth Evaluation Lab Task 1 G. Zuccon 1, A. Holloway 1,2, B. Koopman 1,2, A. Nguyen 1 1 The Australian
Using LSI for Implementing Document Management Systems Turning unstructured data from a liability to an asset.
White Paper Using LSI for Implementing Document Management Systems Turning unstructured data from a liability to an asset. Using LSI for Implementing Document Management Systems By Mike Harrison, Director,
E-Assessment of Free-Text Answers Based on Domain Specic Sublanguages and Knowledge Representation
E-Assessment of Free-Text Answers Based on Domain Specic Sublanguages and Knowledge Representation Grado de Doctor por la Universidad de Vigo con Mención Internacional (PhD Degree by the University of
Transformation of Free-text Electronic Health Records for Efficient Information Retrieval and Support of Knowledge Discovery
Transformation of Free-text Electronic Health Records for Efficient Information Retrieval and Support of Knowledge Discovery Jan Paralic, Peter Smatana Technical University of Kosice, Slovakia Center for
Full-text Search in Intermediate Data Storage of FCART
Full-text Search in Intermediate Data Storage of FCART Alexey Neznanov, Andrey Parinov National Research University Higher School of Economics, 20 Myasnitskaya Ulitsa, Moscow, 101000, Russia [email protected],
BiTeM group report for TREC Medical Records Track 2011
BiTeM group report for TREC Medical Records Track 2011 J. Gobeill a, A. Gaudinat a, E. Pasche b, D. Teodoro b, D. Vishnyakova b, P. Ruch a a BiTeM group, University of Applied Sciences, Information Studies,
Introduction to Logistic Regression
OpenStax-CNX module: m42090 1 Introduction to Logistic Regression Dan Calderon This work is produced by OpenStax-CNX and licensed under the Creative Commons Attribution License 3.0 Abstract Gives introduction
Prediction of Heart Disease Using Naïve Bayes Algorithm
Prediction of Heart Disease Using Naïve Bayes Algorithm R.Karthiyayini 1, S.Chithaara 2 Assistant Professor, Department of computer Applications, Anna University, BIT campus, Tiruchirapalli, Tamilnadu,
Understanding Web personalization with Web Usage Mining and its Application: Recommender System
Understanding Web personalization with Web Usage Mining and its Application: Recommender System Manoj Swami 1, Prof. Manasi Kulkarni 2 1 M.Tech (Computer-NIMS), VJTI, Mumbai. 2 Department of Computer Technology,
Reducing Client Incidents through
Intel IT IT Best Practices Big Data Predictive Analytics December 2013 Reducing Client Incidents through Big Data Predictive Analytics Executive Overview Our new ability to proactively, rather than reactively,
131-1. Adding New Level in KDD to Make the Web Usage Mining More Efficient. Abstract. 1. Introduction [1]. 1/10
1/10 131-1 Adding New Level in KDD to Make the Web Usage Mining More Efficient Mohammad Ala a AL_Hamami PHD Student, Lecturer m_ah_1@yahoocom Soukaena Hassan Hashem PHD Student, Lecturer soukaena_hassan@yahoocom
An integrated. EHR system
An integrated t Expression Repository EHR system Daniel Karlsson, Mikael Nyström, Bengt Kron Project goals Develop and test a system for storing and querying pre- and post-coordinated SNOMED CT expressions
Semantic annotation of requirements for automatic UML class diagram generation
www.ijcsi.org 259 Semantic annotation of requirements for automatic UML class diagram generation Soumaya Amdouni 1, Wahiba Ben Abdessalem Karaa 2 and Sondes Bouabid 3 1 University of tunis High Institute
Clustering Technique in Data Mining for Text Documents
Clustering Technique in Data Mining for Text Documents Ms.J.Sathya Priya Assistant Professor Dept Of Information Technology. Velammal Engineering College. Chennai. Ms.S.Priyadharshini Assistant Professor
Curriculum Reform in Computing in Spain
Curriculum Reform in Computing in Spain Sergio Luján Mora Deparment of Software and Computing Systems Content Introduction Computing Disciplines i Computer Engineering Computer Science Information Systems
Search Result Optimization using Annotators
Search Result Optimization using Annotators Vishal A. Kamble 1, Amit B. Chougule 2 1 Department of Computer Science and Engineering, D Y Patil College of engineering, Kolhapur, Maharashtra, India 2 Professor,
Architecture of an Ontology-Based Domain- Specific Natural Language Question Answering System
Architecture of an Ontology-Based Domain- Specific Natural Language Question Answering System Athira P. M., Sreeja M. and P. C. Reghuraj Department of Computer Science and Engineering, Government Engineering
TRANSFoRm: Vision of a learning healthcare system
TRANSFoRm: Vision of a learning healthcare system Vasa Curcin, Imperial College London Theo Arvanitis, University of Birmingham Derek Corrigan, Royal College of Surgeons Ireland TRANSFoRm is partially
Exploration and Visualization of Post-Market Data
Exploration and Visualization of Post-Market Data Jianying Hu, PhD Joint work with David Gotz, Shahram Ebadollahi, Jimeng Sun, Fei Wang, Marianthi Markatou Healthcare Analytics Research IBM T.J. Watson
Natural Language Processing in the EHR Lifecycle
Insight Driven Health Natural Language Processing in the EHR Lifecycle Cecil O. Lynch, MD, MS [email protected] Health & Public Service Outline Medical Data Landscape Value Proposition of NLP
Volume 2, Issue 12, December 2014 International Journal of Advance Research in Computer Science and Management Studies
Volume 2, Issue 12, December 2014 International Journal of Advance Research in Computer Science and Management Studies Research Article / Survey Paper / Case Study Available online at: www.ijarcsms.com
Text Mining for Health Care and Medicine. Sophia Ananiadou Director National Centre for Text Mining www.nactem.ac.uk
Text Mining for Health Care and Medicine Sophia Ananiadou Director National Centre for Text Mining www.nactem.ac.uk The Need for Text Mining MEDLINE 2005: ~14M 2009: ~18M Overwhelming information in textual,
Clinical and research data integration: the i2b2 FSM experience
Clinical and research data integration: the i2b2 FSM experience Laboratory of Biomedical Informatics for Clinical Research Fondazione Salvatore Maugeri - FSM - Hospital, Pavia, italy Laboratory of Biomedical
Web Mining. Margherita Berardi LACAM. Dipartimento di Informatica Università degli Studi di Bari [email protected]
Web Mining Margherita Berardi LACAM Dipartimento di Informatica Università degli Studi di Bari [email protected] Bari, 24 Aprile 2003 Overview Introduction Knowledge discovery from text (Web Content
Natural Language to Relational Query by Using Parsing Compiler
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 4, Issue. 3, March 2015,
Search and Information Retrieval
Search and Information Retrieval Search on the Web 1 is a daily activity for many people throughout the world Search and communication are most popular uses of the computer Applications involving search
SPATIAL DATA CLASSIFICATION AND DATA MINING
, pp.-40-44. Available online at http://www. bioinfo. in/contents. php?id=42 SPATIAL DATA CLASSIFICATION AND DATA MINING RATHI J.B. * AND PATIL A.D. Department of Computer Science & Engineering, Jawaharlal
Does the Act of Refactoring Really Make Code Simpler? A Preliminary Study
Does the Act of Refactoring Really Make Code Simpler? A Preliminary Study Francisco Zigmund Sokol 1, Mauricio Finavaro Aniche 1, Marco Aurélio Gerosa 1 1 Department of Computer Science University of São
SIPAC. Signals and Data Identification, Processing, Analysis, and Classification
SIPAC Signals and Data Identification, Processing, Analysis, and Classification Framework for Mass Data Processing with Modules for Data Storage, Production and Configuration SIPAC key features SIPAC is
HELP DESK SYSTEMS. Using CaseBased Reasoning
HELP DESK SYSTEMS Using CaseBased Reasoning Topics Covered Today What is Help-Desk? Components of HelpDesk Systems Types Of HelpDesk Systems Used Need for CBR in HelpDesk Systems GE Helpdesk using ReMind
Travis Goodwin & Sanda Harabagiu
Automatic Generation of a Qualified Medical Knowledge Graph and its Usage for Retrieving Patient Cohorts from Electronic Medical Records Travis Goodwin & Sanda Harabagiu Human Language Technology Research
EHR Standards Landscape
EHR Standards Landscape Dr Dipak Kalra Centre for Health Informatics and Multiprofessional Education (CHIME) University College London [email protected] A trans-national ehealth Infostructure Wellness
An Automated Workflow System Geared Towards Consumer Goods and Services Companies
Proceedings of the 2014 International Conference on Industrial Engineering and Operations Management Bali, Indonesia, January 7 9, 2014 An Automated Workflow System Geared Towards Consumer Goods and Services
An Ontology Based Method to Solve Query Identifier Heterogeneity in Post- Genomic Clinical Trials
ehealth Beyond the Horizon Get IT There S.K. Andersen et al. (Eds.) IOS Press, 2008 2008 Organizing Committee of MIE 2008. All rights reserved. 3 An Ontology Based Method to Solve Query Identifier Heterogeneity
Bayesian Spam Filtering
Bayesian Spam Filtering Ahmed Obied Department of Computer Science University of Calgary [email protected] http://www.cpsc.ucalgary.ca/~amaobied Abstract. With the enormous amount of spam messages propagating
A Repository of Semantic Open EHR Archetypes
International Journal of Artificial Intelligence and Interactive Multimedia, Vol. 3, Nº 2. A Repository of Semantic Open EHR Archetypes Fernando Sánchez, Samuel Benavides, Fernando Moreno, Guillermo Garzón,
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
Prediction of Software Development Modication Eort Enhanced by a Genetic Algorithm
Prediction of Software Development Modication Eort Enhanced by a Genetic Algorithm Gerg Balogh, Ádám Zoltán Végh, and Árpád Beszédes Department of Software Engineering University of Szeged, Szeged, Hungary
Web 3.0 image search: a World First
Web 3.0 image search: a World First The digital age has provided a virtually free worldwide digital distribution infrastructure through the internet. Many areas of commerce, government and academia have
The Universe of Discourse Design with Visible Context
The Universe of Discourse Design with Visible Context Rational GUI Andrew U. Frank TU Wien, Department of Geoinformation Gusshausstrasse 27-29/E127.1 A-1040 Vienna, Austria [email protected] for
Combining SAWSDL, OWL DL and UDDI for Semantically Enhanced Web Service Discovery
Combining SAWSDL, OWL DL and UDDI for Semantically Enhanced Web Service Discovery Dimitrios Kourtesis, Iraklis Paraskakis SEERC South East European Research Centre, Greece Research centre of the University
PolyLens: Software for Map-based Visualization and Analysis of Genome-scale Polymorphism Data
PolyLens: Software for Map-based Visualization and Analysis of Genome-scale Polymorphism Data Ryhan Pathan Department of Electrical Engineering and Computer Science University of Tennessee Knoxville Knoxville,
Component visualization methods for large legacy software in C/C++
Annales Mathematicae et Informaticae 44 (2015) pp. 23 33 http://ami.ektf.hu Component visualization methods for large legacy software in C/C++ Máté Cserép a, Dániel Krupp b a Eötvös Loránd University [email protected]
Dr. Anuradha et al. / International Journal on Computer Science and Engineering (IJCSE)
HIDDEN WEB EXTRACTOR DYNAMIC WAY TO UNCOVER THE DEEP WEB DR. ANURADHA YMCA,CSE, YMCA University Faridabad, Haryana 121006,India [email protected] http://www.ymcaust.ac.in BABITA AHUJA MRCE, IT, MDU University
Software development process
OpenStax-CNX module: m14619 1 Software development process Trung Hung VO This work is produced by OpenStax-CNX and licensed under the Creative Commons Attribution License 2.0 Abstract A software development
Intelligent Log Analyzer. André Restivo <[email protected]>
Intelligent Log Analyzer André Restivo 9th January 2003 Abstract Server Administrators often have to analyze server logs to find if something is wrong with their machines.
BUSINESS RULES AND GAP ANALYSIS
Leading the Evolution WHITE PAPER BUSINESS RULES AND GAP ANALYSIS Discovery and management of business rules avoids business disruptions WHITE PAPER BUSINESS RULES AND GAP ANALYSIS Business Situation More
11-792 Software Engineering EMR Project Report
11-792 Software Engineering EMR Project Report Team Members Phani Gadde Anika Gupta Ting-Hao (Kenneth) Huang Chetan Thayur Suyoun Kim Vision Our aim is to build an intelligent system which is capable of
Database Marketing, Business Intelligence and Knowledge Discovery
Database Marketing, Business Intelligence and Knowledge Discovery Note: Using material from Tan / Steinbach / Kumar (2005) Introduction to Data Mining,, Addison Wesley; and Cios / Pedrycz / Swiniarski
The Lindsey-Fox Algorithm for Factoring Polynomials
OpenStax-CNX module: m15573 1 The Lindsey-Fox Algorithm for Factoring Polynomials C. Sidney Burrus This work is produced by OpenStax-CNX and licensed under the Creative Commons Attribution License 2.0
A Typing System for an Optimizing Multiple-Backend Tcl Compiler
The following paper was originally published in the Proceedings of the Fifth Annual Tcl/Tk Workshop Boston, Massachusetts, July 1997 A Typing System for an Optimizing Multiple-Backend Tcl Compiler Forest
White Paper Synoptic Reporting and Structured Data Capture
White Paper Synoptic Reporting and Structured Data Capture Business Case for Synoptic Reporting and Beyond Introduction The amount of clinically significant findings a surgical pathologist is expected
Data Warehousing and Data Mining in Business Applications
133 Data Warehousing and Data Mining in Business Applications Eesha Goel CSE Deptt. GZS-PTU Campus, Bathinda. Abstract Information technology is now required in all aspect of our lives that helps in business
Search and Real-Time Analytics on Big Data
Search and Real-Time Analytics on Big Data Sewook Wee, Ryan Tabora, Jason Rutherglen Accenture & Think Big Analytics Strata New York October, 2012 Big Data: data becomes your core asset. It realizes its
Social Media Mining. Data Mining Essentials
Introduction Data production rate has been increased dramatically (Big Data) and we are able store much more data than before E.g., purchase data, social media data, mobile phone data Businesses and customers
EHR Archetypes in practice: getting feedback from clinicians and the role of EuroRec
EuroRec - EHTEL Conference, Vienna, October 2007 EHR Archetypes in practice: getting feedback from clinicians and the role of EuroRec Dr Dipak Kalra Centre for Health Informatics and Multiprofessional
Legal Informatics Final Paper Submission Creating a Legal-Focused Search Engine I. BACKGROUND II. PROBLEM AND SOLUTION
Brian Lao - bjlao Karthik Jagadeesh - kjag Legal Informatics Final Paper Submission Creating a Legal-Focused Search Engine I. BACKGROUND There is a large need for improved access to legal help. For example,
Theodoros. N. Arvanitis, RT, DPhil, CEng, MIET, MIEEE, AMIA, FRSM
TRANSFoRm Theodoros. N. Arvanitis, RT, DPhil, CEng, MIET, MIEEE, AMIA, FRSM Biomedical Informatics, Signals & Systems Research Laboratory School of Electronic, Electrical & Computer Engineering College
Adaption of Statistical Email Filtering Techniques
Adaption of Statistical Email Filtering Techniques David Kohlbrenner IT.com Thomas Jefferson High School for Science and Technology January 25, 2007 Abstract With the rise of the levels of spam, new techniques
Data Warehouses in the Path from Databases to Archives
Data Warehouses in the Path from Databases to Archives Gabriel David FEUP / INESC-Porto This position paper describes a research idea submitted for funding at the Portuguese Research Agency. Introduction
PREDICTING STUDENTS PERFORMANCE USING ID3 AND C4.5 CLASSIFICATION ALGORITHMS
PREDICTING STUDENTS PERFORMANCE USING ID3 AND C4.5 CLASSIFICATION ALGORITHMS Kalpesh Adhatrao, Aditya Gaykar, Amiraj Dhawan, Rohit Jha and Vipul Honrao ABSTRACT Department of Computer Engineering, Fr.
