Open Access Research of Massive Spatiotemporal Data Mining Technology Based on Cloud Computing



Similar documents
UPS battery remote monitoring system in cloud computing

Cloud Computing based on the Hadoop Platform

Hadoop. MPDL-Frühstück 9. Dezember 2013 MPDL INTERN

Research on Clustering Analysis of Big Data Yuan Yuanming 1, 2, a, Wu Chanle 1, 2

Large-Scale Data Sets Clustering Based on MapReduce and Hadoop


Open Access The Cooperative Study Between the Hadoop Big Data Platform and the Traditional Data Warehouse

Hadoop IST 734 SS CHUNG

Networking in the Hadoop Cluster

Distributed Framework for Data Mining As a Service on Private Cloud

Manifest for Big Data Pig, Hive & Jaql

SPATIAL DATA CLASSIFICATION AND DATA MINING

Applied research on data mining platform for weather forecast based on cloud storage

Big Data with Rough Set Using Map- Reduce

Big Data Storage Architecture Design in Cloud Computing

A STUDY ON HADOOP ARCHITECTURE FOR BIG DATA ANALYTICS

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY

A Novel Cloud Computing Data Fragmentation Service Design for Distributed Systems

Final Project Proposal. CSCI.6500 Distributed Computing over the Internet

Journal of Chemical and Pharmaceutical Research, 2015, 7(3): Research Article. E-commerce recommendation system on cloud computing

Hosting Transaction Based Applications on Cloud

Oracle Big Data SQL Technical Update

Analysis and Research of Cloud Computing System to Comparison of Several Cloud Computing Platforms

Big Data Explained. An introduction to Big Data Science.

Introduction to Hadoop HDFS and Ecosystems. Slides credits: Cloudera Academic Partners Program & Prof. De Liu, MSBA 6330 Harvesting Big Data

Open source software framework designed for storage and processing of large scale data on clusters of commodity hardware

Big Data and Hadoop with components like Flume, Pig, Hive and Jaql

The Construction of Seismic and Geological Studies' Cloud Platform Using Desktop Cloud Visualization Technology

Role of Cloud Computing in Big Data Analytics Using MapReduce Component of Hadoop

Massive Cloud Auditing using Data Mining on Hadoop

DATA SECURITY MODEL FOR CLOUD COMPUTING

Converged, Real-time Analytics Enabling Faster Decision Making and New Business Opportunities

Design of Electric Energy Acquisition System on Hadoop

Cloud Security in Map/Reduce An Analysis July 31, Jason Schlesinger

Mobile Storage and Search Engine of Information Oriented to Food Cloud

CSE 590: Special Topics Course ( Supercomputing ) Lecture 10 ( MapReduce& Hadoop)

White Paper: What You Need To Know About Hadoop

The WAMS Power Data Processing based on Hadoop

W H I T E P A P E R. Deriving Intelligence from Large Data Using Hadoop and Applying Analytics. Abstract

IMPROVED FAIR SCHEDULING ALGORITHM FOR TASKTRACKER IN HADOOP MAP-REDUCE

CIS 4930/6930 Spring 2014 Introduction to Data Science Data Intensive Computing. University of Florida, CISE Department Prof.

Data Refinery with Big Data Aspects

Accelerating and Simplifying Apache

A Professional Big Data Master s Program to train Computational Specialists

Application Development. A Paradigm Shift

Hadoop. Sunday, November 25, 12

Research on Job Scheduling Algorithm in Hadoop

Big Data and Market Surveillance. April 28, 2014

Open source Google-style large scale data analysis with Hadoop

BIG DATA What it is and how to use?

Research on IT Architecture of Heterogeneous Big Data

Cloud computing - Architecting in the cloud

Big Data and Hadoop with Components like Flume, Pig, Hive and Jaql

Snapshots in Hadoop Distributed File System

Query and Analysis of Data on Electric Consumption Based on Hadoop

Big Data and Apache Hadoop s MapReduce

Application and practice of parallel cloud computing in ISP. Guangzhou Institute of China Telecom Zhilan Huang

Lecture 32 Big Data. 1. Big Data problem 2. Why the excitement about big data 3. What is MapReduce 4. What is Hadoop 5. Get started with Hadoop

Data processing goes big

Recognization of Satellite Images of Large Scale Data Based On Map- Reduce Framework

Hadoop: A Framework for Data- Intensive Distributed Computing. CS561-Spring 2012 WPI, Mohamed Y. Eltabakh

Cloud Storage Solution for WSN Based on Internet Innovation Union

A Big Data-driven Model for the Optimization of Healthcare Processes

International Journal of Engineering Research ISSN: & Management Technology November-2015 Volume 2, Issue-6

Managing Cloud Server with Big Data for Small, Medium Enterprises: Issues and Challenges

Hadoop and Map-Reduce. Swati Gore

ICT Perspectives on Big Data: Well Sorted Materials

BIG DATA WEB ORGINATED TECHNOLOGY MEETS TELEVISION BHAVAN GHANDI, ADVANCED RESEARCH ENGINEER SANJEEV MISHRA, DISTINGUISHED ADVANCED RESEARCH ENGINEER

Problem Solving Hands-on Labware for Teaching Big Data Cybersecurity Analysis

White Paper. Big Data and Hadoop. Abhishek S, Java COE. Cloud Computing Mobile DW-BI-Analytics Microsoft Oracle ERP Java SAP ERP

Exploration on Security System Structure of Smart Campus Based on Cloud Computing. Wei Zhou

Open Access Research of Fast Search Algorithm Based on Hadoop Cloud Platform

How To Handle Big Data With A Data Scientist

Open Access Research on Database Massive Data Processing and Mining Method based on Hadoop Cloud Platform

DATA MINING WITH HADOOP AND HIVE Introduction to Architecture

International Journal of Innovative Research in Computer and Communication Engineering

BIG DATA CHALLENGES AND PERSPECTIVES

How In-Memory Data Grids Can Analyze Fast-Changing Data in Real Time

Transforming the Telecoms Business using Big Data and Analytics

Chapter 7. Using Hadoop Cluster and MapReduce

Research of Railway Wagon Flow Forecast System Based on Hadoop-Hazelcast

Luncheon Webinar Series May 13, 2013

ISSN: CONTEXTUAL ADVERTISEMENT MINING BASED ON BIG DATA ANALYTICS

What is Analytic Infrastructure and Why Should You Care?

BIG DATA IN THE CLOUD : CHALLENGES AND OPPORTUNITIES MARY- JANE SULE & PROF. MAOZHEN LI BRUNEL UNIVERSITY, LONDON

An Oracle White Paper November Leveraging Massively Parallel Processing in an Oracle Environment for Big Data Analytics

Search and Real-Time Analytics on Big Data

Research on Operation Management under the Environment of Cloud Computing Data Center

Analysis of Information Management and Scheduling Technology in Hadoop

Modeling for Web-based Image Processing and JImaging System Implemented Using Medium Model

Federated Cloud-based Big Data Platform in Telecommunications

R.K.Uskenbayeva 1, А.А. Kuandykov 2, Zh.B.Kalpeyeva 3, D.K.Kozhamzharova 4, N.K.Mukhazhanov 5

A REVIEW PAPER ON THE HADOOP DISTRIBUTED FILE SYSTEM

Implement Hadoop jobs to extract business value from large and varied data sets

Distributed file system in cloud based on load rebalancing algorithm

WHAT S NEW IN SAS 9.4

Survey of the Benchmark Systems and Testing Frameworks For Tachyon-Perf

Study on Cloud Computing Resource Scheduling Strategy Based on the Ant Colony Optimization Algorithm

Sentimental Analysis using Hadoop Phase 2: Week 2

Transcription:

Send Orders for Reprints to reprints@benthamscience.ae 2244 The Open Automation and Control Systems Journal, 2015, 7, 2244-2252 Open Access Research of Massive Spatiotemporal Data Mining Technology Based on Cloud Computing Shuxia Wang * and Zenggang Xiong School of Computer and Information Science; Hubei Engineering University, Hubei, Xiaogan, 432000, China Abstract: This paper mainly discusses developing a new mass data analysis system in face of the above challenges, based on cloud computing technology, cloud service provided by Amazon and HBase. Through the research of the EC2, S3 and EMR service provided by Amazon and a detailed analysis of the client, reception server and data handling server of the data analysis system, this paper tries to create a preliminary model of the system. Then it tries to finish each model and whole system according to actual demand. Finally, a system test and performance test proves that the system can solve the problems faced by traditional data analysis system or not. The research runs through the whole process of data analytics system design and implementation. Firstly, present the problem domain and requirements. We through investigate the traditional data analysis system and analyses the problems existing in this system, then raise the developing requirements of the massive data analysis system based on cloud computing, and clear the key points of the new system implementation are receiving server and data processing design. Secondly, design and realize each module of the massive data analysis system. Finally, execute the system test and performance test on this system, the test results proved that the new data analysis system is stable and can supply the requirements. Keywords: Big data, massive spatiotemporal data, data mining, cloud computing. 1. INTRODUCTION With the development of information technology, data analysis is playing an increasing role in business decision and product design. Based on analysis of user behavior data, product designers can design functions with an enhanced user experience and can improve functions of other parts. Decision makers of a company can use the result of data analysis to decide product orientation and the company's developing direction. However, the development of the company, the increase of products and users and other factors lead to a huge increase of the total quantity of data. Facing the challenge of data saving and handling capacity, traditional data analysis system cannot able to accomplish its mission [1]. With the dramatic increase in the amount of data storage, massive data processing and massive data computing have become an important issue in the field of data mining. The traditional serial data mining algorithms are often only able to handle small-scale data, when faced with massive data, their execution speed will be reduced or even not run, so it poses severe challenges and Ordeal for the current data mining. And classification algorithms, as one of the most important part of the data mining, which plays an important role in information retrieval, Web search and CRM. Currently the vast majority of classification algorithms are serial, in the feasibility poor and inefficient when dealing with large *Address correspondence to these authors at the School of Computer and Information Science; Hubei Engineering University, Hubei, Xiaogan, 432000, China; Tel: +86-13407874545; E-mail: shuxia@163.com sets of data, the problem of low classification accuracy have become increasingly prominent, leading to immeasurable computing resources and unlimited extension of implementation time. 2. RELATED THEORY OF DATA MINING AND CLOUD COMPUTING Today's society are dealing with massive data, before the advent of cloud computing, in the past when making data mining always look forward to using high-performance machine or even a large-scale computing device to process; and also in the context of massive data, among mining process need to have a good of development environment and application environment, in such circumstances [2], the use of cloud-based computing approach to data mining is more appropriate. And due to the lack of parallel sorting algorithms currently, large-scale data set increasingly large day by day, traditional data mining system can no longer be effective on these massive data mining and use, how to improve the efficiency and parallelism of the algorithm is a problem which needs to be solved urgently. This thesis is based on an integrated system of comprehensive information security program of laboratory, analysis and research massive data mining-related technologies which were applied in this program. As the data which the public opinion system processes are from the Internet, the amount of data each day to deal with is very large, to training and classifying this massive data set, ensure that the public opinion analysis system can be maintained at a stable and efficient. How to improve the efficiency and performance of 1874-4443/15 2015 Bentham Open

Research of Massive Spatiotemporal Data Mining Technology The Open Automation and Control Systems Journal, 2015, Volume 7 2245 Fig. (1). Architecture of Hadoop. Fig. (2). Hadoop module design. classification of the public opinion analysis system is the main problem of this thesis as shown in Fig. (1). The advanced nature of this thesis is that classification algorithm plays an important role in the public opinion analysis system, according to the public opinion analysis system requirements and system design, a classification algorithm module is designed based on the Strategy pattern for the public opinion analysis system. And design different parallel classification algorithms, through packed classification algorithms under MapReduce framework [3], greatly improving the efficiency of classification algorithms, making the classification algorithms speedup are close to linear speedup. Through this module, the public opinion analysis system can dynamically call different classification algorithms to classify public opinion data, improving the performance and efficiency of classification algorithms of the system, thus greatly improving the stability and reliability of public opinion analysis system as shown in Fig. (2). Cloud computing is a business computing model, it assigns the computing tasks to a large number of computers in the resource pool, it can provide users with computing power, storage capacity and application service capabilities according to their needs; Cloud computing provides cheap and efficient solutions of storing and analyzing mass data [4]. Data mining is the process of discovering information or patterns that are interesting, non-trivial, implicit, previously unknown and potentially useful in large databases. Data mining plays a guiding role on scientific research, business decisions and other fields, with far-reaching social and economic significance. Data mining need to use huge computing and storage resource, so integrate cloud computing and data mining can effectively control computing cost and enhance the efficiency of data mining, breaking the bottleneck of the traditional limitations of data mining. It is very important to research the data mining strategies based on cloud computing from the theoretical view and practical view.

2246 The Open Automation and Control Systems Journal, 2015, Volume 7 Wang and Xiong Fig. (3). MapReduce data stream. Fig. (4). Hive system architecture. Hadoop is the most famous open source distributed computing framework [5], through the use of MapReduce parallel model to effective integration of the computing storage capacity in order to provide a powerful distributed computing capabilities. This paper mainly focuses on the Fig. (3). The development of network techniques brings people in a great deal of information, it also greatly increase the difficulty to find useful knowledge from mass data. The efforts to solve this problem promote the emergence and rapid development of data mining techniques. At present, the data mining technologies and tools have been used in the financial, medical, military, and many other areas of commercial decision-making analysis as shown in Fig. (4). Cloud computing is a style of computing in which dynamically scalable and often virtualized resources are provided as a service over the Internet. Cloud computing platform can be used to development high capability programmers. However, it does not provide data reduction service

Research of Massive Spatiotemporal Data Mining Technology The Open Automation and Control Systems Journal, 2015, Volume 7 2247 Fig. (5). Architecture of data analysis system. which is one of the bases of data-mining. So the method to implement data-mining system on cloud computing platform has not been worked out yet. To solve these problems, in this paper, a data reduction module for accessing isomeric and new type data is designed and implemented to expand Google App Engine cloud platform, further more, a new data-mining system built on top of cloud computing services is designed and implemented to verify the validity of data reduction module and the efficiency of cloud based data-mining process. Data reduction module supports uniform definition of data sets by using meta-data definition, data set definition and data set instance definition to abstract data type, data structure, and location information and so on. Layered thinking model is used and a new motile layered plug in architecture is induced to enhance the scalability of the system. Meanwhile, many important thoughts and methods to design and implement the platform are induced in this paper and a prototype of data-mining platform based on this architecture is introduced at last section of this paper. Experimental results show that the proposed method is very promising. 3. MASSIVE SPATIONTEMPORAL DATA MINING TECHNOLOGY Lots of implicit knowledge and information are exist in GML spatiotemporal data, including spatiotemporal model and its characteristic, the general relationship between spatiotemporal data and common data, and the general data characteristic exists in GML spatiotemporal data and so on. Human can learn about the natural world by means of getting relation, discipline and interaction existing in nature, also these are used to making decision for our production and life. However, due to the GML temporal-spatial and half structural characteristic, it can't use accurate mode to define GML spatiotemporal data. So it is much more complicated to extract information from GML spatiotemporal data than traditional data. At the same time, because of characteristics of quantity and the various computing intensive, [6, 7] these limit the information processing progress some degree.to resolve all of these problems, this paper put forward two parallel clustering mining algorithm in parallel computing Hadoop environment. In this paper the author design a parallel clustering GML mining prototype system, and At last, showing the clustering visualizes to the form of map. This paper gives two kinds of parallel clustering algorithm for mining GML spatiotemporal series data. The first kind is put forward with K-means clustering mining algorithm based on time sequence of GML spatiotemporal similarity, Considering the spatial attribute and temporal series together to metric the similarity between the two spatial adjacent objects. Then through the K-means clustering algorithm conduct the mining. The second is according to the definition of spatial neighborhood to get the GML spatial neighborhood in GML spatiotemporal objects, and then in this neighborhood mature the two objects of temporal series of the attribute based on temporal sequence to getting the similarity, combining parallel DBSCAN (STN-DBSCAN) clustering algorithm for data mining. We can see Fig. (5).

2248 The Open Automation and Control Systems Journal, 2015, Volume 7 Wang and Xiong Fig. (6). Architecture of AP2.0 system. Through constructing Hadoop distributed parallel computing platform, using MapReduce programming model to realize the two kind of clustering algorithm means and DBSCAN parallel process. By designing and implementing a parallel GML spatiotemporal clustering prototype mining system, can running the two parallel clustering algorithm using GML format attributes of the meteorological data. Through the experimental results we verify effectiveness of the two parallel algorithm, and quality of high performance. Good performance of the algorithm can be expanded. Finally, we show the results of cluster income in the form of map. 4. DATA ANALYSIS With the use of computer and Internet technology expansion of many aspects o human society, data shows explosive growth. Now, the process and storage of large data sets has become the new challenge of enterprises. How to mine valuable and understandable knowledge from the massive data in a more rapid, efficient, low-cost way, to help company make decisions is a challenges of Data mining. The emergence of Cloud computing bring new opportunities for data mining technology. Cloud computing distribute the ability of storage and computing among multiple nodes in cloud cluster. So it enabling huge data set storage and computing power. Because you can use a lot of cheap computers for cluster instead of by the high price of the server, cloud computing greatly reduced costs. Together with Cloud computing technology which can provide the large storage capacity and computing power, data mining technology go into the cloudbased data mining era. HADOOP is an open source project of Apache for building cloud platform. HADOOP framework will help us implementation of clusters easier, faster and more effective. HADOOP using HDFS (distributed file system) to achieve large file storage and fault tolerance, and the use MapReduce programming model to computing. To make HADOOP applied to data mining, a key question is how to parallel the traditional data mining algorithms. Some specified traditional data mining algorithm can be paralleled easier because of its own characteristics. But some algorithms are hard to be paralleled. For the algorithms that can be paralleled, combined with MapReduce programming model, we can transplant it to HADOOP platform. Then it will complete data mining tasks more efficient in parallel way. We can see the Fig. (6). Cloud computing is a hot topic at home and abroad, it is the development of current high performance computational model, it is a numerical model of virtualization through the network to provide services dynamically scalable resources. Through the cloud computing, people can get dynamically

Research of Massive Spatiotemporal Data Mining Technology The Open Automation and Control Systems Journal, 2015, Volume 7 2249 Fig. (7). Json string of image_size event definition. Fig. (8). Types of amazon EC2 instance. extensible computing and storage capacity on the network. Cloud computing can improve the data processing efficiency, while reducing the terminal equipment requirements; it can effectively solve the problems which mass data processing faced. Therefore, the topic of cloud computing who based on the distributed data mining platform is a hot research. This paper is based on the practice of the key breakthrough project, this paper analyzes and studies the technology of cloud computing and data mining, it focused on the DBSCAN clustering algorithm based on density. Aiming at the shortcomings of DBSCAN clustering algorithm and combined with the project of charging station data characteristics, this paper proposed a new algorithm. This algorithm is the DBSCAN clustering algorithm based on grid control factor, which is used the fixed grid size DBSCAN algorithm in the project as the foundation. In order to find a better clustering accuracy grid size a grid control factor value is to adjust the size of the grid. The paper has been proved that it has the improved clustering accuracy by the test of charging station data, is also effectively reduce the time complexity as shown in Fig. (7-10). Second important problems to be solved in this paper is to do the parallel processing for the improved algorithm, and then realize it on the cloud computing platform. To carry out the clustering analysis of massive data sets, we must ensure that the system can be maintained at a stable, efficient environment. The paper designs a parallel algorithm based on Hadoop, built a simple Hadoop environment, through the

2250 The Open Automation and Control Systems Journal, 2015, Volume 7 Wang and Xiong Fig. (9). Configuration of aggregation script. Fig. (10). Script of table creation. encapsulation of the DBSCAN clustering algorithm in the framework of MapReduce, it greatly improving the efficiency of this algorithm. Finally, I verified the improved algorithm based on cloud computing by using replication largescale charging station data, the experimental results show that, the DBSCAN algorithm based on cloud computing greatly improved the processing efficiency of large data sets, in the condition of not reducing the DBSCAN clustering quality. To help cloud computing take root, it will be necessary to adapt various mature technologies to the cloud computing paradigm. I list some of them below. As cloud computing software platform is the heart of a cloud computing system, it will require considerable further research. Hadoop is an open source cloud computing software platform, as an alternative to the platforms developed by Google and others. It appears to be a good vehicle as a launching point for research. Yahoo is a major sponsor of the Hadoop project. IBM has adopted Hadoop for its Blue Cloud solution. Facebook uses Hadoop in its data analysis. Google, IBM, and Yahoo have donated cloud computing platforms to 6 US universities. The computing platforms all include Hadoop. Hadoop may become the Linux of cloud computing. Such means of collaboration as chat, instant messaging; Internet phone calling, etc. will be added to various popular applications. Google Docs spreadsheets already make it possible for multiple users to chat while editing a spreadsheet together.

Research of Massive Spatiotemporal Data Mining Technology The Open Automation and Control Systems Journal, 2015, Volume 7 2251 Fig. (11). Data activity analytics. The research on these subjects can leverage the available EAI, EII, and ESB technologies. Transmitting the bulky multimedia data across the network will continue to be a challenge, and it needs further research to speed up cloud computing. Further, as more data gets pushed to the clouds, including user-created data, the need to analyze (mine) such data to derive business-useful knowledge will increase. The data mining and machine learning communities will need to address this need. As the clouds proliferate and the users start plugging into multiple clouds, the problems of discovering and composing services that have been subjects of research in the serviceoriented architecture context will need to be revisited in the cloud computing context. Data activity analytics are shown in Fig. (11). While in many cases more research is needed to make these cryptographic tools sufficiently practical for the cloud, we believe they present the best opportunity for a clear differentiator for cloud computing since these protocols can enable cloud users to benefit from one another s data in a controlled manner. In particular, even encrypted data can enable anomaly detection that is valuable from a business intelligence standpoint. For example, a cloud payroll service might provide, with the agreement of participants, aggregate data about payroll execution time that allows users to identify inefficiencies in their own processes. Taking the vision even further, if the cloud service provider is empowered with some ability to search the encrypted data, the proliferation of cloud data can potentially enable better insider threat detection (e.g. by detecting user activities outside of the norm) and better data loss prevention (DLP) (e.g. through detecting anomalous content). CONCLUSION Currently the cloud computing and data analytics system are in the booming development period. This research runs through research the cloud computing and Amazon could Web services, and used them in enterprise data analysis system implementation. We finished the data analysis system based on could computing, so that the problems which exist in traditional data analysis system have been resolved, and aroused data analyzing and handling capacity of the system, saved the cost, and cut down development cycle. After using cloud computing technology, data analysis system is more flexible in data processing and analysis; it can increase or reduce the computing resource according the actual requirement. Also the spending of equipment maintenance and data backup will be saved. Both from the data processing system performance and cost of the project, using cloud computing technology will be the best solution. In this study, the actual project successfully confirmed and embodies the advantages of cloud computing technology. In the research and application of cloud computing technology, this paper has some certain reference value. CONFLICT OF INTEREST The authors confirm that this article content has no conflict of interest. ACKNOWLEDGEMENTS Declared none.

2252 The Open Automation and Control Systems Journal, 2015, Volume 7 Wang and Xiong REFERENCES [1] J. Dean, and S. Ghemawat, MapReduce: Simplified data processing on large clusters, In: Proc. 6 th Symp. Oper. Syst. Des. Implement., Berkeley: USENIX Association, pp. 137-150, 2014. [2] F. Chang, J. Dean, S. Ghemawat, W.C. Hsieh, D.A. Wallach, M. Burrows, T. Chandra, A. Fikes, R.E. Gruber, Bigtable: A distributed storage system for structured data, In: Proc. 7 th USENIX Symp. Oper. Syst. Design Implement., Berkeley: USENIX A ssociation, pp. 205-218, 2006. [3] Welcome to Hadoop MapReduce [EB/OL]. http://hadoop.apache.org/mapreduce/ [4] Google Docs Glitch Exposes Private Files. http://www.pcworld.com/article/160927/google_docs_glitch_expos es_private_files.html. [5] IT Cloud Services User Survey, pt.2: Top Benefits & Challenges. http://blogs.idc.com/ie/?p=210. [6] D. Boneh, and B. Waters, Conjunctive, subset, and range queries on encrypted data, In The 4 th Theor. Cryptogr. Conf. (TCC 2007), 2007. [7] Lithuania Weathers Cyber Attack, Braces for Round 2. http://blog.washingtonpost.com/securityfix/2008/07/lithuania_weat hers_cyber_attac_1.html. Received: June 16, 2015 Revised: August 10, 2015 Accepted: September 19, 2015 Wang and Xiong; Licensee Bentham Open. This is an open access article licensed under the terms of the (https://creativecommons.org/licenses/by/4.0/legalcode), which permits unrestricted, non-commercial use, distribution and reproduction in any medium, provided the work is properly cited.