A COMPARITIVE AND ANALYTICAL STUDY ON BIG DATA



Similar documents
INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY

Chapter 7. Using Hadoop Cluster and MapReduce

Data Refinery with Big Data Aspects

A Survey on Challenges and Advantages in Big Data

Managing Cloud Server with Big Data for Small, Medium Enterprises: Issues and Challenges

Big Data: Tools and Technologies in Big Data

Statistics for BIG data

Manifest for Big Data Pig, Hive & Jaql

Managing Big Data with Hadoop & Vertica. A look at integration between the Cloudera distribution for Hadoop and the Vertica Analytic Database

Are You Ready for Big Data?

ISSN: CONTEXTUAL ADVERTISEMENT MINING BASED ON BIG DATA ANALYTICS

Volume 3, Issue 6, June 2015 International Journal of Advance Research in Computer Science and Management Studies

A REVIEW PAPER ON THE HADOOP DISTRIBUTED FILE SYSTEM

Hadoop Big Data for Processing Data and Performing Workload

Hadoop. MPDL-Frühstück 9. Dezember 2013 MPDL INTERN

Big Data Analytics OverOnline Transactional Data Set

AN EFFICIENT SELECTIVE DATA MINING ALGORITHM FOR BIG DATA ANALYTICS THROUGH HADOOP

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY

Big Data / FDAAWARE. Rafi Maslaton President, cresults the maker of Smart-QC/QA/QD & FDAAWARE 30-SEP-2015

Keywords: Big Data, Hadoop, cluster, heterogeneous, HDFS, MapReduce

BIG DATA TRENDS AND TECHNOLOGIES

Big Data. White Paper. Big Data Executive Overview WP-BD Jafar Shunnar & Dan Raver. Page 1 Last Updated

Big Data and Market Surveillance. April 28, 2014

BIG DATA & ANALYTICS. Transforming the business and driving revenue through big data and analytics

The 4 Pillars of Technosoft s Big Data Practice

How to use Big Data in Industry 4.0 implementations. LAURI ILISON, PhD Head of Big Data and Machine Learning

Sunnie Chung. Cleveland State University

ISSN: International Journal of Innovative Research in Technology & Science(IJIRTS)

Lecture 32 Big Data. 1. Big Data problem 2. Why the excitement about big data 3. What is MapReduce 4. What is Hadoop 5. Get started with Hadoop

BIG DATA What it is and how to use?

Introduction to Hadoop HDFS and Ecosystems. Slides credits: Cloudera Academic Partners Program & Prof. De Liu, MSBA 6330 Harvesting Big Data

BIG DATA CHALLENGES AND PERSPECTIVES

Keywords: Big Data, HDFS, Map Reduce, Hadoop

How To Handle Big Data With A Data Scientist

Big Data and Data Science: Behind the Buzz Words

Ubuntu and Hadoop: the perfect match

Keywords Big Data; OODBMS; RDBMS; hadoop; EDM; learning analytics, data abundance.

Hadoop. Sunday, November 25, 12

Foundations of Business Intelligence: Databases and Information Management

Big Data on Cloud Computing- Security Issues

International Journal of Advanced Engineering Research and Applications (IJAERA) ISSN: Vol. 1, Issue 6, October Big Data and Hadoop

Big Data: Study in Structured and Unstructured Data

Chapter 6. Foundations of Business Intelligence: Databases and Information Management

SAP Solution Brief SAP HANA. Transform Your Future with Better Business Insight Using Predictive Analytics

Developing Scalable Smart Grid Infrastructure to Enable Secure Transmission System Control

Testing Big data is one of the biggest

Alexander Nikov. 5. Database Systems and Managing Data Resources. Learning Objectives. RR Donnelley Tries to Master Its Data

Open source software framework designed for storage and processing of large scale data on clusters of commodity hardware

International Journal of Advancements in Research & Technology, Volume 3, Issue 5, May ISSN BIG DATA: A New Technology

How Big Is Big Data Adoption? Survey Results. Survey Results Big Data Company Strategy... 6

Big Data With Hadoop

Hadoop Technology for Flow Analysis of the Internet Traffic

Are You Ready for Big Data?

Big Data Analysis: Apache Storm Perspective

A Brief Introduction to Apache Tez

Keywords Big Data, NoSQL, Relational Databases, Decision Making using Big Data, Hadoop

White Paper. Big Data and Hadoop. Abhishek S, Java COE. Cloud Computing Mobile DW-BI-Analytics Microsoft Oracle ERP Java SAP ERP

marlabs driving digital agility WHITEPAPER Big Data and Hadoop

Infomatics. Big-Data and Hadoop Developer Training with Oracle WDP

Data Warehouse design

Chapter 6 8/12/2015. Foundations of Business Intelligence: Databases and Information Management. Problem:

Data, Measurements, Features

A STUDY ON HADOOP ARCHITECTURE FOR BIG DATA ANALYTICS

International Journal of Innovative Research in Computer and Communication Engineering

Associate Professor, Department of CSE, Shri Vishnu Engineering College for Women, Andhra Pradesh, India 2

A Review of Data Mining Techniques

Big Data Explained. An introduction to Big Data Science.

International Journal of Computer Science Trends and Technology (IJCST) Volume 2 Issue 3, May-Jun 2014

Prepared By : Manoj Kumar Joshi & Vikas Sawhney

Application Development. A Paradigm Shift

Hadoop Submitted in partial fulfillment of the requirement for the award of degree of Bachelor of Technology in Computer Science

Bringing Big Data into the Enterprise

Large scale processing using Hadoop. Ján Vaňo

Data Isn't Everything

Getting to Know Big Data

Big Data. Fast Forward. Putting data to productive use

Big Data: Issues, Challenges, Tools and Good Practices

Data processing goes big

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY

Introduction to Hadoop. New York Oracle User Group Vikas Sawhney

Internals of Hadoop Application Framework and Distributed File System

Big Data, Why All the Buzz? (Abridged) Anita Luthra, February 20, 2014

DATA MINING TECHNIQUES AND APPLICATIONS

Architectures for Big Data Analytics A database perspective

Problems to store, transfer and process the Big Data 6/2/2016 GIANG TRAN - TTTGIANG2510@GMAIL.COM 1

Packet Flow Analysis and Congestion Control of Big Data by Hadoop

Navigating Big Data business analytics

How To Scale Out Of A Nosql Database

Learning is a very general term denoting the way in which agents:

An Overview of Knowledge Discovery Database and Data mining Techniques

Architecting for Big Data Analytics and Beyond: A New Framework for Business Intelligence and Data Warehousing

Hadoop Ecosystem B Y R A H I M A.

Big data and its transformational effects

An Oracle White Paper November Leveraging Massively Parallel Processing in an Oracle Environment for Big Data Analytics

Chapter 11 Map-Reduce, Hadoop, HDFS, Hbase, MongoDB, Apache HIVE, and Related

Statistical Challenges with Big Data in Management Science

Keywords Stock Exchange Market, Clustering, Hadoop, Map-Reduce

Transforming the Telecoms Business using Big Data and Analytics

Suresh Lakavath csir urdip Pune, India

Transcription:

A COMPARITIVE AND ANALYTICAL STUDY ON BIG DATA Rishav Shaw School of Computing Science and Engineering VIT University Vellore 632014 ABSTRACT:- In this paper, we are going to review and present the existing issues and challenges of Big Data, along with their proper solutions and algorithms in collaboration with the upcoming ones. These solutions can help us in handling various big data s i.e. firms like Google, ebay, LinkedIn, and Facebook were built around big data from the beginning. So like many new information technologies, big data can bring about dramatic cost reductions, substantial improvements in the time required to perform a computing task, or new product and service offerings. Keywords:- Big Data, Issues, Challenges etc. 1. INTRODUCTION:- Big data is a large amount of data which cannot be analysed or processed by conventional methods or technology. It requires new technology and structures to deal with and extract value from it by analytical process. Recently, in today s fast moving economic and social world the amount of data has increased exponentially. This explosion of available data has been caused due to increasing use of data extensive softwares. Some sources of big data include traffic management and tracking of personal devices such as smartphones. Big data analysis technologies being developed can bring huge benefits to business organisations. So it is necessary to understand the various problems and challenges in adapting this technology. Due to large size of data, it becomes very difficult, inefficient and next to impossible to use conventional methods of processing and analysis. Normal operations like data capture, storage, search, sharing, analysis, etc. cannot be performed using normal methods and here is where the difficulties arise. The challenges in efficient handling of big data arise due to its volume, velocity, variability, complexity among other things. The various challenges faced in large data management include scalability, unstructured data, accessibility, real time analytics, fault tolerance and many more. For instance, Walmart handles more than 1 R S. Publication, rspublicationhouse@gmail.com Page 84

million customer transactions every hour. Facebook handles 40 billion photos from its user base. Decoding the human genome originally took 10years to process; now it can be achieved in one week. But as we can see that all of our existing techniques, seems of no use to us as these data s are rising exponentially. Thus with the help of this paper, we will try to reduce the difficulties and existing problems in handling of Big Data, by proposing some solutions and algorithms along with the existing ones up to a greater level. 2. LITERATURE REVIEW:- There has been lots of work in the field of Big Data but still a lot is needed to be done in this field to enhance the processing of very large quantities of digital information that cannot be analyzed with our traditional computing techniques. There are various other issues and challenges i.e. examining large amount of data, then checking for its authenticity, identification of hidden patterns, unknown correlation, effective marketing, customer satisfaction, and increased revenue. The various issues in this this field are discussed below:- Issues in big data:- The Big Data issues are important to know and crucial to handle for organization to implement the technology effectively. a) Data Volume:- The sheer size of the data set is enormous, making traditional data processing methods impossible. Today, Facebook ingests 500 terabytes of new data every day. Boeing 737 will generate 240 terabytes of flight data during a single flight across the US. This shows the data volume is too bulky to handle. b) Data Variety:- The data can be represented in a wide range of types, structured o run structured, including text, sensor, streaming audio or video, or userclickstreams, to name a few. So for different varieties of data, we require different technologies and methodologies. c) Data Velocity:- Many data need to be processed and analysed in near real-time, definitely difficult to be handled using the existing traditional systems E- Commerce has rapidly increased the speed and richness of data used for different business transactions (for example, web-site clicks). d) Data Complexity:- Working with it using relational databases and desktop statistics/visualization packages, requiring massively parallel software running on tens, hundreds, or even thousands of servers and it is also necessary to connect and R S. Publication, rspublicationhouse@gmail.com Page 85

correlate relationships, hierarchies and multiple data linkages or data can quickly spiral out of control. This tells us about the difficulties and complexities in handling Big Data. e) Storage and Transport Issues:- The quantity of data has exploded each time we have invented a new storage and that to there has been an exponential rise in it. The difference about the most recent data explosion, mainly due to social media, is that there has been no new storage medium. Moreover, data is being created by everyone and everything, (from Mobile Devices to Super Computers) not just, as here to fore, by professionals such as scientist, journalists, writers etc. Current disk technology limits would take longer time to transmit the data from a collection or storage point to a processing point than the time required to actually process it. FIGURE-1: THE INTERNET OF THINGS WAS BORN BETWEEN 2008 AND 2009 R S. Publication, rspublicationhouse@gmail.com Page 86

Challenges in Big Data: The challenges in Big Data are usually the real implementation hurdles which require immediate attention. For instance, take the example of the smart phones we use, the amount of data they create and consume; the sensors embedded into everyday objects will soon result in billions of new, constantly-updated data feeds containing environmental, location, and other information, including video. Ignoring these challenges may lead to the failure of the technology implementation and some unpleasant results. a) Privacy and Security:- It is the most important challenges with big data which are sensitive and includes conceptual, technical as well as legal significance. The personal information (e.g. in database of a merchant or social networking website) of a person when combined with external large data sets, leads to the inference of new facts about that person and it s possible that these kinds of facts about the person are secretive and the person might not want the data owner to know or any person to know about them. So these are the most important challenges of big data we face in our day to day life. Big Data used by law enforcement will increase the chances of certain tagged people to suffer from adverse consequences without the ability to fight back or even having knowledge that they are being discriminated. b) Data access and sharing information:- To access the data whenever we need it is necessary and very important to store it in an accurate, neat, complete and timely manner. Randomness in data storage is directly proportional to the difficulty in accessing the data. This makes the data management process a little bit complex adding the necessity to make data open and make it available to government agencies in standardized manner with standardized APIs, formats and metadata hence leading to better decision making and optimised productivity. Sharing data about the clients and operations threatens the culture of secrecy and competitiveness. This helps us in concluding that, the data access and sharing is one of the key concerns of Big Data. c) Analytical Challenges:- The main challenging questions arise as: What if it is not known how to deal the large data? R S. Publication, rspublicationhouse@gmail.com Page 87

Does all data need to be stored and to be analysed? How to find out which data points are crucial? How can the data be used to get the maximum potential? Huge analytical challenges in Big Data require a large number of advance skills i.e. decision making. This can be done by using one of the two techniques: either incorporate massive data volumes in analysis or determine the upfront for which Big data is relevant. d) Human Resources and Manpower:- Since Big data is at its prior-stage and an emerging technology so it needs to attract organizations and youth with diverse new skill sets. These skills should not be bounded to technical ones but also should spread to research, analytical and creative ones. These skills need to be developed in individuals hence requires training programs to be held by the organizations. Moreover the Universities need to introduce curriculum on Big Data to produce skilled employees in this area, which can lead the field by pursuing some research and giving probable solutions to the existing ones. e) Technical Challenges:- Fault Tolerance:- Fault-tolerant computing is extremely tough, involving complex algorithms. Whenever the failure occurs the damage done should be within acceptable area rather than starting the whole task from the scratch. The 100% reliable fault tolerant machines or software cannot be made, thus the main task is to reduce the probability of failure to an "acceptable" level. The reduction in this probability is inversely proportional to the higher cost. Two methods which seem to increase the fault tolerance in Big data are as: o First is, to divide the whole computation being done into tasks and assign these tasks to different nodes for computation. It is like divide and conquer. o Second is, one node is assigned the work of observing that these nodes are working properly. If something happens that particular task is restarted. But the sometimes in the processes like the recursive calls in which the whole computation can t be divided into such split tasks since it requires the previous output as an input. Thus restarting the whole computation becomes cumbersome process. This can be avoided by applying Checkpoints which R S. Publication, rspublicationhouse@gmail.com Page 88

keeps the state of the system at certain intervals of the time. In case of any failure, the computation can restart from last checkpoint maintained. Scalability:- The scalability issue of Big data has pushed towards cloud computing, which now aggregates dissimilar workloads with varying performance goals into very large clusters. This also put a question on efficiency of work. It also requires dealing with the system failures in an efficient manner which occurs more frequently if operating on large clusters. These factors combined put the concern on how to express the programs, even complex machine learning tasks. The shift in usage of technologies has raised a question that what kinds of storage devices are to be used. Quality of Data:- Business Leaders will always want more and more data storage whereas the IT Leaders will take all technical aspects in mind before storing all the data. Collection of huge amount of data and its storage comes at a cost; hence it put the stress on the notion of Quality over Quantity. This further leads to various questions like how it can be ensured that which data is relevant, how much data would be enough for decision making and whether the stored data is accurate or not to draw conclusions from it etc. Heterogeneous Data:- Randomness in data is increasing day-by-day. Unstructured data represents almost every kind of data being produced like social media interactions, to recorded meetings, to handling of PDF documents, fax transfers, to emails and more. Diversity in data is cumbersome and of course costly too. Well organised data is always easy to handle, but converting all this unstructured data into structured one is also not feasible. 3. EXISTING TECHNIQUES:- Big data requires exceptional methods and technologies to efficiently handle it within tolerable time. A 2011 McKinsey report suggests suitable technologies include A/B testing, crowdsourcing, data fusion and integration, genetic algorithms, machine learning, natural language processing among others. Multidimensional big data can be handled using tensor-based computation. The software frameworks available for handling big data include Hadoop, Hive, HBase, etc. A/B TESTING:- In marketing and business intelligence, A/B testing is jargon for a randomised experiment with two variants, A and B, which are the control and R S. Publication, rspublicationhouse@gmail.com Page 89

treatment in the controlled experiment. It is statistical hypothesis testing with two variants leading to the technical term, Two-sample hypothesis testing. In online settings, such as web design (especially user experience design); the goal is to identify changes to web pages that increase or maximise an outcome of interest (e.g., click-through rate for a banner advertisement). As the name implies, two versions or hypotheses (A and B) are compared, which are identical except for one variation that might affect a user's behaviour. Version A might be the currently used version (control), while Version B is modified in some respect (treatment). For instance, on an e-commerce website the purchase funnel is typically a good candidate for A/B testing, as even marginal improvements in drop-off rates can represent a significant gain in sales. Significant improvements can sometimes be seen through testing elements like copy text, layouts, images and colours, but not always. The vastly larger group of statistics broadly referred to as Multivariate testing or multinomial testing is similar to A/B testing, but may test more than two different versions at the same time and/or has more controls, etc. Simple A/B tests are not valid for observational, quasiexperimental or other non-experimental situations, as is common with survey data, offline data, and other, more complex phenomena. A/B testing has been marketed by some as a change in philosophy and business strategy in certain niches, though the approach is identical to a between-subjects design, which is commonly used in a variety of research traditions. A/B testing as a philosophy of web development brings the field into line with a broader movement toward evidence-based practice. CROWDSOURCING:- Crowdsourcing is the process of obtaining needed content and services from a large group of people who are not the traditional suppliers. This large group of people is usually from an online community. This method is used to divide tedious work. The Crowd Capability construct contains three dimensions: structure, content, and process, representing the components of engaging a Crowd. The structure component represents the IT used to engage a Crowd; the content dimension represents the type of input that is desired from a Crowd, while the process dimension represents the organizational processes used to foment resources from R S. Publication, rspublicationhouse@gmail.com Page 90

the Crowd. All three components of Crowd Capability must be implemented by an organization to potentially generate heterogeneous resources from dispersed knowledge. MACHINE LEARNING:-Machine learning refers to the algorithms that have the capability of learning from the data. Machine learning his strongly related to artificial intelligence and optimisation. These algorithms operate on model based inputs and they use that to make predictions or decisions. Example applications include spam filtering, optical character recognition (OCR), search engines and computer vision. In 1959, Arthur Samuel defined machine learning as a "Field of study that gives computers the ability to learn without being explicitly programmed". a) Decision tree learning:- Decision tree learning uses a decision tree as a predictive model which maps observations about an item to conclusions about the item's target value. b) Association rule learning:- Association rule learning is a method for discovering interesting relations between variables in large databases. c) Artificial neural networks:- An artificial neural network (ANN) learning algorithm, usually called "neural network" (NN), is a learning algorithm that is inspired by the structure and functional aspects of biological neural networks. Computations are structured in terms of an interconnected group of artificial neurons, processing information using a connectionist approach to computation. Modern neural networks are non-linear statistical data modelling tools. They are usually used to model complex relationships between inputs and outputs, to find patterns in data, or to capture the statistical structure in an unknown joint probability distribution between observed variables. d) Inductive logic programming:- Inductive logic programming (ILP) is an approach to rule learning using logic programming as a uniform representation for input examples, background knowledge, and hypotheses. Given an encoding of the known background knowledge and a set of examples represented as a logical database of facts, an ILP system will derive a hypothesized logic program which entails all the positive and none of the negative examples. Inductive programming is a related field that R S. Publication, rspublicationhouse@gmail.com Page 91

considers any kind of programming languages for representing hypotheses (and not only logic programming), such as functional programs. e) Support vector machines:- Support vector machines (SVMs) are a set of related supervised learning methods used for classification and regression. Given a set of training examples, each marked as belonging to one of two categories, an SVM training algorithm builds a model that predicts whether a new example falls into one category or the other. f) Clustering:- Cluster analysis is the assignment of a set of observations into subsets (called clusters) so that observations within the same cluster are similar according to some predesigned criterion or criteria, while observations drawn from different clusters are dissimilar. Different clustering techniques make different assumptions on the structure of the data, often defined by some similarity metric and evaluated for example by internal compactness (similarity between members of the same cluster) and separation between different clusters. Other methods are based on estimated density and graph connectivity. Clustering is a method of unsupervised learning, and a common technique for statistical data analysis. g) Bayesian networks:- A Bayesian network, belief network or directed acyclic graphical model is a probabilistic graphical model that represents a set of random variables and their conditional independencies via a directed acyclic graph (DAG). For example, a Bayesian network could represent the probabilistic relationships between diseases and symptoms. Given symptoms, the network can be used to compute the probabilities of the presence of various diseases. Efficient algorithms exist that perform inference and learning. h) Reinforcement learning:- Reinforcement learning is concerned with how an agent ought to take actions in an environment so as to maximize some notion of long-term reward. Reinforcement learning algorithms attempt to find a policy that maps states of the world to the actions the agent ought to take in those states. Reinforcement learning differs from the supervised learning problem in that correct input/output pairs are never presented, nor sub-optimal actions explicitly corrected. i) Representation learning:- Several learning algorithms, mostly unsupervised learning algorithms, aim at discovering better representations R S. Publication, rspublicationhouse@gmail.com Page 92

of the inputs provided during training. Classical examples include principal components analysis and cluster analysis. Representation learning algorithms often attempt to preserve the information in their input but transform it in a way that makes it useful, often as a pre-processing step before performing classification or predictions, allowing reconstructing the inputs coming from the unknown data generating distribution, while not being necessarily faithful for configurations that are implausible under that distribution. Manifold learning algorithms attempt to do so under the constraint that the learned representation is low-dimensional. Sparse coding algorithms attempt to do so under the constraint that the learned representation is sparse (has many zeros). Multilinear subspace learning algorithms aim to learn low-dimensional representations directly from tensor representations for multidimensional data, without reshaping them into (high-dimensional) vectors. Deep learning algorithms discover multiple levels of representation, or a hierarchy of features, with higher-level, more abstract features defined in terms of (or generating) lower-level features. It has been argued that an intelligent machine is one that learns a representation that disentangles the underlying factors of variation that explain the observed data. j) Similarity and metric learning:- In this problem, the learning machine is given pairs of examples that are considered similar and pairs of less similar objects. It then needs to learn a similarity function (or a distance metric function) that can predict if new objects are similar. It is sometimes used in Recommendation systems. k) Genetic algorithms:- A genetic algorithm (GA) is a search heuristic that mimics the process of natural selection, and uses methods such as mutation and crossover to generate new genotype in the hope of finding good solutions to a given problem. In machine learning, genetic algorithms found some uses in the 1980s and 1990s. R S. Publication, rspublicationhouse@gmail.com Page 93

FIGURE-2: BIG DATA PROCESSING FLOW CHART HADOOP:- Apache Hadoop, a nine-year-old open-source data-processing platform first used by Internet giants including Yahoo and Facebook, leads the big-data revolution. Cloudera introduced commercial support for enterprises in 2008, and MapR and Hortonworks piled on in 2009 and 2011, respectively. Among data-management incumbents, IBM and EMC-spinout Pivotal each has introduced its own Hadoop distribution. Microsoft and Teradata offer complementary software and first-line support for Horton works' platform. Oracle resells and supports Cloud era, while HP, SAP, and others act more like Switzerland, working with multiple Hadoop software providers. The Hadoop Distributed File System (HDFS) splits files into large blocks and distributed them amongst nodes of clusters of commodity hardware. For processing the data, the Hadoop Map/Reduce ships code (specifically Jar files) to the nodes that have the required data and the nodes then process the data in parallel. This approach takes advantage of data locality. The base Apache Hadoop framework is composed of the following modules:- a) Hadoop Common: contains libraries and utilities needed by other Hadoop modules. b) Hadoop Distributed File System (HDFS): a distributed file-system that stores data on commodity machines, providing very high aggregate bandwidth across the cluster. R S. Publication, rspublicationhouse@gmail.com Page 94

c) Hadoop YARN: A resource-management platform responsible for managing computes resources in clusters and using them for scheduling of users' applications. d) Hadoop Map Reduce: a programming model for large scale data processing. 4. CONCLUSION:- Thus we have seen why big data is of great significance in today s world and how it has the potential to influence the upcoming technologies and the future market. Our newest research finds that organizations are using big data to target customer-centric outcomes, tap into internal data and build a better information ecosystem on database and data analytics market. It offers commercial opportunities of a comparable scale to enterprise software in the late 1980s. And the Internet boom of the 1990s, and the social media explosion of today. We have seen the challenges big data handling presents and the available solutions. Though these solutions are efficient, still research is being carried out to better the existing techniques as well as develop new ones. We have also presented the on-going and future research into this sphere. A lot of time and money is being spent in this research because big data handling methods and techniques are the future. 5. RESULT:- Big data is of great potential value. For example, it is of $300 billion potential annual value to the US health care. Real-time big data is not just a process of storing petabytes or exabytes of data, it is the ability to make better decisions and take proper actions at the right time. A lot of money is being spent on research. In this paper we have successfully presented the issues, challenges, existing solutions, future works as well as the potential and benefits of big data. 6. FUTURE OF BIG DATA:- Big data itself is the future. A lot of capital is being invested into research into this sphere. The big data industry is worth more than $100 billion! Moreover it is growing at the rate of 10% a year. In February 2012, the open source analyst firm Wikibon released the first market forecast for Big Data, listing $5.1B revenue in 2012 with growth to $53.4B in 2017. The McKinsey Global Institute estimates that data volume is growing 40% per year, and will grow 44x between 2009 and 2020. R S. Publication, rspublicationhouse@gmail.com Page 95

Big Data is a very big domain and it has lot of scope for its development and enhancement. Hopefully the upcoming mechanisms will be furthermore beneficial in handling the Big Data in effective and efficient manner. 7. REFERENCES:- 1. James, J. "How Much Data is Created Every Minute?" Retrieved November 3, 2012 from http://www.domo.com/blog/2012/06/how-much-data-is-created-everyminute/?dkw=socf3 2. IBM Corporation. "What is big data?" from http://www- 01.ibm.com/software/data/bigdata/ 3. Aveksa Inc. (2013). Ensuring Big Data Security with Identity and Access Management. Waltham, MA: Aveksa.. 4. Kaisler, S., Armour, F., Espinosa, J. A., & Money, W. (2013). Big Data: Issues and Challenges Moving Forward. International Confrence on System Sciences (pp. 995-1004). Hawaii: IEEE Computer Soceity. 5. Patel, A. B., Birla, M., & Nair, U. (2013). Addressing Big Data Problem Using Hadoop and. Nirma University, Gujrat: Nirma University. 6. Katal, A., Wazid, M., & Goudar, R. H. (2013). Big Data: Issues, Challenges, Tools and Good Practices. IEEE, 404-409. 7. Singh, S., & Singh, N. (2012). Big Data Analytics. International Conference on Communication, Information & Computing Technology (ICCICT) (pp. 1-4). Mumbai: IEEE. R S. Publication, rspublicationhouse@gmail.com Page 96