2. requirements of those executing and consuming the results of the process [1]. This
administrative behavior was the underlying theme for Herbert Simon’s view of
bounded rationality. The dispute in decision making and administrative behaviors
makes a sense, as we bound the data in the process of modeling, applying algo-
rithmic applications, and have always been seeking discrete relationships within the
data as opposed to the whole picture [3]. The big data analytics evolved to support
the decision making process by collecting, organizing, storing, retrieving, and
managing big data.
1.1 Need for Big Data Analytics
Globalization and personalization of services are some crucial changes that moti-
vate Big Data Analytics. Many issues in real applications need to be described
using a graphical structure such as the optimization of railway paths, prediction of
disease outbreaks, and emerging applications such as analysis of social network,
semantic network, analysis of biological information networks, etc. [4]. Figure 1
shows the hierarchy of data analysis work.
1.2 Big Data Processing
Conceptually there are two ways of data processing that are accepted as a de facto
standard. In centralized processing collected data is stored at a centralized location
and then processed on a single machine/server. This centralized machine/server
has very high configurations in terms of memory, processor, and storage. This
Fig. 1 Data layers in big data
analytics
30 G. Pole and P. Gera
3. architecture is well suited for small organizations with one location for production
and service, but for large organizations having different geographical locations is
almost out dated. Distributed processing evolved to overcome the limitations of the
centralized processing, where there is no need to collect and process all data at a
central location. There are several architectures of distributed processing. For
example Cluster architecture and Peer-to-peer architecture (Fig. 2).
2 Big Data Analysis Steps
Big data involves data sets whose size is challenging to present tools and tech-
nologies to manage and process the data within acceptable time [5]. Journey of raw
data to make useful decisions using big data analytics is governed by the following
steps (Fig. 3).
2.1 Data Collection
Data collection is a process of retrieving raw data from real-world data sources. The
data are fetched from different sources and stored to a file system. Inaccurate data
collection can affect the succeeding data analysis procedure and may lead to
unacceptable results, so the data collection process should to be carefully designed
[6]. In the modern world, the distributed data collection from different geographical
systems and networks is pervasive and there is a demand to discover the hidden
patterns of the data stored with these heterogeneous or homogeneous distributed
nodes [7]. Useful Tools: Chukwa, WebCrawler, Pig, and Flume.
Big Data
Processing
Centralized
Processing
Super Computing System
Distributed
Processing
Cluster Architecture
Peer - to peer Architecture
Fig. 2 Fundamental styles for data processing
A Recent Study of Emerging Tools and Technologies … 31
4. 2.2 Data Partition
To handle very large data volume different partitioning techniques like data tagging,
classification, and incremental clustering approaches are used. Many clustering
approaches have been designed to help address a large data [7, 8]. Useful Tools:
Textual ETL, Cassandra. To minimize the complexity of processing large data
several algorithms are emerged. These algorithms are categorized as Text mining,
Data mining, Pattern processing, Mahout, Scalable nearest Neighbor Algorithms [9].
2.3 Data Coordination
Coordination governs the exchange of data among the data warehouses, relational
databases, NoSQL Databases, and different big data technologies. For example,
Sqoop [10] can be used to exchange data to and from relational databases like
MySQL or Oracle to the HDFS. The cross-cloud service requires data aggregation
and transfer among different systems to process large-scale big data from different
sources [11]. Flume is a distributed system that manages large amount of data from
different systems and networks by efficiently collecting, aggregating, and moving it
to a desired location [12]. On the other hand the Zookeeper is a centralized service
which provides distributed synchronization, naming, and group services. It also
maintains the configuration information of the systems. The APIs are available for
various programming languages like Java, Perl, and Python [13].
2.4 Data Transformation
Data transformation refers to the conversion of data values in one format of source
system to another format of destination system. Useful tools: ETL, DMT, and Pig.
Fig. 3 Different steps in big data analysis
32 G. Pole and P. Gera
5. DMT is a Data Migration Tool from TCS. It provides accurate data migration
from conventional databases to Hadoop repository like HDFS or Hbase.
2.5 Data Storage
The main challenge in front of scalable bigdata system is the need to store and
manage large volume of data collected. At the same time, it should provide methods
and functions for efficient access, retrieval, and manipulation to diverse datasets.
Due to a variety of dissimilar data sources and large volume, it is tough to gather
and integrate data with performance guarantee from distributed locations. The big
data analytic system must effectively mine such huge datasets at different levels in
real time or near real time [6]. Useful Tools: Hbase [14], NoSQL [15], Gluster [16],
HDFS [17], GFS [18].
2.6 Data Processing
There is no fixed data processing model to handle the big data because of its
complex nature. The data processing in big data analytics is rather schema less or
non-structured. To process the structured and unstructured data in different format,
a mixture of various technologies like. Hadoop, NoSQL, and MapReduce is used.
MapReduce framework is popular for processing and transformation of very large
data sets [5]. The Qlikview is an example of In-memory data processing solutions
for big data which is very useful for advanced reporting. Infinispan is an extremely
scalable and highly available data processing grid platform [19].
2.7 Extract Data
This phase extracts data from the files and databases and produces result set, which
is used for further analytics, reporting, integration, and visualization of the result.
Data Query Tool Hive is a data query tool used for ad hoc queries, data
summarizations, data extraction, and other analysis of data. It uses HiveQL lan-
guage which supports the flexible querying to the big data [20].
Big Data Search In order to scale to huge data sets, we use the approach of
distributing the data to multiple machines in a compute cluster and perform the
nearest neighbor search using all the machines in parallel [9]. Lucene [21], Solr [22]
are the major examples of such search tools.
A Recent Study of Emerging Tools and Technologies … 33
6. 2.8 Data Analysis
RapidMiner is an open source system that uses data and text mining for predictive
analysis [23]. Pentaho is an open source business intelligence product which pro-
vides data integration, OLAP services, reporting, and ETL capabilities. Talend [24]
and SpagoBI are examples of widely used open source BI tool [25]. WEKA tool, a
part of large machine learning project, consists of data mining algorithms that can
be applied for data analysis. The machine learning and data mining communities
have developed different tools and algorithms for tackling all sorts of analysis
problems [26].
2.9 Data Visualization
The Big data is stored in files and not in table structure or format, so it is less
interactive in a visualization situation. Because of this, big data need statistical
softwares like R, SAS, and KXEN, where the predefined models for different
statistical functions can be used for data extraction and results are integrated for
statistical visualizations. Some of the well-known visualization tools are DIVE [27],
and Orange [28].
3 Programming for Big Data
There are various programes useful for big data processing. The most prominent
and powerful languages are R, Python, Java, Storm, etc. Python is very useful to
programers for doing statistics and efficient manipulation of statistical data. This
includes typically vectors, lists, and data frames that represent datasets organized in
rows and columns, Whereas R outshines in the range of statistical libraries it
compromises. Almost all statistical tests/methods are part of an R library. It is very
easy to learn language with a vast amount of inbuilt methods. NumPy library in
Python encompasses homogeneous, multidimensional array that offers various
methods for data manipulation. It has several functions for performing advanced
arithmetic and statistics [29]. Java and Storm [30] also play a major role in big data
programming.
34 G. Pole and P. Gera
7. 4 Conclusion
A changed paradigm of modern business and an advancement in communication
and technology has given a new face to the analytics. Like the light fastening in
computers, now people need superfast decision making and it is possible with big
data analytics. Advancement in tools and technologies made it possible. Best
practices have emerged to help big data processing. An interesting fact is that many
of these practices are the new empowered, flexible extensions of the old one. The
main thing behind the popularity of big data analysis is that it helps an organization
to take corrective actions and useful decisions without much of data processing
latency. Thus, big data enables decision capacity, nearly in run time environment.
References
1. X. Wu, X. Zhu, G.-Q. Wu, and W. Ding, “Data mining with big data,” IEEE Trans. Knowl.
Data Eng., vol. 26, no. 1, pp. 97–107, 2014.
2. L. Wang, K. Lu, P. Liu, R. Ranjan, and L. Chen, “IK-SVD: Dictionary Learning for Spatial
Big Data via Incremental Atom Update,” vol. XX, no. Xx, pp. 1–12, 2014.
3. M. Augier, “Sublime Simon: The consistent vision of economic psychology’s Nobel laureate,”
J. Econ. Psychol., vol. 22, no. 3, pp. 307–334, 2001.
4. Y. Liu, B. Wu, H. Wang, and P. Ma, “BPGM : A Big Graph Mining Tool,” vol. 19, no. 1,
2014.
5. S. Meng, W. Dou, X. Zhang, J. Chen, and S. Member, “KASR : A Keyword-Aware Service
Recommendation Method on MapReduce for Big Data Applications,” vol. 25, no. 12, pp. 1–
11, 2013.
6. D. Takaishi, S. Member, H. Nishiyama, and S. Member, “in Densely Distributed Sensor
Networks,” vol. 2, no. 3, 2014.
7. P. Shen and C. Li, “Distributed Information Theoretic Clustering,” vol. 62, no. 13, pp. 3442–
3453, 2014.
8. Y. Wang, L. Chen, S. Member, and J. Mei, “Incremental Fuzzy Clustering With Multiple
Medoids for Large Data,” vol. 22, no. 6, pp. 1557–1568, 2014.
9. M. Muja, “Scalable Nearest Neighbour Methods for High Dimensional Data,” vol. 36, no.
April, pp. 2227–2240, 2013.
10. (2012), sqoop [online]. Available: https://sqoop.apache.org/docs/1.4.2/SqoopUserGuide.htm.
11. W. Dou, X. Zhang, J. Liu, J. Chen, and S. Member, “HireSome -II : Towards Privacy-Aware
Cross- Cloud Service Composition for Big Data Applications,” vol. 26, no. 2, pp. 1–11, 2013.
12. (2013), Flume [online]. Available: https://flume.apache.org/FlumeUserGuide.html.
13. (2014), Zookeeper [online]. Available: https://zookeeper.apache.org/releases.html.
14. (2013). HBase [Online]. Available: http://hbase.apache.org/.
15. R. Cattell, “Scalable SQL and NoSQL data stores,’’ SIGMOD Rec., vol. 39, no. 4, pp. 12_27,
2011.
16. (2014), Gluster [online]. Available: http://www.gluster.org/.
17. (2013). Hadoop Distributed File System [Online]. Available: http://hadoop.apache.org/docs/
r1.0.4/hdfsdesign.html.
18. S. Ghemawat, H. Gobioff, and S.-T. Leung, “The Google file system,’’ in Proc. 19th ACM
Symp. Operating Syst. Principles, 2003, pp.29_43.
19. (2015), Infinispan [online]. Available: http://infinispan.org/documentation/.
A Recent Study of Emerging Tools and Technologies … 35
8. 20. A. Thusoo et al., “Hive: A warehousing solution over a Map-Reduceframework,’’ Proc. VLDB
Endowment, vol. 2, no. 2, pp. 1626_1629, 2009.
21. (2014), Lucene [online]. Available: https://lucene.apache.org/.
22. (2013). Solr [Online]. Available: http://lucene.apache.org/solr/.
23. (2013). Rapidminer [Online]. Available: https://rapidminer.com.
24. (2015). Talend [Online]. Available: https://www.talend.com/.
25. (2015). SpagoBI [Online]. Available: http://www.spagobi.org/.
26. D. Breuker, “Towards Model-Driven Engineering for Big Data Analytics – An Exploratory
Analysis of Domain-Specific Languages for Machine Learning,” 2014 47th Hawaii Int. Conf.
Syst. Sci., pp. 758–767, 2014.
27. S. J. Rysavy, D. Bromley, and V. Daggett, “DIVE: A graph-based visual-analytics framework
for big data,” IEEE Comput. Graph. Appl., vol. 34, no. 2, pp. 26–37, 2014.
28. (2015). Orange [Online]. Available: http://orange.biolab.si/.
29. P. Louridas and C. Ebert, “Embedded analytics and statistics for big data,” IEEE Softw., vol.
30, no. 6, pp. 33–39, 2013.
30. (2015). Storm [Online]. Available: http://storm-project.net/.
36 G. Pole and P. Gera