Bigdata (Horton Works HDP-2.0)
Big Data is nothing but collection of data, data sets which
is of unstructured data .
Big Data is useful under the large growing data where it is
unable to manage.
Hadoop V2.0 is for running mapreduce job of the particular
Hadoop V2.0 consists of Mapreduce ,Yarn and HDFS .
Above mentioned three components are important in
Hadoop Distributed file system.(HDFS).
Handles large data with streaming data Access.
Runs on top of all file system.
Uses Blocks to store files.
Frame work for performing calculations on data in HDFS.
Distributed Data Processing.
Resource and Scheduler Manager.
Machine Learning Language
Python V2.7.6 Modules on Dev,staging and
Cloud Components for Staging & Production.
Centos V6.4 Instance .
Bucket for Storage of files.
WordPress Blog with Version 3.8.1.
JQuery on WordPress.
Visualization on instance.
.pem file for connecting Cloud from Local machine.
.ppk file for moving data from Local system to cloud
Public Ip (Elastic Ip for the Instance).
Dev,Staging and Production.
Maintain same version on all the three stages of
Description of Database:
Handles structured, unstructured and polymorphic.
Scale up with Bigdata.
MongoHQ for MongoDB server.
Backup & Restore Data from DB.