SlideShare a Scribd company logo
1 of 4
Manoj Kumar Sahu
Call-08446783942
Manojsahu.a9@gmail.com
SUMMARY
• Over 4 year of IT experience in Hadoop/Bigdata and Linux Support Environment.
• Having Knowledge of Cassandra administration, Application maintenance & support.
• Having knowledge of Hadoop Administration with Nosql Cassandra,Hbase and Mongodb
• Expertise with Hadoop and Cassandra tools like CM,Hortonwork Ops Centre.
• Having knowledge on file System like Hadoop,Cassandra (CFS) and Linux,Unix.
• Having Knowledge on Cassendra API ,Write Operation data and CQL.
• Expertise on Hadoop Architect and Cassandra Architect.
• Having knowledge on Cassandra Data Model, Keyspace, Interface and Cap Theorem.
• Hands on Commissioning, Decommissioning, Bootstrapping and Balancing for Node.
• Having knowledge on Network topology, Thrift, Avro and Gossip Protocols.
• Handling on ATT &T Mobility project Datalake Hadoop with respective team.
• Hands on experience installation and Management of Hadoop Cluster and Cassandra.
• Hands on experience in Hadoop eco system Hive,Pig,Sqoop,Flume,Hbase installations.
• Expertise with Hadoop Master and Slave daemons of HDFS and MapReduce.
• Having knowledge in Rack awareness and Cassandra cluster Centre.
• Hands on perform for Hadoop Distcp.
• Installation of HadoopEco system and Apache Cassandra (1.0 x , 1.2 x& 2.0x ).
• Hands on TWS (8.5.1) jobs monitoring and tracking using logs file.
• Hands on experiences in Nagios and Ganglia monitoring tool.
• Hands on Monitoring and Managing Hadoop respective daemons.
• Hands on experience in install and configuration hadoop Cloudera Manager.
• Monitoring the Cluster and deploying the daemons using CM, and Hortonwork.
• Good knowledge on MR2 (Yarn), HA(High Availability),hadoop JVM
• Good knowledge on Cluster Tuning and Node Benchmarking Processes.
• Hands on experience in installation and Configuration of Linux Redhat,CentOS,Ubuntu.
• Hands on Remote tools Super Putty and VMware workstation 10.
• Having Knowledge on Linux command and setup the Server single Node and Multiple.
• Creating Partitions Formating Mounting and change the Owner group and other
permissions.
• knowledge on LVM,Raid,Kick Start installation for linux
• Creating repository local and online through Rpm and Yum package.
• Good knowledge on linux NFS,FTP,LDAP.
Software Skillset:
• Hadoop: HDFS, MapReduce, Claudra Manager, Hortonwork.
• HadoopEcosystem: Hive, Pig, Sqoop, Flume, Hbase, Zookeeper.
• Security: Kerberos.
• Cluster Management Tools: Cloudera Manager, OpsCenter, JMX, Nodetool.
• NoSQL Databases: Cassandra.
• Relational Databases: Microsoft SQL Server, MySQL.
• Languages: C, Basic Java, C#, HTML PHP.
• Operating Systems: Linux, Unix, Cent OS 6.4, Windows
PROFESSIONAL EXPERIENCE:
Amdocs | Pune India Mar, 2014 –
Present
Hadoop/ Ops Administrator
Environment: Hadoop, SQL Server,DB,TWS9.2.
Responsibilities:
• As Hadoop & Cassandra Administrator support production server such as KM,STL,
BHM, Dev and QA.
• Monitor and handle application on TWS 9.2 production and Nagios Alerts.
• Deploying Application on Hadoop Cluster.
• Handling Gerick Framework and IF framework.
• Manage Files Systems and Creating Operation Scripts
• Source File monitoring and Run book Design package Reviews
• Automated and deployed Cassandra environments using Chef recipes.
• Evaluated, benchmarked and tuned data model by running endurance tests using
JMeter, Cassandra Stress Tool and OpsCenter.
• Used Cassandra node tool to manage Cassandra cluster.
• Worked on optimizing the Cassandra cluster by making changes in Cassandra
configuration file and Linux OS configurations.
• Management Reporting and Communication and Development Support (Env Queries,
New application Queries etc.)
• Involved in writing multiple scripts to monitor Cassandra cluster and OpsCenter.
• Distcp for Dual-Loaded sources and Monitoring of publishing feeds
• Handling of on-request jobs/re-publishing feeds and Adhoc data movement request for
QA Testing
• Job failure report and Overview report maintenance , when new sources added
• Deployment execution and code migrations and Long running job monitoring and
reporting."
Cloudwick Technology, INDIA 28 – Oct, 2011
Hadoop/ Linux Support Engineer
Environment: Hadoop, SQL Server,DB,TWS9.2,Linux 6.4.
Responsibilities:
• Scheduled repair and cleanup process in production environment.
• Administering and monitoring the cluster using tools
• Scheduled moving production data to development environment for testing purpose.
• Installed and configure mongodb server and mongo client package.
• Exported the JSON data into Mongo db.
• Install and configure Hive and Hive metadata and creating hive table to get result.
• Scheduled repair and cleanup process in production environment during off peak
times.
• Exporting Mongodb data into HDFS.
• Loading Hive data to transfer in to MSQL using Sqoop.
• Experience in storing the analyzed results back into the Cassandra cluster.
• Designed and configured gateway node to the cluster.
• Performed Stress and Performance testing, benchmark for the cluster.
• Installing Hadoop-Lzo Compression Using RPM and YUM repo.
• Adding compression properties into Hadoop Configuration file
• Exported the JSON data into Mongo db.
• Exporting Mongodb data into HDFS.
• Install and configure Hive and Hive metadata and creating hive table to get result.
• Creating Lzo file and inserting into HDFS.
• Creating Index and Running map-reduce Job .
• Installing Hdfs and adding property into configuration file.
• Adding the jdbc mysql connecter for Sqoop
• Compresses data import into Mysql using Sqoop.
• Add the property into flume configurations and load the streaming data into HDFS.
• Running streaming data into Hadoop cluster and monitoring with Ganglia.
• Monitoring the Health of Node using Nagios.
• Hands on experience in installation and Configuration of Linux Redhat,Cent OS,Ubuntu
• Hands on Remote tools VNC viewer,Teamviewer and VMware worksation 10.
• Configuration and managing SSH keygen for cluster.
• Basic knowledge on LVM,Raid, installestion for linux and LAMP.
• Creating repository local and online through Rpm and Yum package.

More Related Content

What's hot

Avi Jain - Resume Descriptive 2015
Avi Jain - Resume Descriptive 2015Avi Jain - Resume Descriptive 2015
Avi Jain - Resume Descriptive 2015AVI JAIN
 
SUSE, Hadoop and Big Data Update. Stephen Mogg, SUSE UK
SUSE, Hadoop and Big Data Update. Stephen Mogg, SUSE UKSUSE, Hadoop and Big Data Update. Stephen Mogg, SUSE UK
SUSE, Hadoop and Big Data Update. Stephen Mogg, SUSE UKhuguk
 
KOTI_RESUME_(1) (2)
KOTI_RESUME_(1) (2)KOTI_RESUME_(1) (2)
KOTI_RESUME_(1) (2)ch koti
 
TareqZantoot-Haddop_Administrator_Security
TareqZantoot-Haddop_Administrator_SecurityTareqZantoot-Haddop_Administrator_Security
TareqZantoot-Haddop_Administrator_SecurityTarek Elzanaty
 
Hadoop in the Cloud - The what, why and how from the experts
Hadoop in the Cloud - The what, why and how from the expertsHadoop in the Cloud - The what, why and how from the experts
Hadoop in the Cloud - The what, why and how from the expertsDataWorks Summit/Hadoop Summit
 
Hadoop Security and Compliance - StampedeCon 2016
Hadoop Security and Compliance - StampedeCon 2016Hadoop Security and Compliance - StampedeCon 2016
Hadoop Security and Compliance - StampedeCon 2016StampedeCon
 
Big data architecture on cloud computing infrastructure
Big data architecture on cloud computing infrastructureBig data architecture on cloud computing infrastructure
Big data architecture on cloud computing infrastructuredatastack
 
Hadoop World 2011: Hadoop and RDBMS with Sqoop and Other Tools - Guy Harrison...
Hadoop World 2011: Hadoop and RDBMS with Sqoop and Other Tools - Guy Harrison...Hadoop World 2011: Hadoop and RDBMS with Sqoop and Other Tools - Guy Harrison...
Hadoop World 2011: Hadoop and RDBMS with Sqoop and Other Tools - Guy Harrison...Cloudera, Inc.
 
Hadoop Ecosystem at a Glance
Hadoop Ecosystem at a GlanceHadoop Ecosystem at a Glance
Hadoop Ecosystem at a GlanceNeev Technologies
 
Data Protection in Hybrid Enterprise Data Lake Environment
Data Protection in Hybrid Enterprise Data Lake EnvironmentData Protection in Hybrid Enterprise Data Lake Environment
Data Protection in Hybrid Enterprise Data Lake EnvironmentDataWorks Summit
 
Hadoop in three use cases
Hadoop in three use casesHadoop in three use cases
Hadoop in three use casesJoey Echeverria
 
sam_resume - updated
sam_resume - updatedsam_resume - updated
sam_resume - updatedsam k
 
Hadoop Innovation Summit 2014
Hadoop Innovation Summit 2014Hadoop Innovation Summit 2014
Hadoop Innovation Summit 2014Data Con LA
 
HPC and cloud distributed computing, as a journey
HPC and cloud distributed computing, as a journeyHPC and cloud distributed computing, as a journey
HPC and cloud distributed computing, as a journeyPeter Clapham
 
DataStax - Analytics on Apache Cassandra - Paris Tech Talks meetup
DataStax - Analytics on Apache Cassandra - Paris Tech Talks meetupDataStax - Analytics on Apache Cassandra - Paris Tech Talks meetup
DataStax - Analytics on Apache Cassandra - Paris Tech Talks meetupVictor Coustenoble
 
October 2016 HUG: The Pillars of Effective Data Archiving and Tiering in Hadoop
October 2016 HUG: The Pillars of Effective Data Archiving and Tiering in HadoopOctober 2016 HUG: The Pillars of Effective Data Archiving and Tiering in Hadoop
October 2016 HUG: The Pillars of Effective Data Archiving and Tiering in HadoopYahoo Developer Network
 

What's hot (20)

Avi Jain - Resume Descriptive 2015
Avi Jain - Resume Descriptive 2015Avi Jain - Resume Descriptive 2015
Avi Jain - Resume Descriptive 2015
 
SUSE, Hadoop and Big Data Update. Stephen Mogg, SUSE UK
SUSE, Hadoop and Big Data Update. Stephen Mogg, SUSE UKSUSE, Hadoop and Big Data Update. Stephen Mogg, SUSE UK
SUSE, Hadoop and Big Data Update. Stephen Mogg, SUSE UK
 
KOTI_RESUME_(1) (2)
KOTI_RESUME_(1) (2)KOTI_RESUME_(1) (2)
KOTI_RESUME_(1) (2)
 
TareqZantoot-Haddop_Administrator_Security
TareqZantoot-Haddop_Administrator_SecurityTareqZantoot-Haddop_Administrator_Security
TareqZantoot-Haddop_Administrator_Security
 
Hadoop in the Cloud - The what, why and how from the experts
Hadoop in the Cloud - The what, why and how from the expertsHadoop in the Cloud - The what, why and how from the experts
Hadoop in the Cloud - The what, why and how from the experts
 
Hadoop Security and Compliance - StampedeCon 2016
Hadoop Security and Compliance - StampedeCon 2016Hadoop Security and Compliance - StampedeCon 2016
Hadoop Security and Compliance - StampedeCon 2016
 
Big data architecture on cloud computing infrastructure
Big data architecture on cloud computing infrastructureBig data architecture on cloud computing infrastructure
Big data architecture on cloud computing infrastructure
 
Introduction to Hadoop
Introduction to HadoopIntroduction to Hadoop
Introduction to Hadoop
 
HDFS
HDFSHDFS
HDFS
 
Hadoop World 2011: Hadoop and RDBMS with Sqoop and Other Tools - Guy Harrison...
Hadoop World 2011: Hadoop and RDBMS with Sqoop and Other Tools - Guy Harrison...Hadoop World 2011: Hadoop and RDBMS with Sqoop and Other Tools - Guy Harrison...
Hadoop World 2011: Hadoop and RDBMS with Sqoop and Other Tools - Guy Harrison...
 
Hadoop Ecosystem at a Glance
Hadoop Ecosystem at a GlanceHadoop Ecosystem at a Glance
Hadoop Ecosystem at a Glance
 
Data Protection in Hybrid Enterprise Data Lake Environment
Data Protection in Hybrid Enterprise Data Lake EnvironmentData Protection in Hybrid Enterprise Data Lake Environment
Data Protection in Hybrid Enterprise Data Lake Environment
 
Hadoop in three use cases
Hadoop in three use casesHadoop in three use cases
Hadoop in three use cases
 
Apache Hadoop 3
Apache Hadoop 3Apache Hadoop 3
Apache Hadoop 3
 
sam_resume - updated
sam_resume - updatedsam_resume - updated
sam_resume - updated
 
Hadoop Innovation Summit 2014
Hadoop Innovation Summit 2014Hadoop Innovation Summit 2014
Hadoop Innovation Summit 2014
 
PPT on Hadoop
PPT on HadoopPPT on Hadoop
PPT on Hadoop
 
HPC and cloud distributed computing, as a journey
HPC and cloud distributed computing, as a journeyHPC and cloud distributed computing, as a journey
HPC and cloud distributed computing, as a journey
 
DataStax - Analytics on Apache Cassandra - Paris Tech Talks meetup
DataStax - Analytics on Apache Cassandra - Paris Tech Talks meetupDataStax - Analytics on Apache Cassandra - Paris Tech Talks meetup
DataStax - Analytics on Apache Cassandra - Paris Tech Talks meetup
 
October 2016 HUG: The Pillars of Effective Data Archiving and Tiering in Hadoop
October 2016 HUG: The Pillars of Effective Data Archiving and Tiering in HadoopOctober 2016 HUG: The Pillars of Effective Data Archiving and Tiering in Hadoop
October 2016 HUG: The Pillars of Effective Data Archiving and Tiering in Hadoop
 

Similar to Manoj CV

Big data_hadoop_spark_kafka_nosql_training
Big data_hadoop_spark_kafka_nosql_trainingBig data_hadoop_spark_kafka_nosql_training
Big data_hadoop_spark_kafka_nosql_trainingKamal A
 
Distro-independent Hadoop cluster management
Distro-independent Hadoop cluster managementDistro-independent Hadoop cluster management
Distro-independent Hadoop cluster managementDataWorks Summit
 
Amit Anand - devops
Amit Anand - devopsAmit Anand - devops
Amit Anand - devopsAmit Anand
 
HariKrishna4+_cv
HariKrishna4+_cvHariKrishna4+_cv
HariKrishna4+_cvrevuri
 
Hd insight essentials quick view
Hd insight essentials quick viewHd insight essentials quick view
Hd insight essentials quick viewRajesh Nadipalli
 
HdInsight essentials Hadoop on Microsoft Platform
HdInsight essentials Hadoop on Microsoft PlatformHdInsight essentials Hadoop on Microsoft Platform
HdInsight essentials Hadoop on Microsoft Platformnvvrajesh
 
Hd insight essentials quick view
Hd insight essentials quick viewHd insight essentials quick view
Hd insight essentials quick viewRajesh Nadipalli
 
Hadoop Administration Online Training.pdf
Hadoop Administration Online Training.pdfHadoop Administration Online Training.pdf
Hadoop Administration Online Training.pdfSpiritsoftsTraining
 
Learn Hadoop at your Leisure time
Learn Hadoop at your Leisure time Learn Hadoop at your Leisure time
Learn Hadoop at your Leisure time Saantosh Rohera
 
Big data - Online Training
Big data - Online TrainingBig data - Online Training
Big data - Online TrainingLearntek1
 
Deployment and Management of Hadoop Clusters
Deployment and Management of Hadoop ClustersDeployment and Management of Hadoop Clusters
Deployment and Management of Hadoop ClustersAmal G Jose
 
Hadoop Operations - Best practices from the field
Hadoop Operations - Best practices from the fieldHadoop Operations - Best practices from the field
Hadoop Operations - Best practices from the fieldUwe Printz
 
Tcloud Computing Hadoop Family and Ecosystem Service 2013.Q3
Tcloud Computing Hadoop Family and Ecosystem Service 2013.Q3Tcloud Computing Hadoop Family and Ecosystem Service 2013.Q3
Tcloud Computing Hadoop Family and Ecosystem Service 2013.Q3tcloudcomputing-tw
 
Infrastructure Around Hadoop
Infrastructure Around HadoopInfrastructure Around Hadoop
Infrastructure Around HadoopDataWorks Summit
 

Similar to Manoj CV (20)

Big data_hadoop_spark_kafka_nosql_training
Big data_hadoop_spark_kafka_nosql_trainingBig data_hadoop_spark_kafka_nosql_training
Big data_hadoop_spark_kafka_nosql_training
 
Hadoop_Admin_eVenkat
Hadoop_Admin_eVenkatHadoop_Admin_eVenkat
Hadoop_Admin_eVenkat
 
Distro-independent Hadoop cluster management
Distro-independent Hadoop cluster managementDistro-independent Hadoop cluster management
Distro-independent Hadoop cluster management
 
Amit Anand - devops
Amit Anand - devopsAmit Anand - devops
Amit Anand - devops
 
HariKrishna4+_cv
HariKrishna4+_cvHariKrishna4+_cv
HariKrishna4+_cv
 
Hd insight essentials quick view
Hd insight essentials quick viewHd insight essentials quick view
Hd insight essentials quick view
 
HdInsight essentials Hadoop on Microsoft Platform
HdInsight essentials Hadoop on Microsoft PlatformHdInsight essentials Hadoop on Microsoft Platform
HdInsight essentials Hadoop on Microsoft Platform
 
Hd insight essentials quick view
Hd insight essentials quick viewHd insight essentials quick view
Hd insight essentials quick view
 
Hadoop Administration Online Training.pdf
Hadoop Administration Online Training.pdfHadoop Administration Online Training.pdf
Hadoop Administration Online Training.pdf
 
Learn Hadoop at your Leisure time
Learn Hadoop at your Leisure time Learn Hadoop at your Leisure time
Learn Hadoop at your Leisure time
 
Big data - Online Training
Big data - Online TrainingBig data - Online Training
Big data - Online Training
 
Deployment and Management of Hadoop Clusters
Deployment and Management of Hadoop ClustersDeployment and Management of Hadoop Clusters
Deployment and Management of Hadoop Clusters
 
hadoop exp
hadoop exphadoop exp
hadoop exp
 
SQL Server 2012 and Big Data
SQL Server 2012 and Big DataSQL Server 2012 and Big Data
SQL Server 2012 and Big Data
 
shaziaresume.
shaziaresume.shaziaresume.
shaziaresume.
 
Hadoop Operations - Best practices from the field
Hadoop Operations - Best practices from the fieldHadoop Operations - Best practices from the field
Hadoop Operations - Best practices from the field
 
Anujit CV
Anujit CV Anujit CV
Anujit CV
 
Anujit CV
Anujit CV Anujit CV
Anujit CV
 
Tcloud Computing Hadoop Family and Ecosystem Service 2013.Q3
Tcloud Computing Hadoop Family and Ecosystem Service 2013.Q3Tcloud Computing Hadoop Family and Ecosystem Service 2013.Q3
Tcloud Computing Hadoop Family and Ecosystem Service 2013.Q3
 
Infrastructure Around Hadoop
Infrastructure Around HadoopInfrastructure Around Hadoop
Infrastructure Around Hadoop
 

Manoj CV

  • 1. Manoj Kumar Sahu Call-08446783942 Manojsahu.a9@gmail.com SUMMARY • Over 4 year of IT experience in Hadoop/Bigdata and Linux Support Environment. • Having Knowledge of Cassandra administration, Application maintenance & support. • Having knowledge of Hadoop Administration with Nosql Cassandra,Hbase and Mongodb • Expertise with Hadoop and Cassandra tools like CM,Hortonwork Ops Centre. • Having knowledge on file System like Hadoop,Cassandra (CFS) and Linux,Unix. • Having Knowledge on Cassendra API ,Write Operation data and CQL. • Expertise on Hadoop Architect and Cassandra Architect. • Having knowledge on Cassandra Data Model, Keyspace, Interface and Cap Theorem. • Hands on Commissioning, Decommissioning, Bootstrapping and Balancing for Node. • Having knowledge on Network topology, Thrift, Avro and Gossip Protocols. • Handling on ATT &T Mobility project Datalake Hadoop with respective team. • Hands on experience installation and Management of Hadoop Cluster and Cassandra. • Hands on experience in Hadoop eco system Hive,Pig,Sqoop,Flume,Hbase installations. • Expertise with Hadoop Master and Slave daemons of HDFS and MapReduce. • Having knowledge in Rack awareness and Cassandra cluster Centre. • Hands on perform for Hadoop Distcp. • Installation of HadoopEco system and Apache Cassandra (1.0 x , 1.2 x& 2.0x ). • Hands on TWS (8.5.1) jobs monitoring and tracking using logs file. • Hands on experiences in Nagios and Ganglia monitoring tool. • Hands on Monitoring and Managing Hadoop respective daemons. • Hands on experience in install and configuration hadoop Cloudera Manager. • Monitoring the Cluster and deploying the daemons using CM, and Hortonwork. • Good knowledge on MR2 (Yarn), HA(High Availability),hadoop JVM • Good knowledge on Cluster Tuning and Node Benchmarking Processes. • Hands on experience in installation and Configuration of Linux Redhat,CentOS,Ubuntu. • Hands on Remote tools Super Putty and VMware workstation 10. • Having Knowledge on Linux command and setup the Server single Node and Multiple. • Creating Partitions Formating Mounting and change the Owner group and other permissions. • knowledge on LVM,Raid,Kick Start installation for linux • Creating repository local and online through Rpm and Yum package. • Good knowledge on linux NFS,FTP,LDAP.
  • 2. Software Skillset: • Hadoop: HDFS, MapReduce, Claudra Manager, Hortonwork. • HadoopEcosystem: Hive, Pig, Sqoop, Flume, Hbase, Zookeeper. • Security: Kerberos. • Cluster Management Tools: Cloudera Manager, OpsCenter, JMX, Nodetool. • NoSQL Databases: Cassandra. • Relational Databases: Microsoft SQL Server, MySQL. • Languages: C, Basic Java, C#, HTML PHP. • Operating Systems: Linux, Unix, Cent OS 6.4, Windows PROFESSIONAL EXPERIENCE: Amdocs | Pune India Mar, 2014 – Present Hadoop/ Ops Administrator Environment: Hadoop, SQL Server,DB,TWS9.2. Responsibilities: • As Hadoop & Cassandra Administrator support production server such as KM,STL, BHM, Dev and QA. • Monitor and handle application on TWS 9.2 production and Nagios Alerts. • Deploying Application on Hadoop Cluster. • Handling Gerick Framework and IF framework. • Manage Files Systems and Creating Operation Scripts • Source File monitoring and Run book Design package Reviews • Automated and deployed Cassandra environments using Chef recipes. • Evaluated, benchmarked and tuned data model by running endurance tests using JMeter, Cassandra Stress Tool and OpsCenter. • Used Cassandra node tool to manage Cassandra cluster. • Worked on optimizing the Cassandra cluster by making changes in Cassandra configuration file and Linux OS configurations. • Management Reporting and Communication and Development Support (Env Queries, New application Queries etc.) • Involved in writing multiple scripts to monitor Cassandra cluster and OpsCenter. • Distcp for Dual-Loaded sources and Monitoring of publishing feeds • Handling of on-request jobs/re-publishing feeds and Adhoc data movement request for QA Testing
  • 3. • Job failure report and Overview report maintenance , when new sources added • Deployment execution and code migrations and Long running job monitoring and reporting." Cloudwick Technology, INDIA 28 – Oct, 2011 Hadoop/ Linux Support Engineer Environment: Hadoop, SQL Server,DB,TWS9.2,Linux 6.4. Responsibilities: • Scheduled repair and cleanup process in production environment. • Administering and monitoring the cluster using tools • Scheduled moving production data to development environment for testing purpose. • Installed and configure mongodb server and mongo client package. • Exported the JSON data into Mongo db. • Install and configure Hive and Hive metadata and creating hive table to get result. • Scheduled repair and cleanup process in production environment during off peak times. • Exporting Mongodb data into HDFS. • Loading Hive data to transfer in to MSQL using Sqoop. • Experience in storing the analyzed results back into the Cassandra cluster. • Designed and configured gateway node to the cluster. • Performed Stress and Performance testing, benchmark for the cluster. • Installing Hadoop-Lzo Compression Using RPM and YUM repo. • Adding compression properties into Hadoop Configuration file • Exported the JSON data into Mongo db. • Exporting Mongodb data into HDFS. • Install and configure Hive and Hive metadata and creating hive table to get result. • Creating Lzo file and inserting into HDFS. • Creating Index and Running map-reduce Job . • Installing Hdfs and adding property into configuration file. • Adding the jdbc mysql connecter for Sqoop
  • 4. • Compresses data import into Mysql using Sqoop. • Add the property into flume configurations and load the streaming data into HDFS. • Running streaming data into Hadoop cluster and monitoring with Ganglia. • Monitoring the Health of Node using Nagios. • Hands on experience in installation and Configuration of Linux Redhat,Cent OS,Ubuntu • Hands on Remote tools VNC viewer,Teamviewer and VMware worksation 10. • Configuration and managing SSH keygen for cluster. • Basic knowledge on LVM,Raid, installestion for linux and LAMP. • Creating repository local and online through Rpm and Yum package.