SlideShare a Scribd company logo
1 of 4
Anil Kumar
Phone:+918588899595
Mailto: anil.mmec@gmail.com
 Proficient in Hadoop and its ecosystem (MapReduce, Pig, Hive, Sqoop.etc) as a developer.
 Exposure to MAPR cluster and MCS.
 MapR certified Hadoop Professional: Developer (MCHP: D).
 About 3 years of experience in IT industry aligned to perform the tasks assigned deliver value and satisfy
customer.
 Expertise in Core Java
 Proficient communication skills, professional outlook towards knowledge sharing and team work and mostly
worked in customer facing roles.
Role Responsibilities
 Interactwiththe clienttogather requirements.
 Involvedinidentifyingthe use casesanddesign.
 InstalledandconfiguredMapRHadoopclustermultiple times.
 Involvedin MapRclustermaintenance andconfigurationtasks.
 DevelopedGenericMapReduce Algorithmsthatcanbe usedacrossvarietyof problems.
 Implemented various concepts (like: Joins, Secondary sort) in Map Reduce applications using Java.
 WrittenHive queries foranalytics.
Environment
Hadoop, HDFS, MapR, MapReduce, R, SpakSQL, Pig, Hive, Ansible, Sqoop, Core Java, Linux, Window 7 and AWS
Technical Skills
Software MAPR, MCS, Map-Reduce, HDFS, Hadoop Ecosystem (Hive, Pig, Sqoop ), R, AWS,
Linux, Windows XP, Windows 7, MySQL ,
Hardware Intel x86 architecture machine
Tools Eclipse Galileo, Eclipse Helios, Apache Tomcat 6.0.33
SDLC Methodology Water Fall ,Agile
Domain Experience Java /J2EE, BigData Technologies.
Business
Development
Approach towards feasible and qualitative business development.
Achievements
.
 Got “STAR PERFORMAR AWARD” from “DESS INDIA HEAD” for continues learning attitude and Dedication
shown in deliverables.
 Received Appreciation Mails from “American Express” for hard work and Dedication shown in deliverables.
 Received Kudos from TCS Leadership during ILP.
 Consistently enhancing competencies in the field of Big-Data - Hadoop Map-Reduce, PIG and HIVE.
 Done IBM Big data university certifications in Hadoop, Mapreduce, Hive, Pig, and Data transfer tools.
 Teaching experience during M.Tech from NIT Kurukshetra.
 Secured 98 percentile in GATE 2010.
TCS Experience Summary
Project BigData Charge Back, Pfizer: Pfizer Inc. is an American multinational pharmaceutical
corporation headquartered in New York City, New York, with its research headquarters
in Groton, Connecticut. It is among one of the world's largest pharmaceutical companies
by revenues.
Role Hadoop Developer and Cluster Maintenance
Period Mar’15 to current
Tools & Technologies MAPR Cluster Installation, MapR Upgrade, MCS, MapR cluster maintenance and
configuration tasks, Basics of R Programming, R Studio, Shell Scripts, SpakSQL,
Ansible, Ipython, Zeppelin, Hive and Hue.
Description  Architecting and evaluating various technical components for Data Lake
implementation in Pfizer.
 Pfizer’s vision is to provide Business Analytics & Insights (BAI) with collaborative,
global data discovery, and analytics data platform and tools.
 Installed and configured MapR Hadoop cluster with 5 nodes multiple times.
 Automated Shell scripts to carry the MapR cluster maintenance and configuration
tasks (eg: service restart, volume management).
 Upgrading MapR from 4.0.1 to 4.0.2.
 Executing Teragen, TeraSort, DFSIO for MapR and Spark cluster.
 MapR cluster deployment on AWS using customized playbooks of Ansible.
 Configuring ecosystem components in the cluster.
 Configuring web UI for Spark.
 Configuring Spark on YARN.
 Configuring Data Wrangling tools on the cluster
 Cluster monitoring using Ganglia.
Project Out Of Pattern Analysis for Amex client
Role Hadoop Developer
Period Apr'14 to Feb'15
Technology Stack Operating System : Linux
Programming Language : Java, JDBC
Hadoop Vendor : MapR 3.0.2
Data Storage : HDFS, SQLServer2005, Hive
Data Processing : MapReduce
Data Access : Hive, Sqoop
Description  Contributed as Hadoop developer in Big Data Project for leading banking client.
 Worked on a cluster of 600 nodes.
 Proposed different Use cases and closely involved in Requirement gathering with
BA.
 The project scope cover, handling huge structured (CSV, RDBMS) and semi
structured data (XML files ) stored in Mainframe System.
 FTP Client was designed in Java to FTP data from Mainframe to Hadoop
Platform
 Generic MapReduce programs were designed for Data Cleansing and Data
Filtering.
 Statistics were calculated on raw data using control file through Mapreduce.
 If anything come out of pattern than need to generate alert on Stats Using
hadoop System..
 Data was loaded into Hive tables for further Analysis.
Project Analytics, Big Data and Information Management for TCS
Role Hadoop Developer
Period Jan’2014 –Mar’2014
Technology Stack Operating System : Centos 6.4 (64 bit)
Programming Language : Java
Hadoop Vendor : Cloudera CDH4u0 with Cloudera Manager 4.0
Data Storage : HDFS, MySQL
Data Processing : MapReduce 1.0
Data Access :Pig, Hive, Sqoop
Visualization : Tableau
Description of Project
Activities
 Worked on internal case studies such as credit card fraud analysis, where we
need to find out the fraudulent transactions done with a credit card.
 A Cluster of 7 nodes was setup using Cloudera Manager 4.0.
 Various data sources & formats like csv, xml, flat files and RDBMS were used to
provide different parts the data.
 The data from different sources was transferred to HDFS through SQOOP and
Java Hadoop API.
 Filtration part of junk and bad record was performed through pig scripts.
 HIVE queries were used to perform analysis.
 Visual analysis was implemented using TABLEAU.
Project Analytics, Big Data and Information Management for TCS
Role Hadoop Developer
Period Nov’2013 –Dec’2013
Technology Stack Operating System : Centos 6.4 (64 bit)
Programming Language : Java
Hadoop Vendor : Cloudera CDH3U4 with Cloudera Manager 4.0
Data Storage : HDFS, MySQL
Data Processing : MapReduce 1.0
Data Access :Pig, Hive, Sqoop
Visualization : Tableau
Description of Project
Activities
 Worked on internal case studies such as Benchmarking MapReduce framework
with respect to data processing and then analyze data to find hidden trend.
 A Cluster of 7 nodes was setup using Cloudera Manager 4.0
 Various data sources like csv, xml and RDBMS were used as data source.
 The data from different sources was transferred to HDFS through SQOOP.
 Implemented MapReduce on different size of data.
 We Increased and decreased number of mappers and reducers at run time.
 Checked the performance by introducing concept of combiner and partitioner.
 Filtration of junk and bad records was done using Pig Scripts.
 Hive Queries were used to perform analysis.
 Visual Analysis was done using Tableau.
Project Initial Learning Program
Role Team Leader
Technology Stack Operating System : Window 7
Programming Language: Java, J2EE, Html, and Oracle.
Achievement Got ILP Kudos Award at ILP Hyderabad.
Description E-Recruitment website that will be enable the establishment of a streamlined, consistent
recruitment process that reduce manual effort and improve the experience for applicants
and for staff in faculties and division seeking to fill vacant position.
Experience Details
TCS Experience 2 Year(s),10 Month(s)
Prev. Experience 0 Year(s),0 Month(s)
Total Experience 2 Year(s),10 Month(s)
Education Summary
Qualification Category College Subject
Bachelor Of Technology MMEC Mullana Information Technology
Master Of Technology NIT Kurukshetra Computer Engineering

More Related Content

What's hot

What's hot (19)

Hadoop Big Data Resume
Hadoop Big Data ResumeHadoop Big Data Resume
Hadoop Big Data Resume
 
Resume (2)
Resume (2)Resume (2)
Resume (2)
 
Resume
ResumeResume
Resume
 
Sasmita bigdata resume
Sasmita bigdata resumeSasmita bigdata resume
Sasmita bigdata resume
 
kishore resume hadoop
kishore resume hadoopkishore resume hadoop
kishore resume hadoop
 
hadoop exp
hadoop exphadoop exp
hadoop exp
 
Resume_VipinKP
Resume_VipinKPResume_VipinKP
Resume_VipinKP
 
Suresh_Yadav_Hadoop_Fresher_Resume
Suresh_Yadav_Hadoop_Fresher_ResumeSuresh_Yadav_Hadoop_Fresher_Resume
Suresh_Yadav_Hadoop_Fresher_Resume
 
Rajeev kumar apache_spark & scala developer
Rajeev kumar apache_spark & scala developerRajeev kumar apache_spark & scala developer
Rajeev kumar apache_spark & scala developer
 
Resume
ResumeResume
Resume
 
Borja González - Resume ​Big Data Architect
Borja González - Resume ​Big Data ArchitectBorja González - Resume ​Big Data Architect
Borja González - Resume ​Big Data Architect
 
Pallavi_Resume
Pallavi_ResumePallavi_Resume
Pallavi_Resume
 
BigData_Krishna Kumar Sharma
BigData_Krishna Kumar SharmaBigData_Krishna Kumar Sharma
BigData_Krishna Kumar Sharma
 
ATHOKPAM NABAKUMAR SINGH's HADOOP ADMIN
ATHOKPAM NABAKUMAR SINGH's HADOOP ADMINATHOKPAM NABAKUMAR SINGH's HADOOP ADMIN
ATHOKPAM NABAKUMAR SINGH's HADOOP ADMIN
 
Owez_IBM_Hadoop_Admin
Owez_IBM_Hadoop_AdminOwez_IBM_Hadoop_Admin
Owez_IBM_Hadoop_Admin
 
HimaBindu
HimaBinduHimaBindu
HimaBindu
 
Nagarjuna_Damarla
Nagarjuna_DamarlaNagarjuna_Damarla
Nagarjuna_Damarla
 
Vijay_hadoop admin
Vijay_hadoop adminVijay_hadoop admin
Vijay_hadoop admin
 
Amith_Hadoop_Admin_CV
Amith_Hadoop_Admin_CVAmith_Hadoop_Admin_CV
Amith_Hadoop_Admin_CV
 

Similar to Anil_BigData Resume

Sanath pabba hadoop resume 1.0
Sanath pabba hadoop resume 1.0Sanath pabba hadoop resume 1.0
Sanath pabba hadoop resume 1.0Pabba Gupta
 
Sunshine consulting mopuru babu cv_java_j2ee_spring_bigdata_scala
Sunshine consulting mopuru babu cv_java_j2ee_spring_bigdata_scalaSunshine consulting mopuru babu cv_java_j2ee_spring_bigdata_scala
Sunshine consulting mopuru babu cv_java_j2ee_spring_bigdata_scalaMopuru Babu
 
Manikyam_Hadoop_5+Years
Manikyam_Hadoop_5+YearsManikyam_Hadoop_5+Years
Manikyam_Hadoop_5+YearsManikyam M
 
sam_resume - updated
sam_resume - updatedsam_resume - updated
sam_resume - updatedsam k
 
HariKrishna4+_cv
HariKrishna4+_cvHariKrishna4+_cv
HariKrishna4+_cvrevuri
 
Nagarjuna_Damarla_Resume
Nagarjuna_Damarla_ResumeNagarjuna_Damarla_Resume
Nagarjuna_Damarla_ResumeNag Arjun
 
Srikanth hadoop 3.6yrs_hyd
Srikanth hadoop 3.6yrs_hydSrikanth hadoop 3.6yrs_hyd
Srikanth hadoop 3.6yrs_hydsrikanth K
 
Vishnu_HadoopDeveloper
Vishnu_HadoopDeveloperVishnu_HadoopDeveloper
Vishnu_HadoopDevelopervishnu ch
 
Big Data with hadoop, Spark and BigQuery (Google cloud next Extended 2017 Kar...
Big Data with hadoop, Spark and BigQuery (Google cloud next Extended 2017 Kar...Big Data with hadoop, Spark and BigQuery (Google cloud next Extended 2017 Kar...
Big Data with hadoop, Spark and BigQuery (Google cloud next Extended 2017 Kar...Imam Raza
 

Similar to Anil_BigData Resume (20)

Madhu
MadhuMadhu
Madhu
 
Resume (1)
Resume (1)Resume (1)
Resume (1)
 
RESUME_N
RESUME_NRESUME_N
RESUME_N
 
sudipto_resume
sudipto_resumesudipto_resume
sudipto_resume
 
Sanath pabba hadoop resume 1.0
Sanath pabba hadoop resume 1.0Sanath pabba hadoop resume 1.0
Sanath pabba hadoop resume 1.0
 
Sunshine consulting mopuru babu cv_java_j2ee_spring_bigdata_scala
Sunshine consulting mopuru babu cv_java_j2ee_spring_bigdata_scalaSunshine consulting mopuru babu cv_java_j2ee_spring_bigdata_scala
Sunshine consulting mopuru babu cv_java_j2ee_spring_bigdata_scala
 
Sureh hadoop 3 years t
Sureh hadoop 3 years tSureh hadoop 3 years t
Sureh hadoop 3 years t
 
Manikyam_Hadoop_5+Years
Manikyam_Hadoop_5+YearsManikyam_Hadoop_5+Years
Manikyam_Hadoop_5+Years
 
Nagesh Hadoop Profile
Nagesh Hadoop ProfileNagesh Hadoop Profile
Nagesh Hadoop Profile
 
sam_resume - updated
sam_resume - updatedsam_resume - updated
sam_resume - updated
 
Prashanth Kumar_Hadoop_NEW
Prashanth Kumar_Hadoop_NEWPrashanth Kumar_Hadoop_NEW
Prashanth Kumar_Hadoop_NEW
 
HariKrishna4+_cv
HariKrishna4+_cvHariKrishna4+_cv
HariKrishna4+_cv
 
Nagarjuna_Damarla_Resume
Nagarjuna_Damarla_ResumeNagarjuna_Damarla_Resume
Nagarjuna_Damarla_Resume
 
hadoop resume
hadoop resumehadoop resume
hadoop resume
 
Nikhil Sinha.
Nikhil Sinha.Nikhil Sinha.
Nikhil Sinha.
 
Poorna Hadoop
Poorna HadoopPoorna Hadoop
Poorna Hadoop
 
Srikanth hadoop 3.6yrs_hyd
Srikanth hadoop 3.6yrs_hydSrikanth hadoop 3.6yrs_hyd
Srikanth hadoop 3.6yrs_hyd
 
Resume - Narasimha Rao B V (TCS)
Resume - Narasimha  Rao B V (TCS)Resume - Narasimha  Rao B V (TCS)
Resume - Narasimha Rao B V (TCS)
 
Vishnu_HadoopDeveloper
Vishnu_HadoopDeveloperVishnu_HadoopDeveloper
Vishnu_HadoopDeveloper
 
Big Data with hadoop, Spark and BigQuery (Google cloud next Extended 2017 Kar...
Big Data with hadoop, Spark and BigQuery (Google cloud next Extended 2017 Kar...Big Data with hadoop, Spark and BigQuery (Google cloud next Extended 2017 Kar...
Big Data with hadoop, Spark and BigQuery (Google cloud next Extended 2017 Kar...
 

Anil_BigData Resume

  • 1. Anil Kumar Phone:+918588899595 Mailto: anil.mmec@gmail.com  Proficient in Hadoop and its ecosystem (MapReduce, Pig, Hive, Sqoop.etc) as a developer.  Exposure to MAPR cluster and MCS.  MapR certified Hadoop Professional: Developer (MCHP: D).  About 3 years of experience in IT industry aligned to perform the tasks assigned deliver value and satisfy customer.  Expertise in Core Java  Proficient communication skills, professional outlook towards knowledge sharing and team work and mostly worked in customer facing roles. Role Responsibilities  Interactwiththe clienttogather requirements.  Involvedinidentifyingthe use casesanddesign.  InstalledandconfiguredMapRHadoopclustermultiple times.  Involvedin MapRclustermaintenance andconfigurationtasks.  DevelopedGenericMapReduce Algorithmsthatcanbe usedacrossvarietyof problems.  Implemented various concepts (like: Joins, Secondary sort) in Map Reduce applications using Java.  WrittenHive queries foranalytics. Environment Hadoop, HDFS, MapR, MapReduce, R, SpakSQL, Pig, Hive, Ansible, Sqoop, Core Java, Linux, Window 7 and AWS Technical Skills Software MAPR, MCS, Map-Reduce, HDFS, Hadoop Ecosystem (Hive, Pig, Sqoop ), R, AWS, Linux, Windows XP, Windows 7, MySQL , Hardware Intel x86 architecture machine Tools Eclipse Galileo, Eclipse Helios, Apache Tomcat 6.0.33 SDLC Methodology Water Fall ,Agile Domain Experience Java /J2EE, BigData Technologies. Business Development Approach towards feasible and qualitative business development. Achievements .  Got “STAR PERFORMAR AWARD” from “DESS INDIA HEAD” for continues learning attitude and Dedication shown in deliverables.  Received Appreciation Mails from “American Express” for hard work and Dedication shown in deliverables.  Received Kudos from TCS Leadership during ILP.  Consistently enhancing competencies in the field of Big-Data - Hadoop Map-Reduce, PIG and HIVE.  Done IBM Big data university certifications in Hadoop, Mapreduce, Hive, Pig, and Data transfer tools.  Teaching experience during M.Tech from NIT Kurukshetra.  Secured 98 percentile in GATE 2010.
  • 2. TCS Experience Summary Project BigData Charge Back, Pfizer: Pfizer Inc. is an American multinational pharmaceutical corporation headquartered in New York City, New York, with its research headquarters in Groton, Connecticut. It is among one of the world's largest pharmaceutical companies by revenues. Role Hadoop Developer and Cluster Maintenance Period Mar’15 to current Tools & Technologies MAPR Cluster Installation, MapR Upgrade, MCS, MapR cluster maintenance and configuration tasks, Basics of R Programming, R Studio, Shell Scripts, SpakSQL, Ansible, Ipython, Zeppelin, Hive and Hue. Description  Architecting and evaluating various technical components for Data Lake implementation in Pfizer.  Pfizer’s vision is to provide Business Analytics & Insights (BAI) with collaborative, global data discovery, and analytics data platform and tools.  Installed and configured MapR Hadoop cluster with 5 nodes multiple times.  Automated Shell scripts to carry the MapR cluster maintenance and configuration tasks (eg: service restart, volume management).  Upgrading MapR from 4.0.1 to 4.0.2.  Executing Teragen, TeraSort, DFSIO for MapR and Spark cluster.  MapR cluster deployment on AWS using customized playbooks of Ansible.  Configuring ecosystem components in the cluster.  Configuring web UI for Spark.  Configuring Spark on YARN.  Configuring Data Wrangling tools on the cluster  Cluster monitoring using Ganglia. Project Out Of Pattern Analysis for Amex client Role Hadoop Developer Period Apr'14 to Feb'15 Technology Stack Operating System : Linux Programming Language : Java, JDBC Hadoop Vendor : MapR 3.0.2 Data Storage : HDFS, SQLServer2005, Hive Data Processing : MapReduce Data Access : Hive, Sqoop Description  Contributed as Hadoop developer in Big Data Project for leading banking client.  Worked on a cluster of 600 nodes.  Proposed different Use cases and closely involved in Requirement gathering with BA.  The project scope cover, handling huge structured (CSV, RDBMS) and semi structured data (XML files ) stored in Mainframe System.  FTP Client was designed in Java to FTP data from Mainframe to Hadoop Platform  Generic MapReduce programs were designed for Data Cleansing and Data Filtering.  Statistics were calculated on raw data using control file through Mapreduce.
  • 3.  If anything come out of pattern than need to generate alert on Stats Using hadoop System..  Data was loaded into Hive tables for further Analysis. Project Analytics, Big Data and Information Management for TCS Role Hadoop Developer Period Jan’2014 –Mar’2014 Technology Stack Operating System : Centos 6.4 (64 bit) Programming Language : Java Hadoop Vendor : Cloudera CDH4u0 with Cloudera Manager 4.0 Data Storage : HDFS, MySQL Data Processing : MapReduce 1.0 Data Access :Pig, Hive, Sqoop Visualization : Tableau Description of Project Activities  Worked on internal case studies such as credit card fraud analysis, where we need to find out the fraudulent transactions done with a credit card.  A Cluster of 7 nodes was setup using Cloudera Manager 4.0.  Various data sources & formats like csv, xml, flat files and RDBMS were used to provide different parts the data.  The data from different sources was transferred to HDFS through SQOOP and Java Hadoop API.  Filtration part of junk and bad record was performed through pig scripts.  HIVE queries were used to perform analysis.  Visual analysis was implemented using TABLEAU. Project Analytics, Big Data and Information Management for TCS Role Hadoop Developer Period Nov’2013 –Dec’2013 Technology Stack Operating System : Centos 6.4 (64 bit) Programming Language : Java Hadoop Vendor : Cloudera CDH3U4 with Cloudera Manager 4.0 Data Storage : HDFS, MySQL Data Processing : MapReduce 1.0 Data Access :Pig, Hive, Sqoop Visualization : Tableau Description of Project Activities  Worked on internal case studies such as Benchmarking MapReduce framework with respect to data processing and then analyze data to find hidden trend.  A Cluster of 7 nodes was setup using Cloudera Manager 4.0  Various data sources like csv, xml and RDBMS were used as data source.  The data from different sources was transferred to HDFS through SQOOP.  Implemented MapReduce on different size of data.
  • 4.  We Increased and decreased number of mappers and reducers at run time.  Checked the performance by introducing concept of combiner and partitioner.  Filtration of junk and bad records was done using Pig Scripts.  Hive Queries were used to perform analysis.  Visual Analysis was done using Tableau. Project Initial Learning Program Role Team Leader Technology Stack Operating System : Window 7 Programming Language: Java, J2EE, Html, and Oracle. Achievement Got ILP Kudos Award at ILP Hyderabad. Description E-Recruitment website that will be enable the establishment of a streamlined, consistent recruitment process that reduce manual effort and improve the experience for applicants and for staff in faculties and division seeking to fill vacant position. Experience Details TCS Experience 2 Year(s),10 Month(s) Prev. Experience 0 Year(s),0 Month(s) Total Experience 2 Year(s),10 Month(s) Education Summary Qualification Category College Subject Bachelor Of Technology MMEC Mullana Information Technology Master Of Technology NIT Kurukshetra Computer Engineering