MUKUL UPADHYAY
Tel: 01396 275232 ~ Mobile No.: 09911408870 ~ E-Mail: mukul_upadhyay2006@yahoo.co.in
Seeking assignments in Big Data Technology, Hadoop and Data Analysis with a technology driven organisation
of repute in IT – Software / Software Services sector
Professional Summery
 5 years 10 month of extensive IT experience with domestic and multinational clients which includes 4 years of
Hadoop related architecture experience developing Big Data / Hadoop applications.
 4 years of hands on experience in Hadoop Technology (Development and Administration) and its ecosystem like
Map-reduce Programming, Hive, HBase.
 Strong experience in architecture, design and development of Big Data platform including large clusters, Hadoop
ecosystem projects.
 Experience in custom map-reduce job using Java, User management, cluster management, Setup NOSQL
databases.
 Hands on experience in manual testing using MR unit.
 Experience in Amazon, Horton works and Cloudera Hadoop distributions.
 Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice -versa.
 Extending Hive functionality by writing custom UDFs.
 Tuning, and Monitoring Hadoop jobs and clusters in a production environment.
 Experience in installing, configuring, testing Hadoop ecosystem components.
 Experience on Core JAVA, JDBC.
 Expertise in managing Hadoop Services like Namenode, Datanode, Jobtracker, Tasktracker.
 Developing the Backup & Recovery strategies.
 Experienced with performing real time analytics on NoSQL databases like HBase.
 Familiar with RDBMS concepts and worked on MySql.
 Experience in job workflow scheduling and monitoring tool Oozie.
 Experienced with build tool Maven.
 Experience in all facets of Software Development Life Cycle (Analysis, Design, Development, Testing and
maintenance) using Waterfall and Agile methodologies.
 Highly adept at promptly and thoroughly mastering new technologies with a keen awareness of new industry
developments and the evolution of next generation programming solutions.
Organisational Exposure
Knowledge Momentum India : Since November ’2014 –till now
Designation: Team Lead (Development)
SKILL SET
Technical Skills
 Database : MySql
 Application Software : Jasper ,Eclipse, Netbeans, iReport, SVN, GitHub, Maven
 Language : Core Java
 Operating System : MS Windows ,Linux (Centos,Ubantu,Open Suse)
 Package : MS Office, Hadoop ,Hive, Open Source Solution,Amazone EC2, Apache Ambari,
HDP Sandbox2.3
Projects Executed
Project Title : OVALEDGE
Description: Ovaledge Inc. is an American company, working on a Hadoop based workbench for data analytics.
OvalEdge brings all your business data together in one unified Hadoop platform. With OvalEdge, it's easy to find,
understand and analyze important information in one place, and to use it to make faster, better‑informed decisions.
Client : OvalEgde Inc GA USA Nov 2014 to till continuing
Role : Team Lead (Hadoop Development)
Responsibilities:
 Expert in implementing advanced procedures like text analytics and processing using Hive Hooks.
 Worked on the core Hive Stack and customized own Hive Flavour (OvalEdge Hive).
 Developed and executed shell scripts to automate the jobs.
 Wrote complex Hive queries and UDFs.
 Worked o n reading multiple data formats on HDFS using Map-reduce and Apache Tika.
 Developed multiple POCs using Hadoop and deployed on the Yarn cluster.
 Involved in loading data from UNIX file system to HDFS.
 Extracted the data from various RDBMS into HDFS using Sqoop.
 Handled importing of data from various data sources, performed transformations using Hive, Map-reduce and
loaded data into HDFS.
 Manage and review Hadoop log files.
 Involved in analysis, design, testing phases and responsible for documenting technical specifications .
 Very good understanding ofPartitions,Bucketing concepts in Hive and designed both Managed and External tables
in Hive to optimize performance.
 Handling structured and unstructured data and applying ETL processes.
 Experienced in managing and reviewing Hadoop log files.
 Experienced in running Hadoop streaming jobs to process terabytes data.
 Involved in importing the real time data to hadoop using Kafka and implemented the Oozie job for daily imports.
Environment: Hadoop, HDFS, Hive, Map-reduce, Hive Hooks, HCatalog, Sqoop, Eclipse, UNIX Shell Scripting,
Hortonworks Data Platform (HDP2.2.4), Git
Project Title : Radius M2M
The purpose of the project is to store terabytes of log information generated by the company’s millions devices and
extract meaningful information outof it. The solution is based on the open source BigData s/w Hadoop. The data will be
stored in Hadoop file system and processed using Map/Reduce jobs which includes getting the raw byte or text data
from the devices, process the such file to obtain device’s stack and heartbeat information, extract various reports out of
the product various status information and export the information for further processing.
Client : Radius Synergies Pvt. Ltd. Jun 2012 to May 2014
Role : Hadoop Developer
Responsibilities:
 Worked on analyzing Hadoop cluster and different big data analytic tools including Map Reduce, Hive.
 Involved in loading data from LINUX file system, servers.
 Prepared Linux shell scripts for automating the process.
 Implemented map-reduce jobs for data analysis.
 Write Utility Software and Simulators for product testing.
 Implemented reports using Jasper Reports with Hive.
 Rendered and delivered reports in desired formats by using Jasper Report.
 Automated all the jobs from pulling data from Storage to loading data into MySQL using Shell Scripts
 Developed the Map Reduce programs to parse the raw data and store the pre Aggregated data in the portioned
tables.
 Involved in Installing, Configuring Hadoop Eco System using HDP(Hortonworks).
 Installed and Configured Hadoop cluster using Amazon Web Services (AWS) for POC purposes.
 Load and transform large sets ofstructured,semi structured,and unstructured data with map reduce, hive, and pig.
 Implemented Hive complex UDF's to execute business logic with Hive Queries.
 Exporting of result set from HIVE to MySQL using Sqoop export tool for further processing.
 Automation of all the jobs starting from pulling the Data from different Data Sources like MySQL and pushing th e
result dataset to Hadoop Distributed File System and running MR, PIG, and Hive jobs using Oozie (Work Flow
management)
 Worked with NoSQL databases like HBase in creating tables to load large sets ofsemi structured data coming from
various sources.
 Installed and configured Hive and also written Hive UDFs.
 Created partitioned tables in Hive, mentored analyst and test team for writing Hive Queries.
 Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS.
 Involved in cluster setup, monitoring, test benchmarks for results.
 Worked in agile methodology, attended daily standup meetings, Sprint plan meetings.
Environment: Hadoop,HDFS, Pig, Hive, Sqoop,Oozie, Hbase,My SQL, Linux CentOS, Eclipse,Hortonworks Data
Platform (HDP2.2.1)
Project Title : NANONET
NENO stands for Neutral Service Network Operations. This EMS application developed for the GPON (Gigabyte Passive
Optical Network) machine. Which provide various services on a single optical fibre line connection. This application
provides devices configurations, Network Optimization (Multiple Services) configurations, Maintenance of devices,
Component operational configurations.
Client : Radius Infratel Pvt. Ltd. Mar 2011 to Apr 2012
Role : Java developer
Responsibilities:
 Requirements gathering, analysis and design of GPON Components.
 Involved in Configurations Analysis, Architecture Design, Data Modelling, Security Modelling, Functional Design,
Specification review, Test case reviews and Programming.
 Developing GUI using Java Swing.
 Designing database tables for new modules.
 Assign various services to device’s port using configuration module.
 Participate in weekly project meetings and technical discussions.
 Testing the developed module before delivering to final Quality Control Assurance.
 Involved in configuring applications in various environments like Dev, QA and Production.
 Heavily involved in migration projects from Microsoft applications to Java platform .
 Involved in creating Security layer, Database layer and Presentation layer.
Environment: Core Java , Swing, NetBeans, Linux, MySql, Subversion(Tortoise).
Project Title MIS Radius
This application has developed for FTTH (Fibre to the Home) department.Which caters the needs of various reports and
tracking activities like DPR (Daily Progress Report), MPR (Monthly Progress Reports),Inventory Management, Society
Details, Customer Details and Report Generation.
Client : Radius Infratel Pvt. Ltd. Oct 2010 to Feb 2011
Role : Java developer
Responsibilities:
 Designed use cases, activities, states, objects and components.
 Developed the UI pages using HTML, Java script, Ajax, JSP.
 Developed front-end screens using JSP and Tag Libraries.
 Performing validations between various users.
 Design of Java Servlets and Objects using J2EE standards.
 Coded HTML, JSP and Servlets.
 Designing database tables for new modules.
 Created new connections through application coding for better access to MySql database, involved in writing SQL
queries and integration with Jasper Report.
 Implement generic data up-loader module using MS-Excel.
 Involved in testing and deploying in the development server.
 Involved in the design tables of the database in MySql. Involved in the design tables of the database in mySql.
Environment: Java, J2EE, Apache Tomcat, JSP, Servlets, Jasper iReport, MySql, Subversion(Tortoise).
Scholastics
 B-Tech (Computer Science & Engineering) from S.I.E.T. Meerut (U.P.T.U.) Meerut in 2010 (68%).
 10+2 from UP Board Allahabad 2004 (65 %).
 10th from UP Board Allahabad 2002 (66 %).
Personal Dossier
Date of Birth 11th
July 1987
Father’s Name Sh. Sunder Lal
Nationality Indian
Religion Hindu
Gender Male
Language Know. Hindi & English
Hobbies Experimental Cooking,Yoga
Permanent Address HN-640, Near Shalini Nursing Home , Saini Nagar Khatauli-251201(U.P.)
Correspondent Address 3/647, Sec-3, Vasundhra , Gaziabad-201012
Declaration
I hereby declare that the above information is correctto the bestof my knowledge.
Date :
Place : (Mukul Upadhyay)

Mukul-Resume

  • 1.
    MUKUL UPADHYAY Tel: 01396275232 ~ Mobile No.: 09911408870 ~ E-Mail: mukul_upadhyay2006@yahoo.co.in Seeking assignments in Big Data Technology, Hadoop and Data Analysis with a technology driven organisation of repute in IT – Software / Software Services sector Professional Summery  5 years 10 month of extensive IT experience with domestic and multinational clients which includes 4 years of Hadoop related architecture experience developing Big Data / Hadoop applications.  4 years of hands on experience in Hadoop Technology (Development and Administration) and its ecosystem like Map-reduce Programming, Hive, HBase.  Strong experience in architecture, design and development of Big Data platform including large clusters, Hadoop ecosystem projects.  Experience in custom map-reduce job using Java, User management, cluster management, Setup NOSQL databases.  Hands on experience in manual testing using MR unit.  Experience in Amazon, Horton works and Cloudera Hadoop distributions.  Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice -versa.  Extending Hive functionality by writing custom UDFs.  Tuning, and Monitoring Hadoop jobs and clusters in a production environment.  Experience in installing, configuring, testing Hadoop ecosystem components.  Experience on Core JAVA, JDBC.  Expertise in managing Hadoop Services like Namenode, Datanode, Jobtracker, Tasktracker.  Developing the Backup & Recovery strategies.  Experienced with performing real time analytics on NoSQL databases like HBase.  Familiar with RDBMS concepts and worked on MySql.  Experience in job workflow scheduling and monitoring tool Oozie.  Experienced with build tool Maven.  Experience in all facets of Software Development Life Cycle (Analysis, Design, Development, Testing and maintenance) using Waterfall and Agile methodologies.  Highly adept at promptly and thoroughly mastering new technologies with a keen awareness of new industry developments and the evolution of next generation programming solutions. Organisational Exposure Knowledge Momentum India : Since November ’2014 –till now Designation: Team Lead (Development) SKILL SET Technical Skills  Database : MySql  Application Software : Jasper ,Eclipse, Netbeans, iReport, SVN, GitHub, Maven  Language : Core Java  Operating System : MS Windows ,Linux (Centos,Ubantu,Open Suse)  Package : MS Office, Hadoop ,Hive, Open Source Solution,Amazone EC2, Apache Ambari, HDP Sandbox2.3 Projects Executed Project Title : OVALEDGE Description: Ovaledge Inc. is an American company, working on a Hadoop based workbench for data analytics. OvalEdge brings all your business data together in one unified Hadoop platform. With OvalEdge, it's easy to find, understand and analyze important information in one place, and to use it to make faster, better‑informed decisions. Client : OvalEgde Inc GA USA Nov 2014 to till continuing Role : Team Lead (Hadoop Development)
  • 2.
    Responsibilities:  Expert inimplementing advanced procedures like text analytics and processing using Hive Hooks.  Worked on the core Hive Stack and customized own Hive Flavour (OvalEdge Hive).  Developed and executed shell scripts to automate the jobs.  Wrote complex Hive queries and UDFs.  Worked o n reading multiple data formats on HDFS using Map-reduce and Apache Tika.  Developed multiple POCs using Hadoop and deployed on the Yarn cluster.  Involved in loading data from UNIX file system to HDFS.  Extracted the data from various RDBMS into HDFS using Sqoop.  Handled importing of data from various data sources, performed transformations using Hive, Map-reduce and loaded data into HDFS.  Manage and review Hadoop log files.  Involved in analysis, design, testing phases and responsible for documenting technical specifications .  Very good understanding ofPartitions,Bucketing concepts in Hive and designed both Managed and External tables in Hive to optimize performance.  Handling structured and unstructured data and applying ETL processes.  Experienced in managing and reviewing Hadoop log files.  Experienced in running Hadoop streaming jobs to process terabytes data.  Involved in importing the real time data to hadoop using Kafka and implemented the Oozie job for daily imports. Environment: Hadoop, HDFS, Hive, Map-reduce, Hive Hooks, HCatalog, Sqoop, Eclipse, UNIX Shell Scripting, Hortonworks Data Platform (HDP2.2.4), Git Project Title : Radius M2M The purpose of the project is to store terabytes of log information generated by the company’s millions devices and extract meaningful information outof it. The solution is based on the open source BigData s/w Hadoop. The data will be stored in Hadoop file system and processed using Map/Reduce jobs which includes getting the raw byte or text data from the devices, process the such file to obtain device’s stack and heartbeat information, extract various reports out of the product various status information and export the information for further processing. Client : Radius Synergies Pvt. Ltd. Jun 2012 to May 2014 Role : Hadoop Developer Responsibilities:  Worked on analyzing Hadoop cluster and different big data analytic tools including Map Reduce, Hive.  Involved in loading data from LINUX file system, servers.  Prepared Linux shell scripts for automating the process.  Implemented map-reduce jobs for data analysis.  Write Utility Software and Simulators for product testing.  Implemented reports using Jasper Reports with Hive.  Rendered and delivered reports in desired formats by using Jasper Report.  Automated all the jobs from pulling data from Storage to loading data into MySQL using Shell Scripts  Developed the Map Reduce programs to parse the raw data and store the pre Aggregated data in the portioned tables.  Involved in Installing, Configuring Hadoop Eco System using HDP(Hortonworks).  Installed and Configured Hadoop cluster using Amazon Web Services (AWS) for POC purposes.  Load and transform large sets ofstructured,semi structured,and unstructured data with map reduce, hive, and pig.  Implemented Hive complex UDF's to execute business logic with Hive Queries.  Exporting of result set from HIVE to MySQL using Sqoop export tool for further processing.  Automation of all the jobs starting from pulling the Data from different Data Sources like MySQL and pushing th e result dataset to Hadoop Distributed File System and running MR, PIG, and Hive jobs using Oozie (Work Flow management)  Worked with NoSQL databases like HBase in creating tables to load large sets ofsemi structured data coming from various sources.  Installed and configured Hive and also written Hive UDFs.  Created partitioned tables in Hive, mentored analyst and test team for writing Hive Queries.  Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS.  Involved in cluster setup, monitoring, test benchmarks for results.  Worked in agile methodology, attended daily standup meetings, Sprint plan meetings. Environment: Hadoop,HDFS, Pig, Hive, Sqoop,Oozie, Hbase,My SQL, Linux CentOS, Eclipse,Hortonworks Data Platform (HDP2.2.1)
  • 3.
    Project Title :NANONET NENO stands for Neutral Service Network Operations. This EMS application developed for the GPON (Gigabyte Passive Optical Network) machine. Which provide various services on a single optical fibre line connection. This application provides devices configurations, Network Optimization (Multiple Services) configurations, Maintenance of devices, Component operational configurations. Client : Radius Infratel Pvt. Ltd. Mar 2011 to Apr 2012 Role : Java developer Responsibilities:  Requirements gathering, analysis and design of GPON Components.  Involved in Configurations Analysis, Architecture Design, Data Modelling, Security Modelling, Functional Design, Specification review, Test case reviews and Programming.  Developing GUI using Java Swing.  Designing database tables for new modules.  Assign various services to device’s port using configuration module.  Participate in weekly project meetings and technical discussions.  Testing the developed module before delivering to final Quality Control Assurance.  Involved in configuring applications in various environments like Dev, QA and Production.  Heavily involved in migration projects from Microsoft applications to Java platform .  Involved in creating Security layer, Database layer and Presentation layer. Environment: Core Java , Swing, NetBeans, Linux, MySql, Subversion(Tortoise). Project Title MIS Radius This application has developed for FTTH (Fibre to the Home) department.Which caters the needs of various reports and tracking activities like DPR (Daily Progress Report), MPR (Monthly Progress Reports),Inventory Management, Society Details, Customer Details and Report Generation. Client : Radius Infratel Pvt. Ltd. Oct 2010 to Feb 2011 Role : Java developer Responsibilities:  Designed use cases, activities, states, objects and components.  Developed the UI pages using HTML, Java script, Ajax, JSP.  Developed front-end screens using JSP and Tag Libraries.  Performing validations between various users.  Design of Java Servlets and Objects using J2EE standards.  Coded HTML, JSP and Servlets.  Designing database tables for new modules.  Created new connections through application coding for better access to MySql database, involved in writing SQL queries and integration with Jasper Report.  Implement generic data up-loader module using MS-Excel.  Involved in testing and deploying in the development server.  Involved in the design tables of the database in MySql. Involved in the design tables of the database in mySql. Environment: Java, J2EE, Apache Tomcat, JSP, Servlets, Jasper iReport, MySql, Subversion(Tortoise). Scholastics  B-Tech (Computer Science & Engineering) from S.I.E.T. Meerut (U.P.T.U.) Meerut in 2010 (68%).  10+2 from UP Board Allahabad 2004 (65 %).  10th from UP Board Allahabad 2002 (66 %).
  • 4.
    Personal Dossier Date ofBirth 11th July 1987 Father’s Name Sh. Sunder Lal Nationality Indian Religion Hindu Gender Male Language Know. Hindi & English Hobbies Experimental Cooking,Yoga Permanent Address HN-640, Near Shalini Nursing Home , Saini Nagar Khatauli-251201(U.P.) Correspondent Address 3/647, Sec-3, Vasundhra , Gaziabad-201012 Declaration I hereby declare that the above information is correctto the bestof my knowledge. Date : Place : (Mukul Upadhyay)