SlideShare a Scribd company logo
Nagarjuna Damarla
Phone: +91-9941618664 e-mail: nagarjuna.bca@gmail.com
Professional Summary:
 3+ years of overall IT experience in Application Development using Big Data Hadoop, Java and
BI technologies.
 Proficient working experience in Hadoop components like HDFS, Map Reduce, Hive, Pig,
HBase, Sqoop and Flume.
 Good designing skills in writing Map Reduce programs.
 Involved in writing Pig and Hive scripts to reduce the job execution time.
 Able to handle export and import data to other database through Sqoop.
 Developed API to interact with MySQL data using java swings.
 Good communication, interpersonal, analytical skills, and strong ability to perform as part of
team.
 Interested in learning new concepts to keep updated in technology trends.
 Smart working and enthusiastic.
 Knowledge on FLUME and No-SQL Databases like Mongo DB.
 Got appreciations from clients and got Q1 Quarterly Award – 2015 for my contribution in
project.
Professional Experience:
 Programmer Analyst in Cognizant Technology Solutions, Hyderabad, India since 2013.
Education:
 Master of Computer Applications, Osmania University.
Technical Skills:
Skills BIG DATA MapReduce, Pig, Sqoop, Pig, Hive, Hbase, JAVA/J2EE
Technologies, Swings, JDBC, OJDBC
Frameworks Hadoop
Java IDEs Eclipse and NetBeans
Databases SQL, Mysql, MongoDB
Operating Systems Windows XP, 2000, 2003, Unix and Linux
Project Details:
Project #1
Title : Target Re-hosting of WebIntelligence Project
Environment : Hadoop , Apache Pig, Hive, SQOOP,Java , Unix , PHP , MySQL
Role : Hadoop Developer.
Hardware : Virtual Machines, UNIX.
Duration : March 2015 to Till Date
Description:
The purpose of the project is to store terabytes of log information generated by the ecommerce
website and extract meaningful information out of it. The solution is based on the open source BigData
s/w Hadoop .The data will be stored in Hadoop file system and processed using Map/Reduce jobs. Which
intern includes getting the raw html data from the websites ,Process the html to obtain product and pricing
information,Extract various reports out of the product pricing information and Export the information for
further processing.
This project is mainly for the re-platforming of the current existing system which is running on
WebHarvest a third party JAR and in MySQL DB to a new cloud solution technology called Hadoop which
can able to process large date sets (i.e. Tera bytes and Peta bytes of data) in order to meet the client
requirements with the incresing competion from his retailers.
Contributions:
1. Moved all crawl data flat files generated from various retailers to HDFS for further processing.
2. Written the Apache PIG scripts to process the HDFS data.
3. Created Hive tables to store the processed results in a tabular format.
4. Developed the sqoop scripts inorder to make the intraction between Pig and MySQL Database.
5. Involved in gathering the requirements, designing, development and testing
6. Writing CLI commands using HDFS.
7. Completely involved in Hadoop, hive, pig and mysql installation setup.
8. Analyzed log file to understand the user behavior.
9. Unit testing of MapReduce and Pig scripts.
Project #2
Title : Device Fault Predection
Environment : Hadoop , MapReduce, Hive, Sqoop, pig.
Role : Hadoop Developer.
Hardware : Virtual Machines, UNIX.
Duration : Jan 2014 – Feb 2015
Description:
Cisco’s support team on a day-to-day basis deals with huge volumes of issues related to their network
products like routers, switches etc. The support teams have been operating on a reactive model i.e.
based on the customer tickets/queries being raised. Hence, to improve customer satisfaction, they would
like the system to predict network faults based on the logs being generated by various network devices
i.e. by loading them into Hadoop cluster and analyzing them using some of the machine learning
algorithms implemented in Apache Mahout or custom built.
Responsibilities:
1. Moved all log files generated by various network devices into hdfs location
2. Written MapReduce code that will take input as log files and parse the logs and structure them in
tabular format to facilitate effective querying on the log data.
3. Created External Hive Table on top of parsed data
4. .Developed the sqoop scripts inorder to make the intraction between Pig and MySQL Database.
5. Involved in gathering the requirements, designing, development and testing
6. Analyzed log file to understand the user behavior.
7. Unit testing of MapReduce and Pig scripts.
Project #3
Title : Endeca iPlus 3.1
Environment : Endeca 3.1, SQL developer
Role : Endeca 3.1 developer
Hardware : Virtual Machines, UNIX.
Duration : Mar 2013– Dec 2013
Description:
Deliver the best Incentive System, to meet the needs of Customers and Dealers, provide a state of the art
system that will allow to be more efficient, flexible and capture increased sales, market share and profit ,
the existing SIMS R2.2 Business Processes have been modified, new functionalities and new reports
have been added. Business Intelligence has been introduced to provide better reporting solutions through
Endeca. Following are the key changes that have been implemented in ISYS.
Responsibilities:
1. Create Endeca pages with respective components like charts, results table, crosstabs.
2. Fine Tuning of queries, for the better performance.
3. Configure the instances according to the business rules and filter conditions.
4. Involving in Requirement gathering analysis.
5. Validate 3.1 endeca reports against 2.2.1 reports.
6. Prepared unit test cases for the reports.

More Related Content

What's hot

Bigdata.sunil_6+yearsExp
Bigdata.sunil_6+yearsExpBigdata.sunil_6+yearsExp
Bigdata.sunil_6+yearsExpbigdata sunil
 
Anil_BigData Resume
Anil_BigData ResumeAnil_BigData Resume
Anil_BigData ResumeAnil Sokhal
 
InfoSphere BigInsights - Analytics power for Hadoop - field experience
InfoSphere BigInsights - Analytics power for Hadoop - field experienceInfoSphere BigInsights - Analytics power for Hadoop - field experience
InfoSphere BigInsights - Analytics power for Hadoop - field experience
Wilfried Hoge
 
Shiv shakti resume
Shiv shakti resumeShiv shakti resume
Shiv shakti resume
Shiv Shakti
 
2014 10 09 Top reasons to use IBM BigInsights as your Big Data Hadoop system
2014 10 09 Top reasons to use IBM BigInsights as your Big Data Hadoop system2014 10 09 Top reasons to use IBM BigInsights as your Big Data Hadoop system
2014 10 09 Top reasons to use IBM BigInsights as your Big Data Hadoop systemToby Woolfe
 
Hadoop data access layer v4.0
Hadoop data access layer v4.0Hadoop data access layer v4.0
Hadoop data access layer v4.0
SpringPeople
 
Get Started Quickly with IBM's Hadoop as a Service
Get Started Quickly with IBM's Hadoop as a ServiceGet Started Quickly with IBM's Hadoop as a Service
Get Started Quickly with IBM's Hadoop as a Service
IBM Cloud Data Services
 
Bharath Hadoop Resume
Bharath Hadoop ResumeBharath Hadoop Resume
Bharath Hadoop ResumeBharath Kumar
 
Cascading User Group Meet
Cascading User Group MeetCascading User Group Meet
Cascading User Group Meet
Vinoth Kannan
 
Hadoop Distriubted File System (HDFS) presentation 27- 5-2015
Hadoop Distriubted File System (HDFS) presentation 27- 5-2015Hadoop Distriubted File System (HDFS) presentation 27- 5-2015
Hadoop Distriubted File System (HDFS) presentation 27- 5-2015
Abdul Nasir
 
Comparison among rdbms, hadoop and spark
Comparison among rdbms, hadoop and sparkComparison among rdbms, hadoop and spark
Comparison among rdbms, hadoop and spark
AgnihotriGhosh2
 
Big data and hadoop
Big data and hadoopBig data and hadoop
Big data and hadoop
Sri Kanth
 
Unlocking Operational Intelligence from the Data Lake
Unlocking Operational Intelligence from the Data LakeUnlocking Operational Intelligence from the Data Lake
Unlocking Operational Intelligence from the Data Lake
MongoDB
 
Marketing Digital Command Center
Marketing Digital Command CenterMarketing Digital Command Center
Marketing Digital Command Center
DataWorks Summit
 
SplunkLive! Hunk Technical Overview
SplunkLive! Hunk Technical OverviewSplunkLive! Hunk Technical Overview
SplunkLive! Hunk Technical OverviewSplunk
 
Big Data , Big Problem?
Big Data , Big Problem?Big Data , Big Problem?
Big Data , Big Problem?
Mohammadhasan Farazmand
 
NETWORK TRAFFIC ANALYSIS: HADOOP PIG VS TYPICAL MAPREDUCE
NETWORK TRAFFIC ANALYSIS: HADOOP PIG VS TYPICAL MAPREDUCENETWORK TRAFFIC ANALYSIS: HADOOP PIG VS TYPICAL MAPREDUCE
NETWORK TRAFFIC ANALYSIS: HADOOP PIG VS TYPICAL MAPREDUCE
csandit
 
Pouring the Foundation: Data Management in the Energy Industry
Pouring the Foundation: Data Management in the Energy IndustryPouring the Foundation: Data Management in the Energy Industry
Pouring the Foundation: Data Management in the Energy Industry
DataWorks Summit
 
Big data course
Big data  courseBig data  course
Big data course
kiruthikab6
 

What's hot (20)

Bigdata.sunil_6+yearsExp
Bigdata.sunil_6+yearsExpBigdata.sunil_6+yearsExp
Bigdata.sunil_6+yearsExp
 
Anil_BigData Resume
Anil_BigData ResumeAnil_BigData Resume
Anil_BigData Resume
 
InfoSphere BigInsights - Analytics power for Hadoop - field experience
InfoSphere BigInsights - Analytics power for Hadoop - field experienceInfoSphere BigInsights - Analytics power for Hadoop - field experience
InfoSphere BigInsights - Analytics power for Hadoop - field experience
 
Shiv shakti resume
Shiv shakti resumeShiv shakti resume
Shiv shakti resume
 
2014 10 09 Top reasons to use IBM BigInsights as your Big Data Hadoop system
2014 10 09 Top reasons to use IBM BigInsights as your Big Data Hadoop system2014 10 09 Top reasons to use IBM BigInsights as your Big Data Hadoop system
2014 10 09 Top reasons to use IBM BigInsights as your Big Data Hadoop system
 
Hadoop data access layer v4.0
Hadoop data access layer v4.0Hadoop data access layer v4.0
Hadoop data access layer v4.0
 
Get Started Quickly with IBM's Hadoop as a Service
Get Started Quickly with IBM's Hadoop as a ServiceGet Started Quickly with IBM's Hadoop as a Service
Get Started Quickly with IBM's Hadoop as a Service
 
Bharath Hadoop Resume
Bharath Hadoop ResumeBharath Hadoop Resume
Bharath Hadoop Resume
 
Cascading User Group Meet
Cascading User Group MeetCascading User Group Meet
Cascading User Group Meet
 
Hadoop Distriubted File System (HDFS) presentation 27- 5-2015
Hadoop Distriubted File System (HDFS) presentation 27- 5-2015Hadoop Distriubted File System (HDFS) presentation 27- 5-2015
Hadoop Distriubted File System (HDFS) presentation 27- 5-2015
 
Comparison among rdbms, hadoop and spark
Comparison among rdbms, hadoop and sparkComparison among rdbms, hadoop and spark
Comparison among rdbms, hadoop and spark
 
Prasanna Resume
Prasanna ResumePrasanna Resume
Prasanna Resume
 
Big data and hadoop
Big data and hadoopBig data and hadoop
Big data and hadoop
 
Unlocking Operational Intelligence from the Data Lake
Unlocking Operational Intelligence from the Data LakeUnlocking Operational Intelligence from the Data Lake
Unlocking Operational Intelligence from the Data Lake
 
Marketing Digital Command Center
Marketing Digital Command CenterMarketing Digital Command Center
Marketing Digital Command Center
 
SplunkLive! Hunk Technical Overview
SplunkLive! Hunk Technical OverviewSplunkLive! Hunk Technical Overview
SplunkLive! Hunk Technical Overview
 
Big Data , Big Problem?
Big Data , Big Problem?Big Data , Big Problem?
Big Data , Big Problem?
 
NETWORK TRAFFIC ANALYSIS: HADOOP PIG VS TYPICAL MAPREDUCE
NETWORK TRAFFIC ANALYSIS: HADOOP PIG VS TYPICAL MAPREDUCENETWORK TRAFFIC ANALYSIS: HADOOP PIG VS TYPICAL MAPREDUCE
NETWORK TRAFFIC ANALYSIS: HADOOP PIG VS TYPICAL MAPREDUCE
 
Pouring the Foundation: Data Management in the Energy Industry
Pouring the Foundation: Data Management in the Energy IndustryPouring the Foundation: Data Management in the Energy Industry
Pouring the Foundation: Data Management in the Energy Industry
 
Big data course
Big data  courseBig data  course
Big data course
 

Viewers also liked

Career@in solar insolare
Career@in solar   insolareCareer@in solar   insolare
Career@in solar insolare
Abhishek Sengupta
 
Free verse
Free verseFree verse
Free versekdelga2
 
Group motivation in open source
Group motivation in open sourceGroup motivation in open source
Group motivation in open sourceTom Hickerson
 
Изопроцессы Калинич Влада
Изопроцессы Калинич ВладаИзопроцессы Калинич Влада
Изопроцессы Калинич ВладаJimmy_ZigZag
 
Integration stories with OpenClinica and OpenXData
Integration stories with OpenClinica and OpenXDataIntegration stories with OpenClinica and OpenXData
Integration stories with OpenClinica and OpenXData
Tom Hickerson
 
Изопроцессы Калинич Влада
Изопроцессы Калинич ВладаИзопроцессы Калинич Влада
Изопроцессы Калинич ВладаJimmy_ZigZag
 
Stvorennya blogu
Stvorennya bloguStvorennya blogu
Free verse Hoops
Free verse HoopsFree verse Hoops
Free verse Hoopskdelga2
 
Searching Billions of Product Logs in Real Time (Use Case)
Searching Billions of Product Logs in Real Time (Use Case)Searching Billions of Product Logs in Real Time (Use Case)
Searching Billions of Product Logs in Real Time (Use Case)Ryan Tabora
 
газовые законы Калинич Влада
газовые законы Калинич Владагазовые законы Калинич Влада
газовые законы Калинич ВладаJimmy_ZigZag
 
Real Time Search and Analytics on Big Data
Real Time Search and Analytics on Big DataReal Time Search and Analytics on Big Data
Real Time Search and Analytics on Big DataRyan Tabora
 
газовые законы Калинич Влада
газовые законы Калинич Владагазовые законы Калинич Влада
газовые законы Калинич ВладаJimmy_ZigZag
 
The New Tech Stack for Device Data
The New Tech Stack for Device DataThe New Tech Stack for Device Data
The New Tech Stack for Device Data
Ryan Tabora
 
Poem recitation
Poem recitationPoem recitation
Poem recitationkdelga2
 
Free verse
Free verseFree verse
Free versekdelga2
 
Poempresentation
PoempresentationPoempresentation
Poempresentationkdelga2
 

Viewers also liked (17)

Career@in solar insolare
Career@in solar   insolareCareer@in solar   insolare
Career@in solar insolare
 
Free verse
Free verseFree verse
Free verse
 
Group motivation in open source
Group motivation in open sourceGroup motivation in open source
Group motivation in open source
 
Изопроцессы Калинич Влада
Изопроцессы Калинич ВладаИзопроцессы Калинич Влада
Изопроцессы Калинич Влада
 
Integration stories with OpenClinica and OpenXData
Integration stories with OpenClinica and OpenXDataIntegration stories with OpenClinica and OpenXData
Integration stories with OpenClinica and OpenXData
 
Изопроцессы Калинич Влада
Изопроцессы Калинич ВладаИзопроцессы Калинич Влада
Изопроцессы Калинич Влада
 
Stvorennya blogu
Stvorennya bloguStvorennya blogu
Stvorennya blogu
 
Free verse Hoops
Free verse HoopsFree verse Hoops
Free verse Hoops
 
Searching Billions of Product Logs in Real Time (Use Case)
Searching Billions of Product Logs in Real Time (Use Case)Searching Billions of Product Logs in Real Time (Use Case)
Searching Billions of Product Logs in Real Time (Use Case)
 
3D PRINTING
3D PRINTING3D PRINTING
3D PRINTING
 
газовые законы Калинич Влада
газовые законы Калинич Владагазовые законы Калинич Влада
газовые законы Калинич Влада
 
Real Time Search and Analytics on Big Data
Real Time Search and Analytics on Big DataReal Time Search and Analytics on Big Data
Real Time Search and Analytics on Big Data
 
газовые законы Калинич Влада
газовые законы Калинич Владагазовые законы Калинич Влада
газовые законы Калинич Влада
 
The New Tech Stack for Device Data
The New Tech Stack for Device DataThe New Tech Stack for Device Data
The New Tech Stack for Device Data
 
Poem recitation
Poem recitationPoem recitation
Poem recitation
 
Free verse
Free verseFree verse
Free verse
 
Poempresentation
PoempresentationPoempresentation
Poempresentation
 

Similar to Nagarjuna_Damarla_Resume

Nagarjuna_Damarla
Nagarjuna_DamarlaNagarjuna_Damarla
Nagarjuna_DamarlaNag Arjun
 
Sourav banerjee resume
Sourav banerjee   resumeSourav banerjee   resume
Sourav banerjee resume
Sourav Banerjee
 
Kalyan Hadoop
Kalyan HadoopKalyan Hadoop
Kalyan HadoopCanarys
 
Manikyam_Hadoop_5+Years
Manikyam_Hadoop_5+YearsManikyam_Hadoop_5+Years
Manikyam_Hadoop_5+YearsManikyam M
 
VAMSHI KRISHNA GADDAM IDRBT Experienced RESUME
VAMSHI KRISHNA GADDAM IDRBT Experienced RESUMEVAMSHI KRISHNA GADDAM IDRBT Experienced RESUME
VAMSHI KRISHNA GADDAM IDRBT Experienced RESUMEvamshi krishna
 
Shubham, 7.5+ years exp, mcp, map r spark-hive-bi-etl-azure-dataengineer-ml
Shubham, 7.5+ years exp, mcp, map r spark-hive-bi-etl-azure-dataengineer-mlShubham, 7.5+ years exp, mcp, map r spark-hive-bi-etl-azure-dataengineer-ml
Shubham, 7.5+ years exp, mcp, map r spark-hive-bi-etl-azure-dataengineer-ml
Shubham Mallick
 
IJSRED-V2I3P43
IJSRED-V2I3P43IJSRED-V2I3P43
IJSRED-V2I3P43
IJSRED
 
Maharshi_Amin_416
Maharshi_Amin_416Maharshi_Amin_416
Maharshi_Amin_416mamin1411
 
Rama prasad owk etl hadoop_developer
Rama prasad owk etl hadoop_developerRama prasad owk etl hadoop_developer
Rama prasad owk etl hadoop_developer
ramaprasad owk
 

Similar to Nagarjuna_Damarla_Resume (20)

Nagarjuna_Damarla
Nagarjuna_DamarlaNagarjuna_Damarla
Nagarjuna_Damarla
 
Resume (1)
Resume (1)Resume (1)
Resume (1)
 
Sourav banerjee resume
Sourav banerjee   resumeSourav banerjee   resume
Sourav banerjee resume
 
Pushpendra
PushpendraPushpendra
Pushpendra
 
Yasar resume 2
Yasar resume 2Yasar resume 2
Yasar resume 2
 
ChandraSekhar CV
ChandraSekhar CVChandraSekhar CV
ChandraSekhar CV
 
Robin_Hadoop
Robin_HadoopRobin_Hadoop
Robin_Hadoop
 
Kalyan Hadoop
Kalyan HadoopKalyan Hadoop
Kalyan Hadoop
 
Manikyam_Hadoop_5+Years
Manikyam_Hadoop_5+YearsManikyam_Hadoop_5+Years
Manikyam_Hadoop_5+Years
 
Resume_VipinKP
Resume_VipinKPResume_VipinKP
Resume_VipinKP
 
hadoop_bigdata
hadoop_bigdatahadoop_bigdata
hadoop_bigdata
 
VAMSHI KRISHNA GADDAM IDRBT Experienced RESUME
VAMSHI KRISHNA GADDAM IDRBT Experienced RESUMEVAMSHI KRISHNA GADDAM IDRBT Experienced RESUME
VAMSHI KRISHNA GADDAM IDRBT Experienced RESUME
 
Shubham, 7.5+ years exp, mcp, map r spark-hive-bi-etl-azure-dataengineer-ml
Shubham, 7.5+ years exp, mcp, map r spark-hive-bi-etl-azure-dataengineer-mlShubham, 7.5+ years exp, mcp, map r spark-hive-bi-etl-azure-dataengineer-ml
Shubham, 7.5+ years exp, mcp, map r spark-hive-bi-etl-azure-dataengineer-ml
 
hadoop resume
hadoop resumehadoop resume
hadoop resume
 
sudipto_resume
sudipto_resumesudipto_resume
sudipto_resume
 
IJSRED-V2I3P43
IJSRED-V2I3P43IJSRED-V2I3P43
IJSRED-V2I3P43
 
Neelima_Resume
Neelima_ResumeNeelima_Resume
Neelima_Resume
 
Mukul-Resume
Mukul-ResumeMukul-Resume
Mukul-Resume
 
Maharshi_Amin_416
Maharshi_Amin_416Maharshi_Amin_416
Maharshi_Amin_416
 
Rama prasad owk etl hadoop_developer
Rama prasad owk etl hadoop_developerRama prasad owk etl hadoop_developer
Rama prasad owk etl hadoop_developer
 

Nagarjuna_Damarla_Resume

  • 1. Nagarjuna Damarla Phone: +91-9941618664 e-mail: nagarjuna.bca@gmail.com Professional Summary:  3+ years of overall IT experience in Application Development using Big Data Hadoop, Java and BI technologies.  Proficient working experience in Hadoop components like HDFS, Map Reduce, Hive, Pig, HBase, Sqoop and Flume.  Good designing skills in writing Map Reduce programs.  Involved in writing Pig and Hive scripts to reduce the job execution time.  Able to handle export and import data to other database through Sqoop.  Developed API to interact with MySQL data using java swings.  Good communication, interpersonal, analytical skills, and strong ability to perform as part of team.  Interested in learning new concepts to keep updated in technology trends.  Smart working and enthusiastic.  Knowledge on FLUME and No-SQL Databases like Mongo DB.  Got appreciations from clients and got Q1 Quarterly Award – 2015 for my contribution in project. Professional Experience:  Programmer Analyst in Cognizant Technology Solutions, Hyderabad, India since 2013. Education:  Master of Computer Applications, Osmania University. Technical Skills: Skills BIG DATA MapReduce, Pig, Sqoop, Pig, Hive, Hbase, JAVA/J2EE Technologies, Swings, JDBC, OJDBC Frameworks Hadoop Java IDEs Eclipse and NetBeans Databases SQL, Mysql, MongoDB Operating Systems Windows XP, 2000, 2003, Unix and Linux Project Details: Project #1 Title : Target Re-hosting of WebIntelligence Project Environment : Hadoop , Apache Pig, Hive, SQOOP,Java , Unix , PHP , MySQL Role : Hadoop Developer. Hardware : Virtual Machines, UNIX. Duration : March 2015 to Till Date
  • 2. Description: The purpose of the project is to store terabytes of log information generated by the ecommerce website and extract meaningful information out of it. The solution is based on the open source BigData s/w Hadoop .The data will be stored in Hadoop file system and processed using Map/Reduce jobs. Which intern includes getting the raw html data from the websites ,Process the html to obtain product and pricing information,Extract various reports out of the product pricing information and Export the information for further processing. This project is mainly for the re-platforming of the current existing system which is running on WebHarvest a third party JAR and in MySQL DB to a new cloud solution technology called Hadoop which can able to process large date sets (i.e. Tera bytes and Peta bytes of data) in order to meet the client requirements with the incresing competion from his retailers. Contributions: 1. Moved all crawl data flat files generated from various retailers to HDFS for further processing. 2. Written the Apache PIG scripts to process the HDFS data. 3. Created Hive tables to store the processed results in a tabular format. 4. Developed the sqoop scripts inorder to make the intraction between Pig and MySQL Database. 5. Involved in gathering the requirements, designing, development and testing 6. Writing CLI commands using HDFS. 7. Completely involved in Hadoop, hive, pig and mysql installation setup. 8. Analyzed log file to understand the user behavior. 9. Unit testing of MapReduce and Pig scripts. Project #2 Title : Device Fault Predection Environment : Hadoop , MapReduce, Hive, Sqoop, pig. Role : Hadoop Developer. Hardware : Virtual Machines, UNIX. Duration : Jan 2014 – Feb 2015 Description: Cisco’s support team on a day-to-day basis deals with huge volumes of issues related to their network products like routers, switches etc. The support teams have been operating on a reactive model i.e. based on the customer tickets/queries being raised. Hence, to improve customer satisfaction, they would like the system to predict network faults based on the logs being generated by various network devices i.e. by loading them into Hadoop cluster and analyzing them using some of the machine learning algorithms implemented in Apache Mahout or custom built. Responsibilities: 1. Moved all log files generated by various network devices into hdfs location 2. Written MapReduce code that will take input as log files and parse the logs and structure them in tabular format to facilitate effective querying on the log data. 3. Created External Hive Table on top of parsed data 4. .Developed the sqoop scripts inorder to make the intraction between Pig and MySQL Database.
  • 3. 5. Involved in gathering the requirements, designing, development and testing 6. Analyzed log file to understand the user behavior. 7. Unit testing of MapReduce and Pig scripts. Project #3 Title : Endeca iPlus 3.1 Environment : Endeca 3.1, SQL developer Role : Endeca 3.1 developer Hardware : Virtual Machines, UNIX. Duration : Mar 2013– Dec 2013 Description: Deliver the best Incentive System, to meet the needs of Customers and Dealers, provide a state of the art system that will allow to be more efficient, flexible and capture increased sales, market share and profit , the existing SIMS R2.2 Business Processes have been modified, new functionalities and new reports have been added. Business Intelligence has been introduced to provide better reporting solutions through Endeca. Following are the key changes that have been implemented in ISYS. Responsibilities: 1. Create Endeca pages with respective components like charts, results table, crosstabs. 2. Fine Tuning of queries, for the better performance. 3. Configure the instances according to the business rules and filter conditions. 4. Involving in Requirement gathering analysis. 5. Validate 3.1 endeca reports against 2.2.1 reports. 6. Prepared unit test cases for the reports.