1. “Transforming large, unruly data sets into competitive advantages”
To work in a professional and solution oriented environment where I can get enough opportunity to continuously innovate
and improve Software Products as well as myself, which in the long run will allow me to learn and establish processes and
standards.
Areas of Expertise
1 year of Big Data Ecosystem related technologies with development, implementation and deployment on Linux/Windows,
accompanying with experience in application development using Java/J2ee.
• Hands on experience on HDFS, HIVE, PIG, Hadoop Map Reduce framework and SQOOP.
• Worked extensively with HIVE DDLs and Hive Query language (HQLs).
• Developed UDF, UDAF, UDTF functions and implemented it in HIVE Queries.
• Developed PIG Latin scripts for handling business transformations.
• Implemented SQOOP for large dataset transfer between Hadoop and RDBMs.
• Hands on experience in Sequence files, RC files, Combiners, Counters, Dynamic Partitions, Bucketing for best
practice and performance improvement.
• Experience in setting up Hadoop on Pseudo distributed environment.
• Having Knowledge on SPARK and SCALA.
• Familiarity with common computing environment (e.g. Linux, Shell Scripting).
• Proficient in Java/J2EE application development.
• Good team player with ability to solve problems, organize and prioritize multiple tasks.
• Excellent communication and inter-personal skills.
Education
MCA (Master of Computer Applications) from Dr.L.B.P.G.College, Andhra University with an aggregate of 68% in 2009.
Technical Skills
Languages Java, Java Script, HTML, XML, XSD, XSL , Web Services, Map Reduce, Pig, Sqoop,
Hive, Hbase.
J2EE Technologies : JSP, Servlets, JDBC and EJB
Servers IBM Web Sphere Application Server 7.0, Web Logic and Tomcat
Frameworks Hadoop.
GANESH YEDUVAKA M o b i l e : + 9 1 9 0 3 2 9 1 9 0 6 5
Java & hadoop developer E - m a i i d : g a n 9 9 4 9 1 4 8 8 1 5 @ g m a i l . c o m
2. Java IDEs Eclipse.
Databases DB2 9.x, Oracle, SQL (DDL, DML, DCL) and PL/SQL.
Design Skills J2EE design patterns, Object Oriented Analysis and Design (OOAD), UML.
Operating Systems Windows7, Windows XP, windows 2000, Unix and Linux.
Professional Experience
INTERNATIONAL HOME SWAPPING Dec 2008
Java Developer
Astron IT Services Pvt. Ltd, Hyderabad, INDIA
“INTERNATIONAL HOME SWAPPING” is a Web Based Project, mainly deals with the exchange of homes of different people
across the countries, who wish to need hospitality during the tours for long time.
Modules:
The modules involved in the home swappers:
Admin Module:
The responsibility of the admin is to maintain the database and the web pages. He is responsible for adding the
authenticate members.
User(s) Module:
The user(s) should be two or many. Gets the required information from list placed in the internet. Requests the
authenticator, if he want any additional information.
Project Name : CMETRICA (Correlated METRIC Analysis)
Environment : Hadoop, Apache Pig, Hive, Cassandra, SQOOP, Flume, Java, UNIX, MySQL.
Role : Hadoop Developer.
Team Size : 8
Duration : Feb 2014 to Till Date
Description:
3. CMETRICA NxG platform is a domain-agnostic, configuration-based, Big Data processing platform. It supports real-time
as well as batch data processing and data analytics. This data exploratory platform of advanced analytic technique can
handle very large, diverse data sets that include different types such as structured/unstructured and batch. It has evolved
from its inception to support the need of CFO, Data Scientist, Analyst and several Business Users in various avatars
irrespective of the industry vertical or horizontal.
Roles and Responsibilities:
Completely involved in the requirement analysis phase.
Written different kinds of pig algorithms like Cube Generation, Summarization, Profiling, Ranking, Segmentation,
TopX, BottomX, Trending, Offer Management.
Written different UDF’s in PIG.
Created Hive tables to store the processed results in a tabular format.
Written different hive query’s to generate reports.
Developed the Unix shell scripts for creating the reports from Hive data.
Developed REST layer using Jersey to make available resource class methods.
Involved in implementation of DAO module.
Design and build Hadoop solutions for big data problems.
POC Handled
Project name : Sensex Info Processing System with Hadoop.
Environment : Hadoop 1.0.4-cdh3u5 Single Node Cluster running on Linux,
HDFS, Map Reduce, Apache Pig, JAVA, Hive, Sqoop.
Hardware : Virtual Machines.
Role : Hadoop Developer.
Description:
4. Sensex data in flat files which contains huge data consists of Daily and Weekly Sensex records. After storing it into HDFS,
based on clients conditions developed a Map Reduce Custom Partitioned program and with the resulted Reducer files
developed a Pig Script that separates Daily and Weekly data. With these two separated files data should be stored in Hive
External tables with names Daily and Weekly respectively. By developing Python script the data which is in Hive tables
will be loaded into Excel Sheets as Daily and weekly.
Roles and Responsibilities:
• Writing scripts to store data into HDFS.
• Developed Custom Partition Map Reduce program.
• Developed Pig Script to divide the records.
• Developed a UNIX Shell Script to run multiple commands one by one.
• Created Hive External tables to avoid data lose.
Project : Customer Valued Store(CVS)
Environment : JAVA, J2EE, Servlets, JSP, JDBC, and Tomcat, Oracle
Duration : Aug 2012 – Jan 2013
Description:
CVS Pharmacy is the second largest pharmacy chain after Walgreens in the United States, with more than 7,600 stores.
CVS sells prescription drugs and a wide assortment of general merchandise, including over-the-counter drugs, beauty
products and cosmetics
Responsibilities:
• Translate business needs into technical requirements and solutions.
• Develop code and implement change requests to fulfil client requirements.
• Perform code review, unit testing, debugging.
• Write batches to upload data from files to Data Base.