1. KARTHICK
EMail:skarthicece@gmail.com
Mobile: +919095318544
Objective
Intend to build a career with leading corporate of Hi-tech environment with committed and
dedicated people, which will help me to explore myself fully and realize my potential.
PROFESSIONAL EXPERIENCE:
Having 3.6 years of experience in Software Development, Involved in developing of
applications on Hadoop, Java on windows and Linux
Hands on experience in installing, configuring and using Apache Hadoop ecosystem
components like HDFS, Hadoop MapReduce, Apache Pig, Apache Hive, Apache
Sqoop, Oozie, and Neo4j, Apache Mahout .
Expertise in Neo4j - Cypher Query Language(CQL) for importing large data sets and
creating Nodes and Relationships
Strong Experience in Profile Recommendation by using Collaborative Filtering
Algorithm and making Bi-Directional Relationship in Neo4j
Analysed and Explored machine learning algorithm (Log Likelihood) and making
Profile Recommendation using Apache Mahout
Involved in running Hadoop jobs for processing millions of records of text data.
Analysed large data sets by running Pig scripts.
Designed and implemented Pig UDF's for evaluation, filtering, loading and storing of
data.
Developed multiple MapReduce jobs in java for data cleaning and pre-processing.
Involved in unit testing using MR unit for MapReduce jobs.
Involved in creating Hive tables, and loading and analysing data using hive queries.
The Hive tables created as per requirement were Internal or External tables defined
with appropriate Static and Dynamic partitions, intended for efficiency.
Experience in importing and exporting data using Sqoop from HDFS to Relational
Database Systems (RDBMS) and vice-versa.
Basic knowledge in Spark
Worked on large cluster having 200 nodes
Completed IBM certifications on BIG DATA
Completed 6 months course on Big Data Hadoop
Applications on Java, Hadoop on windows and Linux(Ubuntu)
2. Educational Profile
B.E (Electronics & Communication Engineering) 7.22 CGPA 2012
HSC 84% 2008
SSLC 84% 2006
Working Experience
Working as Software Engineer (Research and Development) for Matrimony.com Ltd
from July 2014 - till date
Worked on Hadoop Developer for SI System from June 2012 to July 2014
Project Details
Project Title: Viewed Profile also Viewed
Description
This is one of the portion which is present in each division of Matrimony.com site. Here,
people can see profiles as recommendation based on who are all viewed their own profile.
Role and Responsibility
Developer:
Installed and Configured the Neo4j Graph Database
Imported all the viewed history files into Neo4j graph database using Cypher Query
Language(CQL)
Learned and Converted Collaborative Filtering algorithm into Neo4j need
ExploredBi-Directional Relationship between viewer andviewed list of Idsusing CQL
Brought up Recommendation Ids from who are all viewed a particular Id using CQL
Tools: Neo4j-2.2.1
Project Title: Search Results
Description
In each division of Matrimony.com websites, there are some option to categorize the
searching results. People can Decline and Ignore profiles by their wish. Once they have done,
they could make fast searching and get new recommended profiles.
Role and Responsibility
3. Developer:
Installed and Configured Apache Hadoop and Apache Pig
Imported all the four category files into HDFS
Loaded all the files into Apache Pig
Joined all the category files based on the Filtration need in Apache Pig
Filtered all the Declined and Ignored Ids from the source and acceptance files
Tools: Apache Pig-0.13.0, Apache Hadoop-2.4.0
Project Title: Profile Recommendation
Description
This is Profile Recommendation to users when they are seeing any profile. This will give
profile as recommendation by collecting the user interest from the user viewed history and
bring results which matches their profile.
Role and Responsibility
Developer:
Installed and Configured the Apache Mahout with Apache Hadoop
Explored MapReduce program to make input file from viewed history which is
acceptable by Apache Mahout
Explored Log Likelihood machine learning algorithm in Mahout to make
Recommendation for each Id present in the viewed history file.
Tools: Apache Mahout-0.9, Apache Hadoop-1.2.1, MapReduce
Project: Customer Insights Platform
Client: Kohl’s Corporation
Duration: May 2013 to June 2014
Description
Kohl’s and its associated retail companies have very vast customer base both from Brick
and motor & online stores. With a generation capacity of around 10,000 transactions per
hour across all the entities, which comprises of 2.5 petabytes of data, Kohl’s is focused on
creating a central hub for customer specific information using BIGDATA technologies.
This project aims at providing analytics dashboard for business and data scientists for data
analysis and to generate new insights.
Role and Responsibility
Developer:
Working in an agile methodology, understand the requirements of the user stories and
prepare High-level design documentation for approval
4. On approval, derive Low level design documents for development
Develop the draft version of the scripts in Java Map reduce and Pig (Data transformation)
and HiveQL script (if it involves ad-hoc querying)
Fine tune the process based on the Map Reduce jobs processed
Using Apache Sqoop, the data from various database sources like Oracle, Teradata, DB2
and Informix is extracted to HDFS
Based on the requirement, schedule for each Job setup in Apache Oozie, Data from
multiple database sources loaded into Hive tables.
Apache Hive is being extensively used to build various tables, which are specific to user
requirement
Tools:
Hadoop (CDH4), HDFS, Sqoop, Hive, Map Reduce, Pig, Oozie
Hadoop: POC
Project: Web Log Analyzer
Description
On clicking the menus of the given organization the date, time, ip address from which the
application' URL was accessed are logged into the log file. The log file is moved to HDFS. The
HDFS file is processed using Map Reduce and Pig to find the maximum number of customer
who visited a particular URL. The result is stored in HBase and the report is displayed to the
user
Role and Responsibility
Developer:
Involved in requirements gathering, analysis
Understanding and designing the architecture based on the requirements
Storage of data in HDFS
Processing of data using Map Reduce and Pig
Processed result is stored in Hbase
Retrieval of reports from Hbase and displayed to user
Tools:
Hadoop (CDH4), HDFS, Sqoop, Hbase, Map Reduce, Pig, Oozie
Software Proficiency
Frameworks : Apache Hadoop, Mapreduce, Spark
Hadoop Echo Systems : Mahout, Pig, Hive, Hbase, Sqoop, Oozie
Programming Languages : Core Java
5. Database Connectivity : Neo4j, MySQL
Operating System : Linux(Ubuntu), Windows
Packages : WinSCP, Putty, MS office
IBM certifications on BIGDATA
Hadoop Fundamental
Introduction to Map Reduce Programming
Introduction to PIG
Accessing Hadoop Data using Hive
Controlling Hadoop Jobs using Oozie
Achievements
Course completed in Big Data- Hadoop
Course completed in Diploma in Computer Application
Course completed in Junior Diploma in Computer science.
KEY SOFT SKILL
Flexible and good in team work.
Listen to the feedback and work on it.
Hardworking, Trustworthy and Patient.
Personal Profile
Father’s Name Selvaraj.K
Sex Male
Marital Status Single
Date of Birth 22.11.1990
Languages Known English, Tamil
Declaration
I declare that the details given above are true to the best of my knowledge and
belief.
Karthick. S