1. DEEPESH REHI
Email: rehi.deepesh@gmail.com +91 9962 726 080
Professional Summary:
• IT services experience with exclusive focus on Hadoop, Scala/Spark technologies and integrating
the Applications for BFSI and Communications domain in Cognizant Technology Solutions.
• Experience in designing & developing application based on Big data technologies using Hadoop
framework using Spark, Hive, Scala, Sqoop, Flume.
• Experience in working with Scala with maven building spark jobs.
• Have involved in several POC development for getting new projects for organization.
• Experience in projects under core-publishing, banking, telecommunication solutions.
• Experience in working with tools- hortonworks, cloudera framework.
• Experience in working in Unix environment with languages like java.
• Experience in sql databases like MySql and also working with NoSql databases like Hbase and
neo4j.
Technical Skills:
Big data Technologies Hadoop, Spark, Hive, Sqoop, Flume, Hbase, Neo4j
Programming Language Scala, Core java, sql, Unix shell scripting
Tools and Utilities Cloudera, Hortonworks, eclipse, MobaXterm, winscp, Solr, putty
Database Oracle, DB2, mySql Server
Source Control Git
Operating system Windows XP/2000/7, Unix
Professional Work Experience:
Client: AAA Insurance May‘16- Till
Date
Project & Role Description Programmer Analyst
Design the powerful suite of novel algorithms using statistical and other techniques to monitor Fraud
Detection. The system will develop a very good graphical representation of customer activity, and can
watch for First Party Fraud, Third Party Fraud and Identity Fraud using the unique graph node entity
relations. The bank can use this system to detect a variety of illegal activity, including money
laundering, insider trading, front-running, intra-day manipulation, marking to close, and more. Fast
detection allows the bank to protect itself from considerable losses.
Responsibilities:
• Creating a Staging environment in Hive.
• Spark-neo4j connectivity and using spark to run neo4j.
• Creating nodes, relations, entities into Neo4j using Spark.
• Automating the whole process using scala language.
Environment: Hadoop, Spark, Scala, Neo4j, Hive
Project: Unified data BI Sept ‘15-
Feb’16
Client: CapitalOne
Project & Role Description Programmer Analyst
Unified data BI brings in a single window access to multiple data sources by building a federated query
engine and runtime environment. By defining a virtual database over various sources (currently
RDBMS, flat file & other non RDBMS sources in future) enables a business user or even a power user to
query over federated data from a standalone or web application.
- 1 -
2. Responsibilities:
• Designed modules for the Query Generation and translation using teiid designer framework
and integrating Spark to process the queries.
• Designed code in Scala integrated with teiid.
• Assisted fellow team mate for the design of different modules.
• Wrote ScalaTest for QueryTraslator, PlanGenerator and PlanProcessor modules.
Environment: Hadoop, Java, Scala, Spark
Project: Big Data COE May’15-Sept’16
Project & Role Description Programmer Analyst
Generic Framework puts the data from different platform (like servers, Databases, ftp) to hdfs.
This Application is used for the putting data from anywhere to hdfs. With the help flume, Sqoop, ftp etc.
we had created a generic application by this any format of data can be placed to hdfs and can be used
for the further process.
Responsibilities:
• Designed Sqoop import module to load the data to hdfs (DB level and Query level).
• Designed modules on validation and testing of Flume and Sqoop data import.
• Assisted fellow team mates for the design of the Logging framework using Log4j.
Environment: Hadoop, Flume, Sqoop, Shell Script, Java, Agile Methodology.
Project: nXg Device Analytics Aug’14- Apr’15
Client: Cognizant – COE (Centre of Excellence)
Project & Role Description Programmer Analyst
This Application is used for the analytics of the set top box logs data. The data was available in the AWS
S3 bucket. Data was ingested in the Hadoop Cluster using Flume. The master data which went through
daily update and new inserts were managed through HBase. The transactional data which come in point
in time data, daily new increments were managed through Hive.
After that R-Script runs on the data for the prediction and then data is loaded to qlickview for reporting
and analysis.
Responsibilities:
• Work on flume configuration to load data to Hdfs from Amazon s3bucket.
• W Integration from Hbase to hive as per the requirement.
• Populate the data in Data models.
• Work on Base tables, data mart tables and fact tables in HIVE to meet the business requirements
(for Star schema Data Mart).
Environnent : Hadoop, Hive, Java, Hbase, Flume, S3, Oozie.
Scholastics
• B. Tech (CSE), JECRC UDML– Rajasthan Technical University, Jaipur 2013
• 10+2, Gyan Vihar School, CBSE Board, 2009
• 10 , All Saints Church School, Rajasthan Board, 2007
- 2 -
3. Accolades & Achievements:
• Have been awarded as Young Innovator in Cts 2015.
• Consistently received appraisals by customer as well as senior team member.
• Consistently achieved highest performance rating among team member.
• Have been Part of Organizing Committee of Sarvatra-Tech Fest during 6th
semester.
• North India Trinity Guitar Grade 3 Topper 2013.
• Won many guitar competitions.
Personal Details:
Father’s Name : Kamlesh Rehi
Permanent Address : 340, Krishna Bhawna, Chandi ki Taksal, Jaipur.
Temporary Address : D-140, West Patel Nagar, New Delhi
Declaration:
I hereby declare that the above mentioned information is correct up to my knowledge and I bear the
responsibility for the correctness of the above mentioned particulars.
Date:
Place: Gurgaon (Deepesh Rehi)
- 3 -