Tara Prasad Panda is seeking a career opportunity with opportunities for growth. He has 2 years of experience in data warehousing and Informatica ETL development. Some of his responsibilities include requirement gathering, designing mappings, developing reusable transformation logic, testing, and migrating code to production environments. Currently he works as a Senior Software Engineer at Capgemini India Pvt. Ltd. on a data masking project for Barclays Capital involving masking sensitive personal information from databases, files and message queues.
Hi There , I am a passionate Data Engineer doesn't do just data cleansing and logics but also love building data pipelines from full product life cycle . Getting involved from initial phase to delivery phase (live) makes me feel valued .
I have approx 6 years of professional experience as a Data Engineering in the field of DW/DE
Acquired Competencies are ETL pipeline building, pyspark , Spark SQL, Python informatica,Talend and MySql . AWS Certified Developer.
having said that have also worked on Databricks delta lake , ADF , ADB Notebooks, Hive .
1. Tara Prasad Panda
Flat no. C-301
Icon Linera
Bhumkar Chowk, Pune
Mobile: +91 8796543968
E-Mail: panda.tara@gmail.com
CAREER OBJECTIVE
To build a career that offers career growth with opportunities to enrich my skills
while contributing my best to the organization I work with.
PORFESSIONAL SUMMARY
Have 2 years of experience in data warehousing and informatica, this includes mapping
creation and designing
Extensive hands on experience in informatica 9.6, informatica lifecycle management
Have good experience in Oracle 10g and MSSQL server
Worked with different data sources like database, flat files and MQ
Have knowledge of DEV, UAT and PROD environment
Experience with dimensional modeling using star schema and snowflake models
Created UNIX shell scripts to run the Informatica workflows and controlling the ETL flow
Performed performance tuning at session level to maximize data throughput
Developed a generic query for partitioning at session level
Built reusable transformation and mapping wherever redundancy is needed
Performed Unit testing and maintained test logs and test cases for all the mappings
Parameterized hard-coded values in session level and mapping level
Have good knowledge in HADOOP ecosystem especially in PIG and HIVE
Assisted the other ETL developers in solving complex scenarios and coordinated with
source systems owners with day-to-day ETL progress monitoring
EXPERIENCE SUMMARY
Working as Senior Software Engineer Capgemini India PVT. LTD. from NOV 2014 to till date.
ACADEMIC DETAILS
EXAMINATION SPECIALIZATION INSTITUTION
BOARD / YEAR OF PERCENTAGE /
UNIVERSITY PASSING CGPA DIVISION
Electronics and National Institute of
B.Tech Communication Science and BPUT 2014 8.7/10
Engineering Technology
12th
Science Kendriya Vidyalaya CBSE 2010 83 %
Bhubaneswar
10th
Kendriya Vidyalaya CBSE 2008 91.6 %
Bhubaneswar
2. TECHNICAL SKILLS
ETL Tool:
Informatica Power Center 9.6.
Informatica ILM 9.x
Database:-
Oracle 10g/11g.
SQL Server 2008.
Programming language:-
SQL, PIG, HIVE
o Operating System:-
UNIX (basic shell scripting).
MS DOS, Windows XP/7/8/10.
Tools
Autosys
Toad
SQL server management studio
PROJECT EXPERIENCE
Project -1:- Data masking
Organization: - Capgemini
Client: - Barclays Capital
Duration: - 1.9 years. (Feb-2014 to till date)
Summary:-
Data is vital and to protect personal data is a most challenging task for today's banking
organizations. Mostly data leakages occur from non-production environment. To avoid this we have
provided data masking solution, which will mask all the sensitive personal information before
migrating to non-production environments. We masked all type of sources like databases, flat files,
xmls, MQs, etc. We have designed standard mapplets to mask various information like full name,
user identification number, credit card number, account number, etc.
Challenges:-
Integrity is the most challenging task in masking. Which means particular account number
must be masked with some dummy valid account number throughout all the tables and applications.
To achieve this we have designed secure key and unique key mapplet. Secure key mapplet keeps
integrity while unique key mapplet will randomly mask data with unique value.
Another challenge is to developed complex process for running Informatica jobs. The process
must be complex for maintaining secrecy. No one can easily understand how data is getting masked.
We have used the autosys and unix scripts for running Informatica workflows.
3. Responsibilities:-
Mostly I engaged with development activities. I took the responsibility of whole application. So
far I have masked approximately 10 applications. Whole cycle involves different steps like
requirement gathering, documentation, plans set up, development, unit testing, sign off and
code migration to UAT then to PROD.
Understanding business requirement.
To import source from different data source like databases, flat files, MQ.
To generate code through Informatica ILM. ILM is a widely used tool for generating
Informatica code.
To standardize code with the help of UNIX script.
To prepare autosys jobs by taking care of dependencies. Autosys is a third party tool which is
widely used for running Informatica workflows by setting up dependencies.
To prepare mappings for complex code. Mapping logic involves calculations, implementation
of algorithms, handling integrity especially for many to many relationships.
Developed mappings that perform Extraction, Transformation and load of source data into
derived Masters schema using various power center transformations like Source Qualifier,
Aggregator, Filter, Router, Sequence Generator, look up, Rank, Joiner, Expression, XML
parser, XML generator, Transaction control transformation, Normalizer and update strategy to
meet business logic in the mappings
To develop mapplet. Mostly we create mapplet because the masking logic for particular
columns are same e.g. account number, credit card number has almost same masking logic.
So it is more feasible to use mapplet rather than developing code in each mapping.
To develop workflows and sessions. Workflows may be sequential, parallel, with
dependencies, etc.
Performed partitioning at session level to increase throughput.
To prepare deployment and run book documents. Deployment document contains entire
information how to migrate code in prod or go live. Run books provide guidance to end user
how to run jobs and prepare a mask copy of databases.
ACHIEVEMENTS/ EXTRA-CURRULAR ACTIVITIES
Awarded as the Rising Star for H1 2016.
Attended Cloudera HADOOP training.
Attended PIG and HIVE classroom trainings and did mini project in PIG.
Attended Manage by metrics training session.
Attended Software configuration management training.
PERSONAL PROFILE
Name : Tara Prasad Panda
Date of Birth : 10th
July 1992
Nationality : Indian
Gender : Male
Marital Status : Single
Hobbies : Playing Table Tennis and reading tech news.
Languages known : English, Hindi and Odiya
Declaration:-
I hereby declare that the information furnished above is true to the best of my knowledge and belief.
Date:
Place: Tara Prasad Panda