SlideShare a Scribd company logo
1 of 8
SOURAV BANERJEE
Souravbanerjee423@gmail.com
9482400751/9619860924
CONTACT DETAILS:
Mob:-
9482400751/9619
860924
PERSONAL DATA:
Date of Birth:
14/03/1991
Gender: Male
Nationality:
Indian
Languages:
English, Hindi,
Bengali.
A dynamic & result-oriented Software Developer (Senior Software Engineer) with strong
background in Banking/Credit Card Domain on Hadoop /SPARK Technology and
HP Tandem/Cobol environment.
Also having Proven track record of 4+ years in diverse facets of software development,
Design, Data Analytics and execution of business applications. Also having knowledge on
Data Mining, Machine Learning and Data Science.
PROFESSIONAL CAREER DETAILS:
Domain Knowledge Wholesale Banking / Credit Card Domain.
Functional Areas Effective communicator with ability to convey ideas in
speaking and writing, Excellent analytical and decision
making skills. Have a good knowledge on wholesale
banking/Card (Switch) Concept.
TECHNICAL SKILL SET:
Languages JAVA (J2EE),PYTHON(BASIC),R, SQL, COBOL,C.
Utilities SPOOLCOM,BATCHCOM, GBS(ATLAS),BATCHCOM, PERUSE,
FUP, ENFORM, ,IDCAMS,DFSORT, FTP,INSPECT,TMF,HUE,
Cloudera Manager
Tools
Hadoop Ecosystem
IDE’s
Relational
Databases
STORQM, FRACT, RDDM , MS Excel, ULTRA EDIT, EXPEDITOR,
Streaming Platform, VMWARE
Map Reduce, Sqoop, Hive, Pig, Hbase, HDFS, Zookeeper
ECLIPS, Microsoft SQL server 2010,ATOM
Enscribe File System(SQL/MP), DB2,HDFS
No SQL Database
Framework
Application Servers
HBase
Hadoop Framework
HP TANDEM, IBM MAINFRAME Z/OS
Operating Systems
Scripting Language
Project
Management Tools
IT Operation
Standard
WINDOWS XP/7,UNIX/LINUX, Guardian(Non Stop Kernel),
Z/OS
Shell Script , Python(Basic)
HPSM, JIRA, SNOW
SCRUM and Agile Methodology
EDUCATIONAL BACKGROUND :
Bachelor of Technology.
Year of Passout :June ,2012
CAREER SUMMARY:
 Proactive, flexible, customer focused, innovative with good analytical skills and
Can work under pressure and tight deadlines.
 3 years of extensive experience in COBOL, TACL, TAL, UNIX, and IBM Mainframe
Systems and Tandem Development.
 1+ years of experience as Hadoop developer having sound knowledge of Big Data
Technology Stack– Hadoop, Hive, HDFS, Mapreduce, Sqoop,PIG,Flume,Impala.
 Strong Knowledge of Hadoop and Hive and Hive's analytical functions
 Hands on Experience in working with ecosystems like Hive, Pig, Sqoop, Map Reduce,
Flume, OoZie.
 Implemented Proofs of Concept on Hadoop stack and different big data analytic tools,
migration from different databases (i.e Teradata, Oracle,MYSQL ) to Hadoop.
 Successfully loaded files to Hive and HDFS from HBase
 Efficient in building hive, pig and map Reduce scripts.
 Loaded the dataset into Hive for ETL Operation.
 Experience in using DBvisualizer, Zoo keeper and cloudera Manager.
 Experience in database design using Stored Procedure, Functions, Triggers and strong
experience in writing complex queries for DB2, SQL Server.
 Banking Knowledge particularly in the areas of payment, Interest Accruals, FX
And MM.
 Credit Card Domain Knowledge especially in Issuing and Acquiring.
 Has also experience in Testing (Unit testing, functional analysis testing and user
acceptance testing).
 Ability to pay attention to detail and develop a bigger picture of problem.
 Have a good experience on working in AGILE and Waterfall methodology.
 Experience on working in SCRUM methodology
CERTIFICATION:
 Certification on Basic of Financial Market – NCFM
 Big Data Analytics with HDInsight: Hadoop on Azure – Microsoft
 IBM Certified Big Data and Hadoop Developer - IBM
 Certified Big data and Hadoop Developer - Simplilearn
 Winner of TCS GEMS On-The-Spot Award
 Refer to below mentioned LinkedIn profile for more details.
 Certified Data Science – R Programming - Simplilearn
PROFESSIONAL EXPERIENCE:
Company Mashreq Global Service
(From Nov 2016 to Current)
Technical Role : Hadoop Developer
Working as development - production support executive in Card Domain Technology IN
issuing and acquiring module.
Install raw Hadoop and NoSQL applications and develop programs for analyzing data.
Responsibilities:
• Replaced default Derby metadata storage system for Hive with MySQL system.
• Executed queries using Hive and developed Map-Reduce jobs to analyze data.
• Developed Pig Latin scripts to extract the data from the web server output files
to load into HDFS.
• Developed the Pig UDF's to preprocess the data for analysis.
• Developed Hive queries for the analysts.
• Use Impala in place of Hive for some short application.
• Use Flume for Streaming Data.
• Involved in loading data from LINUX and UNIX file system to HDFS.
• Analyzed the web log data using the HiveQL.
Technology Used: Core Java, Apache Hadoop,HDFS, Pig, Hive, Cassandra, Shell Scripting,
My Sql, LINUX, UNIX
Tata Consultancy Service LTD
(From Oct 2012 to Aug 2016)
Worked as System Engineer for banking and financial service practice. In my tenure of
engagement with TCS, I was involved in performing activities like analyzing business
requirements, technical solution design; develop COBOL codes, quality assurance and
internal reviews.
Client:
ING Wholesale
Banking
Duration
3 year 4 months
(From May 13 –
Aug 2016)
Role:
System Engineer
Key Projects Undertaken:
ATLAS Data Analysis :
Technical Role : Hadoop Developer
Duration: 6 months.
Project Details: ING bank from Netherlands is one of the leading banks in the world.
ING migrated from Atlas to GBS in 2008; however regulatory requirements make it
mandatory to keep data for 10 years.
ING keeps the data for 20 years and with the Atlas servers decommissioned,
maintaining the servers to retain the data is very costly.
Thus all the data from Atlas server was decided to migrate to Hadoop , As the data is
highly structured the data was stored in Hive tables so that further analysis can be done
on the data and Regulatory/customer requirements can be met.
Responsibilities:
• Developed Hive scripts for end user / analyst requirements to perform
• ad hoc analysis
• Solved performance issues in Hive with understanding of Joins, Group and
aggregation and how does it translate to Map Reduce jobs.
• Created External/ Managed Hive tables with optimized partitioning and
bucketing.
• Data is partitioned on the basis of year, sub branch, currency and account type.
• Generate the statements for an account for a particular year.
•
Technology Used: COBOL, UNIX, JAVA, SQOOP, HIVE.
ING Security Monitoring:
Technical Role : Hadoop Developer
Duration: 2 months.
Project Details: Daily activities in the GBS environments are recorded in the log file.
These log files are stored as flat files. At the end of the day after the Pre-EOD run the log
files are FTPed to the local system and then stored in HDFS. The log files are analyzed
using MapReduce to check for any suspicious activity.
Responsibilities:
• Worked as lead developer payment module and Foreign Exchange module.
• Developed problem definition design for enabling project requirements.
• Developed MapReduce program to analyze the log files and check for any
suspicious activity.
• To perform team activities like internal code reviews and quality assurance
Technology Used : MQ, MapReduce, Hadoop Framework,Tandem,UNIX
Statement Generation:
Technical Role : Hadoop Developer
Duration: 2 months
Project Details: As an extension to the daily log analysis for security monitoring, hadoop
can be used to analyze the accounting entry records to generate the statements for
customers on daily basis.The generation of the statements makes up a large portion of
the daily EOD processing.
Responsibilities:
• Developing MapReduce program to read the accounting entries file and
generate the output such and send it PSE (Paper Statement Engine) for further
processing.
• Worked on whole design and development part of the project
• Develop new COBOL codes for new tags in SWIFT messaging.
• Designing application flow and user interaction
• Providing development solutions to meet end user requirements.
• Delivering code with defined standards within given timeframe
Technology Used: UNIX, Mapreduce, COBOL, Hive,SQOOP,HDFS.
ING Wholesale Bank: GBS LCM Release 2014
Technical Role : Tandem Developer
Duration: 8 months
Project Details: It was a combination of different small change requests (CR) which are
solely based on various customer requirement, such as Report Creation, Payment Flow
changes, Database Handling.
Technology used : IBM Mainframe, TANDEM, CICS,UNIX, JAVA
Responsibilities:
• Worked as lead developer
• Develop codes from the very scratch and perform small changes in existing
programs.
• Developing technical solution design for enabling customer requirements.
• Monitor day to day activities.
• To perform team activities like internal code reviews, time estimates and quality
assurance
Technology used : COBOL, MVS/JCL, TAL, UNIX,TANDEM.
ING Wholesale Bank: SWIFT 2013
Technical Role : Tandem Developer
Duration: 4 months
Project Details- It was based on updating of SWIFT messaging based on the new SWIFT
2013 manual. We had to add new tag to accommodate new facility to the customer.
Technology used : COBOL, MVS/JCL, TAL, UNIX,TANDEM.
Responsibilities:
• Worked on whole design and development part of the project
• Designing application flow and user interaction
• Providing development solutions to meet end user requirements.
• Delivering code with defined standards within given timeframe
ING Wholesale Bank: GMAINT 2013
Technical Role : Tandem Developer
Duration: 4 months
Project Details: In this project WSS (Wall Street System) deliver new components to
accommodate in the TANDEM server. We gather the requirement specifically for ING
and modify and update the WSS deliveries as per our own requirement.
Technology used : TANDEM, COBOL, JCL, and DB2
Responsibilities:
• Worked as developer
• Involved in Technical analysis part.
• Worked on section preparation and delivery part.
• Gathering requirements from onsite functional team
• Designing application flow and user interaction
• Worked on documentation part.
ING Wholesale Bank : GBS LCM Release 2013
Technical Role : Tandem Developer
Duration: 8 months
Project Details: It is a combination of different small change requests(CR) which are
Client:
TCS INTERNAL
Duration
4 months
(Jan 2013 – April
2013)
Role:
Software
developer
solely based on various customer requirement ,such as Report Creation, Payment Flow
changes, Database Handling.
Technology Used : IBM Mainframe, TANDEM, CICS,UNIX, JAVA
Responsibilities:
• Worked as developer.
• Develop codes from the very scratch and perform small changes in existing
programs.
• Developing technical solution design for enabling customer requirements.
• Monitor day to day activities.
Non Life Insurance First Quote Generation (Jan 2013 to Apr 2013)
Technical Role : Tandem Developer
Duration: 4 months
Project Details: The Non-Life insurance Quote generation system or Non-life Fast Quote
generation system need to estimate the premium amount for an applicant based on
the given factors including type of insurance, coverage amount, length of coverage, age,
gender, driving history, health and medical history, family history, vehicle history and
approximate rating class. The calculated premium should be displayed on the screen
and the quote has to be saved in the system database after the completion of the
transaction
Technology used : COBOL, JCL, and DB2,IBM Mainframe,UNIX
Responsibilities:
• Developing functional solution design as per business requirement
• To develop COBOL codes and JCL jobs from scratch
• Creating MAPS using CICS for UI purpose.
• Developing technical solution design for enabling customer requirements.
• To perform team activities like time estimates and quality assurance of
documents
REFERENCES
References Available upon Request
LinkedIn Profile : https://www.linkedin.com/in/sourav-banerjee-50b443106/
DECLARATION
I hereby declare that the above-mentioned information is correct up to my knowledge and I bear the
responsibility for the correctness of the above-mentioned particulars.
Place: Bangalore (Sourav Banerjee)

More Related Content

What's hot

Sunshine consulting mopuru babu cv_java_j2ee_spring_bigdata_scala
Sunshine consulting mopuru babu cv_java_j2ee_spring_bigdata_scalaSunshine consulting mopuru babu cv_java_j2ee_spring_bigdata_scala
Sunshine consulting mopuru babu cv_java_j2ee_spring_bigdata_scalaMopuru Babu
 
Vishwanath_M_CV_NL
Vishwanath_M_CV_NLVishwanath_M_CV_NL
Vishwanath_M_CV_NLVishwanath M
 
Has Traditional MDM Finally Met its Match?
Has Traditional MDM Finally Met its Match?Has Traditional MDM Finally Met its Match?
Has Traditional MDM Finally Met its Match?Inside Analysis
 
SACHIN_SINGH_SAP-BW
SACHIN_SINGH_SAP-BWSACHIN_SINGH_SAP-BW
SACHIN_SINGH_SAP-BWsachin singh
 
Ron Mathewson resume
Ron Mathewson resumeRon Mathewson resume
Ron Mathewson resumeRon mathewson
 
Rajendra kori it_project lead_9_cv
Rajendra kori it_project lead_9_cvRajendra kori it_project lead_9_cv
Rajendra kori it_project lead_9_cvRajendra Kori
 
Chandan's_Resume
Chandan's_ResumeChandan's_Resume
Chandan's_ResumeChandan Das
 
Resume_Md ZakirHussain
Resume_Md ZakirHussainResume_Md ZakirHussain
Resume_Md ZakirHussainzakir hussain
 
Jerun profile 2012_08_10
Jerun profile 2012_08_10Jerun profile 2012_08_10
Jerun profile 2012_08_10CapIpad
 
Resume april updated
Resume april updatedResume april updated
Resume april updatedSukanta Saha
 

What's hot (14)

Sunshine consulting mopuru babu cv_java_j2ee_spring_bigdata_scala
Sunshine consulting mopuru babu cv_java_j2ee_spring_bigdata_scalaSunshine consulting mopuru babu cv_java_j2ee_spring_bigdata_scala
Sunshine consulting mopuru babu cv_java_j2ee_spring_bigdata_scala
 
Vishwanath_M_CV_NL
Vishwanath_M_CV_NLVishwanath_M_CV_NL
Vishwanath_M_CV_NL
 
Has Traditional MDM Finally Met its Match?
Has Traditional MDM Finally Met its Match?Has Traditional MDM Finally Met its Match?
Has Traditional MDM Finally Met its Match?
 
Neeti resume 1
Neeti resume 1Neeti resume 1
Neeti resume 1
 
SACHIN_SINGH_SAP-BW
SACHIN_SINGH_SAP-BWSACHIN_SINGH_SAP-BW
SACHIN_SINGH_SAP-BW
 
Ron Mathewson resume
Ron Mathewson resumeRon Mathewson resume
Ron Mathewson resume
 
Rajendra kori it_project lead_9_cv
Rajendra kori it_project lead_9_cvRajendra kori it_project lead_9_cv
Rajendra kori it_project lead_9_cv
 
Chandan's_Resume
Chandan's_ResumeChandan's_Resume
Chandan's_Resume
 
kk_resume
kk_resumekk_resume
kk_resume
 
ZakirHussain
ZakirHussainZakirHussain
ZakirHussain
 
Resume_Md ZakirHussain
Resume_Md ZakirHussainResume_Md ZakirHussain
Resume_Md ZakirHussain
 
NaliniProfile
NaliniProfileNaliniProfile
NaliniProfile
 
Jerun profile 2012_08_10
Jerun profile 2012_08_10Jerun profile 2012_08_10
Jerun profile 2012_08_10
 
Resume april updated
Resume april updatedResume april updated
Resume april updated
 

Similar to Sourav banerjee resume

Srikanth hadoop hyderabad_3.4yeras - copy
Srikanth hadoop hyderabad_3.4yeras - copySrikanth hadoop hyderabad_3.4yeras - copy
Srikanth hadoop hyderabad_3.4yeras - copysrikanth K
 
Nagarjuna_Damarla_Resume
Nagarjuna_Damarla_ResumeNagarjuna_Damarla_Resume
Nagarjuna_Damarla_ResumeNag Arjun
 
Pankaj Resume for Hadoop,Java,J2EE - Outside World
Pankaj Resume for Hadoop,Java,J2EE -  Outside WorldPankaj Resume for Hadoop,Java,J2EE -  Outside World
Pankaj Resume for Hadoop,Java,J2EE - Outside WorldPankaj Kumar
 
Sunshine consulting mopuru babu cv_java_j2_ee_spring_bigdata_scala_Spark
Sunshine consulting mopuru babu cv_java_j2_ee_spring_bigdata_scala_SparkSunshine consulting mopuru babu cv_java_j2_ee_spring_bigdata_scala_Spark
Sunshine consulting mopuru babu cv_java_j2_ee_spring_bigdata_scala_SparkMopuru Babu
 
Sunshine consulting Mopuru Babu CV_Java_J2ee_Spring_Bigdata_Scala_Spark
Sunshine consulting Mopuru Babu CV_Java_J2ee_Spring_Bigdata_Scala_SparkSunshine consulting Mopuru Babu CV_Java_J2ee_Spring_Bigdata_Scala_Spark
Sunshine consulting Mopuru Babu CV_Java_J2ee_Spring_Bigdata_Scala_SparkMopuru Babu
 
Nagarjuna_Damarla
Nagarjuna_DamarlaNagarjuna_Damarla
Nagarjuna_DamarlaNag Arjun
 
Resume quaish abuzer
Resume quaish abuzerResume quaish abuzer
Resume quaish abuzerquaish abuzer
 

Similar to Sourav banerjee resume (20)

Srikanth hadoop hyderabad_3.4yeras - copy
Srikanth hadoop hyderabad_3.4yeras - copySrikanth hadoop hyderabad_3.4yeras - copy
Srikanth hadoop hyderabad_3.4yeras - copy
 
Resume_VipinKP
Resume_VipinKPResume_VipinKP
Resume_VipinKP
 
Resume_2706
Resume_2706Resume_2706
Resume_2706
 
Resume
ResumeResume
Resume
 
hadoop exp
hadoop exphadoop exp
hadoop exp
 
HimaBindu
HimaBinduHimaBindu
HimaBindu
 
Yasar resume 2
Yasar resume 2Yasar resume 2
Yasar resume 2
 
Pushpendra
PushpendraPushpendra
Pushpendra
 
Ankit_Yadav
Ankit_YadavAnkit_Yadav
Ankit_Yadav
 
Nagarjuna_Damarla_Resume
Nagarjuna_Damarla_ResumeNagarjuna_Damarla_Resume
Nagarjuna_Damarla_Resume
 
Hadoop Developer
Hadoop DeveloperHadoop Developer
Hadoop Developer
 
Pankaj Resume for Hadoop,Java,J2EE - Outside World
Pankaj Resume for Hadoop,Java,J2EE -  Outside WorldPankaj Resume for Hadoop,Java,J2EE -  Outside World
Pankaj Resume for Hadoop,Java,J2EE - Outside World
 
Sunshine consulting mopuru babu cv_java_j2_ee_spring_bigdata_scala_Spark
Sunshine consulting mopuru babu cv_java_j2_ee_spring_bigdata_scala_SparkSunshine consulting mopuru babu cv_java_j2_ee_spring_bigdata_scala_Spark
Sunshine consulting mopuru babu cv_java_j2_ee_spring_bigdata_scala_Spark
 
Sunshine consulting Mopuru Babu CV_Java_J2ee_Spring_Bigdata_Scala_Spark
Sunshine consulting Mopuru Babu CV_Java_J2ee_Spring_Bigdata_Scala_SparkSunshine consulting Mopuru Babu CV_Java_J2ee_Spring_Bigdata_Scala_Spark
Sunshine consulting Mopuru Babu CV_Java_J2ee_Spring_Bigdata_Scala_Spark
 
DeepeshRehi
DeepeshRehiDeepeshRehi
DeepeshRehi
 
Nagarjuna_Damarla
Nagarjuna_DamarlaNagarjuna_Damarla
Nagarjuna_Damarla
 
hadoop_bigdata
hadoop_bigdatahadoop_bigdata
hadoop_bigdata
 
Robin_Hadoop
Robin_HadoopRobin_Hadoop
Robin_Hadoop
 
Divya
DivyaDivya
Divya
 
Resume quaish abuzer
Resume quaish abuzerResume quaish abuzer
Resume quaish abuzer
 

More from Sourav Banerjee

Big Data Engineer with Spark
Big Data Engineer with SparkBig Data Engineer with Spark
Big Data Engineer with SparkSourav Banerjee
 
Weka data mining with open source machine learning tool
Weka  data mining with open source machine learning toolWeka  data mining with open source machine learning tool
Weka data mining with open source machine learning toolSourav Banerjee
 
Python a z python for data science with real exercises!
Python a z  python for data science with real exercises!Python a z  python for data science with real exercises!
Python a z python for data science with real exercises!Sourav Banerjee
 
Matplotlib complete tutorial machine learning pre requisite
Matplotlib complete tutorial machine learning pre requisiteMatplotlib complete tutorial machine learning pre requisite
Matplotlib complete tutorial machine learning pre requisiteSourav Banerjee
 
Machine Learning Specialization
Machine Learning SpecializationMachine Learning Specialization
Machine Learning SpecializationSourav Banerjee
 
Microsoft certified professional_certificate
Microsoft certified professional_certificateMicrosoft certified professional_certificate
Microsoft certified professional_certificateSourav Banerjee
 
Mlflow managing the machine learning lifecycle
Mlflow managing the machine learning lifecycleMlflow managing the machine learning lifecycle
Mlflow managing the machine learning lifecycleSourav Banerjee
 
ETL part 1 data extraction
ETL part 1 data extractionETL part 1 data extraction
ETL part 1 data extractionSourav Banerjee
 
ETL Part 2 Transformations and Loads
ETL Part 2 Transformations and LoadsETL Part 2 Transformations and Loads
ETL Part 2 Transformations and LoadsSourav Banerjee
 
Getting started with apache spark sql
Getting started with apache spark sqlGetting started with apache spark sql
Getting started with apache spark sqlSourav Banerjee
 
Jigsaw Academy Certificates-Random Forest
Jigsaw Academy Certificates-Random ForestJigsaw Academy Certificates-Random Forest
Jigsaw Academy Certificates-Random ForestSourav Banerjee
 
Certified Data Scientist with SAS and Excel
Certified Data Scientist with SAS and ExcelCertified Data Scientist with SAS and Excel
Certified Data Scientist with SAS and ExcelSourav Banerjee
 

More from Sourav Banerjee (19)

Big Data Engineer with Spark
Big Data Engineer with SparkBig Data Engineer with Spark
Big Data Engineer with Spark
 
Weka data mining with open source machine learning tool
Weka  data mining with open source machine learning toolWeka  data mining with open source machine learning tool
Weka data mining with open source machine learning tool
 
Python a z python for data science with real exercises!
Python a z  python for data science with real exercises!Python a z  python for data science with real exercises!
Python a z python for data science with real exercises!
 
Matplotlib complete tutorial machine learning pre requisite
Matplotlib complete tutorial machine learning pre requisiteMatplotlib complete tutorial machine learning pre requisite
Matplotlib complete tutorial machine learning pre requisite
 
Machine Learning Specialization
Machine Learning SpecializationMachine Learning Specialization
Machine Learning Specialization
 
Microsoft certified professional_certificate
Microsoft certified professional_certificateMicrosoft certified professional_certificate
Microsoft certified professional_certificate
 
MongoDB Basic - M001
MongoDB Basic - M001MongoDB Basic - M001
MongoDB Basic - M001
 
Structured streaming
Structured streamingStructured streaming
Structured streaming
 
Managed delta lake
Managed delta lakeManaged delta lake
Managed delta lake
 
Mlflow managing the machine learning lifecycle
Mlflow managing the machine learning lifecycleMlflow managing the machine learning lifecycle
Mlflow managing the machine learning lifecycle
 
ETL part 1 data extraction
ETL part 1 data extractionETL part 1 data extraction
ETL part 1 data extraction
 
ETL Part 2 Transformations and Loads
ETL Part 2 Transformations and LoadsETL Part 2 Transformations and Loads
ETL Part 2 Transformations and Loads
 
Getting started with apache spark sql
Getting started with apache spark sqlGetting started with apache spark sql
Getting started with apache spark sql
 
Etl part 3 production
Etl part 3 productionEtl part 3 production
Etl part 3 production
 
Jigsaw Academy Certificates-Random Forest
Jigsaw Academy Certificates-Random ForestJigsaw Academy Certificates-Random Forest
Jigsaw Academy Certificates-Random Forest
 
Certificates
CertificatesCertificates
Certificates
 
Certified Data Scientist with SAS and Excel
Certified Data Scientist with SAS and ExcelCertified Data Scientist with SAS and Excel
Certified Data Scientist with SAS and Excel
 
Final certification
Final certificationFinal certification
Final certification
 
Sourav banerjee
Sourav banerjeeSourav banerjee
Sourav banerjee
 

Recently uploaded

Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Mattias Andersson
 
Bluetooth Controlled Car with Arduino.pdf
Bluetooth Controlled Car with Arduino.pdfBluetooth Controlled Car with Arduino.pdf
Bluetooth Controlled Car with Arduino.pdfngoud9212
 
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationBeyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationSafe Software
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationSlibray Presentation
 
Pigging Solutions Piggable Sweeping Elbows
Pigging Solutions Piggable Sweeping ElbowsPigging Solutions Piggable Sweeping Elbows
Pigging Solutions Piggable Sweeping ElbowsPigging Solutions
 
My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024The Digital Insurer
 
APIForce Zurich 5 April Automation LPDG
APIForce Zurich 5 April  Automation LPDGAPIForce Zurich 5 April  Automation LPDG
APIForce Zurich 5 April Automation LPDGMarianaLemus7
 
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Patryk Bandurski
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsMark Billinghurst
 
Unblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesUnblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesSinan KOZAK
 
Understanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitectureUnderstanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitecturePixlogix Infotech
 
Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Scott Keck-Warren
 
AI as an Interface for Commercial Buildings
AI as an Interface for Commercial BuildingsAI as an Interface for Commercial Buildings
AI as an Interface for Commercial BuildingsMemoori
 
New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024BookNet Canada
 
SIEMENS: RAPUNZEL – A Tale About Knowledge Graph
SIEMENS: RAPUNZEL – A Tale About Knowledge GraphSIEMENS: RAPUNZEL – A Tale About Knowledge Graph
SIEMENS: RAPUNZEL – A Tale About Knowledge GraphNeo4j
 
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)Wonjun Hwang
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024Scott Keck-Warren
 
Making_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptx
Making_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptxMaking_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptx
Making_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptxnull - The Open Security Community
 

Recently uploaded (20)

Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?
 
Bluetooth Controlled Car with Arduino.pdf
Bluetooth Controlled Car with Arduino.pdfBluetooth Controlled Car with Arduino.pdf
Bluetooth Controlled Car with Arduino.pdf
 
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationBeyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck Presentation
 
Pigging Solutions Piggable Sweeping Elbows
Pigging Solutions Piggable Sweeping ElbowsPigging Solutions Piggable Sweeping Elbows
Pigging Solutions Piggable Sweeping Elbows
 
My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024
 
APIForce Zurich 5 April Automation LPDG
APIForce Zurich 5 April  Automation LPDGAPIForce Zurich 5 April  Automation LPDG
APIForce Zurich 5 April Automation LPDG
 
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
 
DMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special EditionDMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special Edition
 
Vulnerability_Management_GRC_by Sohang Sengupta.pptx
Vulnerability_Management_GRC_by Sohang Sengupta.pptxVulnerability_Management_GRC_by Sohang Sengupta.pptx
Vulnerability_Management_GRC_by Sohang Sengupta.pptx
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR Systems
 
Unblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen FramesUnblocking The Main Thread Solving ANRs and Frozen Frames
Unblocking The Main Thread Solving ANRs and Frozen Frames
 
Understanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitectureUnderstanding the Laravel MVC Architecture
Understanding the Laravel MVC Architecture
 
Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024
 
AI as an Interface for Commercial Buildings
AI as an Interface for Commercial BuildingsAI as an Interface for Commercial Buildings
AI as an Interface for Commercial Buildings
 
New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
New from BookNet Canada for 2024: BNC BiblioShare - Tech Forum 2024
 
SIEMENS: RAPUNZEL – A Tale About Knowledge Graph
SIEMENS: RAPUNZEL – A Tale About Knowledge GraphSIEMENS: RAPUNZEL – A Tale About Knowledge Graph
SIEMENS: RAPUNZEL – A Tale About Knowledge Graph
 
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
Bun (KitWorks Team Study 노별마루 발표 2024.4.22)
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024
 
Making_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptx
Making_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptxMaking_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptx
Making_way_through_DLL_hollowing_inspite_of_CFG_by_Debjeet Banerjee.pptx
 

Sourav banerjee resume

  • 1. SOURAV BANERJEE Souravbanerjee423@gmail.com 9482400751/9619860924 CONTACT DETAILS: Mob:- 9482400751/9619 860924 PERSONAL DATA: Date of Birth: 14/03/1991 Gender: Male Nationality: Indian Languages: English, Hindi, Bengali. A dynamic & result-oriented Software Developer (Senior Software Engineer) with strong background in Banking/Credit Card Domain on Hadoop /SPARK Technology and HP Tandem/Cobol environment. Also having Proven track record of 4+ years in diverse facets of software development, Design, Data Analytics and execution of business applications. Also having knowledge on Data Mining, Machine Learning and Data Science. PROFESSIONAL CAREER DETAILS: Domain Knowledge Wholesale Banking / Credit Card Domain. Functional Areas Effective communicator with ability to convey ideas in speaking and writing, Excellent analytical and decision making skills. Have a good knowledge on wholesale banking/Card (Switch) Concept. TECHNICAL SKILL SET: Languages JAVA (J2EE),PYTHON(BASIC),R, SQL, COBOL,C. Utilities SPOOLCOM,BATCHCOM, GBS(ATLAS),BATCHCOM, PERUSE, FUP, ENFORM, ,IDCAMS,DFSORT, FTP,INSPECT,TMF,HUE, Cloudera Manager Tools Hadoop Ecosystem IDE’s Relational Databases STORQM, FRACT, RDDM , MS Excel, ULTRA EDIT, EXPEDITOR, Streaming Platform, VMWARE Map Reduce, Sqoop, Hive, Pig, Hbase, HDFS, Zookeeper ECLIPS, Microsoft SQL server 2010,ATOM Enscribe File System(SQL/MP), DB2,HDFS No SQL Database Framework Application Servers HBase Hadoop Framework HP TANDEM, IBM MAINFRAME Z/OS
  • 2. Operating Systems Scripting Language Project Management Tools IT Operation Standard WINDOWS XP/7,UNIX/LINUX, Guardian(Non Stop Kernel), Z/OS Shell Script , Python(Basic) HPSM, JIRA, SNOW SCRUM and Agile Methodology EDUCATIONAL BACKGROUND : Bachelor of Technology. Year of Passout :June ,2012 CAREER SUMMARY:  Proactive, flexible, customer focused, innovative with good analytical skills and Can work under pressure and tight deadlines.  3 years of extensive experience in COBOL, TACL, TAL, UNIX, and IBM Mainframe Systems and Tandem Development.  1+ years of experience as Hadoop developer having sound knowledge of Big Data Technology Stack– Hadoop, Hive, HDFS, Mapreduce, Sqoop,PIG,Flume,Impala.  Strong Knowledge of Hadoop and Hive and Hive's analytical functions  Hands on Experience in working with ecosystems like Hive, Pig, Sqoop, Map Reduce, Flume, OoZie.  Implemented Proofs of Concept on Hadoop stack and different big data analytic tools, migration from different databases (i.e Teradata, Oracle,MYSQL ) to Hadoop.  Successfully loaded files to Hive and HDFS from HBase  Efficient in building hive, pig and map Reduce scripts.  Loaded the dataset into Hive for ETL Operation.  Experience in using DBvisualizer, Zoo keeper and cloudera Manager.  Experience in database design using Stored Procedure, Functions, Triggers and strong experience in writing complex queries for DB2, SQL Server.  Banking Knowledge particularly in the areas of payment, Interest Accruals, FX And MM.  Credit Card Domain Knowledge especially in Issuing and Acquiring.  Has also experience in Testing (Unit testing, functional analysis testing and user acceptance testing).  Ability to pay attention to detail and develop a bigger picture of problem.  Have a good experience on working in AGILE and Waterfall methodology.  Experience on working in SCRUM methodology
  • 3. CERTIFICATION:  Certification on Basic of Financial Market – NCFM  Big Data Analytics with HDInsight: Hadoop on Azure – Microsoft  IBM Certified Big Data and Hadoop Developer - IBM  Certified Big data and Hadoop Developer - Simplilearn  Winner of TCS GEMS On-The-Spot Award  Refer to below mentioned LinkedIn profile for more details.  Certified Data Science – R Programming - Simplilearn PROFESSIONAL EXPERIENCE: Company Mashreq Global Service (From Nov 2016 to Current) Technical Role : Hadoop Developer Working as development - production support executive in Card Domain Technology IN issuing and acquiring module. Install raw Hadoop and NoSQL applications and develop programs for analyzing data. Responsibilities: • Replaced default Derby metadata storage system for Hive with MySQL system. • Executed queries using Hive and developed Map-Reduce jobs to analyze data. • Developed Pig Latin scripts to extract the data from the web server output files to load into HDFS. • Developed the Pig UDF's to preprocess the data for analysis. • Developed Hive queries for the analysts. • Use Impala in place of Hive for some short application. • Use Flume for Streaming Data. • Involved in loading data from LINUX and UNIX file system to HDFS. • Analyzed the web log data using the HiveQL. Technology Used: Core Java, Apache Hadoop,HDFS, Pig, Hive, Cassandra, Shell Scripting, My Sql, LINUX, UNIX Tata Consultancy Service LTD (From Oct 2012 to Aug 2016) Worked as System Engineer for banking and financial service practice. In my tenure of engagement with TCS, I was involved in performing activities like analyzing business requirements, technical solution design; develop COBOL codes, quality assurance and internal reviews.
  • 4. Client: ING Wholesale Banking Duration 3 year 4 months (From May 13 – Aug 2016) Role: System Engineer Key Projects Undertaken: ATLAS Data Analysis : Technical Role : Hadoop Developer Duration: 6 months. Project Details: ING bank from Netherlands is one of the leading banks in the world. ING migrated from Atlas to GBS in 2008; however regulatory requirements make it mandatory to keep data for 10 years. ING keeps the data for 20 years and with the Atlas servers decommissioned, maintaining the servers to retain the data is very costly. Thus all the data from Atlas server was decided to migrate to Hadoop , As the data is highly structured the data was stored in Hive tables so that further analysis can be done on the data and Regulatory/customer requirements can be met. Responsibilities: • Developed Hive scripts for end user / analyst requirements to perform • ad hoc analysis • Solved performance issues in Hive with understanding of Joins, Group and aggregation and how does it translate to Map Reduce jobs. • Created External/ Managed Hive tables with optimized partitioning and bucketing. • Data is partitioned on the basis of year, sub branch, currency and account type. • Generate the statements for an account for a particular year. • Technology Used: COBOL, UNIX, JAVA, SQOOP, HIVE. ING Security Monitoring: Technical Role : Hadoop Developer Duration: 2 months. Project Details: Daily activities in the GBS environments are recorded in the log file. These log files are stored as flat files. At the end of the day after the Pre-EOD run the log files are FTPed to the local system and then stored in HDFS. The log files are analyzed using MapReduce to check for any suspicious activity. Responsibilities: • Worked as lead developer payment module and Foreign Exchange module. • Developed problem definition design for enabling project requirements. • Developed MapReduce program to analyze the log files and check for any suspicious activity. • To perform team activities like internal code reviews and quality assurance
  • 5. Technology Used : MQ, MapReduce, Hadoop Framework,Tandem,UNIX Statement Generation: Technical Role : Hadoop Developer Duration: 2 months Project Details: As an extension to the daily log analysis for security monitoring, hadoop can be used to analyze the accounting entry records to generate the statements for customers on daily basis.The generation of the statements makes up a large portion of the daily EOD processing. Responsibilities: • Developing MapReduce program to read the accounting entries file and generate the output such and send it PSE (Paper Statement Engine) for further processing. • Worked on whole design and development part of the project • Develop new COBOL codes for new tags in SWIFT messaging. • Designing application flow and user interaction • Providing development solutions to meet end user requirements. • Delivering code with defined standards within given timeframe Technology Used: UNIX, Mapreduce, COBOL, Hive,SQOOP,HDFS. ING Wholesale Bank: GBS LCM Release 2014 Technical Role : Tandem Developer Duration: 8 months Project Details: It was a combination of different small change requests (CR) which are solely based on various customer requirement, such as Report Creation, Payment Flow changes, Database Handling. Technology used : IBM Mainframe, TANDEM, CICS,UNIX, JAVA Responsibilities: • Worked as lead developer • Develop codes from the very scratch and perform small changes in existing programs. • Developing technical solution design for enabling customer requirements. • Monitor day to day activities. • To perform team activities like internal code reviews, time estimates and quality assurance
  • 6. Technology used : COBOL, MVS/JCL, TAL, UNIX,TANDEM. ING Wholesale Bank: SWIFT 2013 Technical Role : Tandem Developer Duration: 4 months Project Details- It was based on updating of SWIFT messaging based on the new SWIFT 2013 manual. We had to add new tag to accommodate new facility to the customer. Technology used : COBOL, MVS/JCL, TAL, UNIX,TANDEM. Responsibilities: • Worked on whole design and development part of the project • Designing application flow and user interaction • Providing development solutions to meet end user requirements. • Delivering code with defined standards within given timeframe ING Wholesale Bank: GMAINT 2013 Technical Role : Tandem Developer Duration: 4 months Project Details: In this project WSS (Wall Street System) deliver new components to accommodate in the TANDEM server. We gather the requirement specifically for ING and modify and update the WSS deliveries as per our own requirement. Technology used : TANDEM, COBOL, JCL, and DB2 Responsibilities: • Worked as developer • Involved in Technical analysis part. • Worked on section preparation and delivery part. • Gathering requirements from onsite functional team • Designing application flow and user interaction • Worked on documentation part. ING Wholesale Bank : GBS LCM Release 2013 Technical Role : Tandem Developer Duration: 8 months Project Details: It is a combination of different small change requests(CR) which are
  • 7. Client: TCS INTERNAL Duration 4 months (Jan 2013 – April 2013) Role: Software developer solely based on various customer requirement ,such as Report Creation, Payment Flow changes, Database Handling. Technology Used : IBM Mainframe, TANDEM, CICS,UNIX, JAVA Responsibilities: • Worked as developer. • Develop codes from the very scratch and perform small changes in existing programs. • Developing technical solution design for enabling customer requirements. • Monitor day to day activities. Non Life Insurance First Quote Generation (Jan 2013 to Apr 2013) Technical Role : Tandem Developer Duration: 4 months Project Details: The Non-Life insurance Quote generation system or Non-life Fast Quote generation system need to estimate the premium amount for an applicant based on the given factors including type of insurance, coverage amount, length of coverage, age, gender, driving history, health and medical history, family history, vehicle history and approximate rating class. The calculated premium should be displayed on the screen and the quote has to be saved in the system database after the completion of the transaction Technology used : COBOL, JCL, and DB2,IBM Mainframe,UNIX Responsibilities: • Developing functional solution design as per business requirement • To develop COBOL codes and JCL jobs from scratch • Creating MAPS using CICS for UI purpose. • Developing technical solution design for enabling customer requirements. • To perform team activities like time estimates and quality assurance of documents REFERENCES References Available upon Request LinkedIn Profile : https://www.linkedin.com/in/sourav-banerjee-50b443106/ DECLARATION I hereby declare that the above-mentioned information is correct up to my knowledge and I bear the responsibility for the correctness of the above-mentioned particulars.