SlideShare a Scribd company logo
Email: aditiwadekar10@gmail.com| Mobile: +91-87933-49212
BIG DATA –PROFESSIONAL
Working in Big Data with a two plus years of IT experience. I am part of Opus Consulting
Solutions Analytics Team. My responsibilities are to setup, configure & manage Hadoop and
MongoDB cluster. Perform Data Migration Activities. I am looking for an organization that
offers professional growth and provides an environment conducive to excellence and enhance
my ability to be an achiever.
Big Data Skills:
MongoDB
 Created Sharded cluster with replica Set on AWS production cluster
 Worked with Authentication and Authorization on cluster
 Installed and configured Nagios for monitoring shard cluster
 Tested Node migration, Cluster upgradation and Backup-recovery on MongoDB cluster
 Worked with Aggregate queries to analyze data
Hadoop
 Hands on experience with Apache Hadoop
 Experimented tools like Hbase, Pig, Hive by inserting, updating and performing query
operations
 Implemented Spark on Real time data
 Installed and configured cloudera-manager for testing purpose
BI-Tool
 Data Migration using Pentaho data Integration tools.
 Used Hunk/Splunk to generate and analyze reports
Hands on Experience with AWS Environment
 EC2, volumes, snapshot, S3, IAM
Other Technical Skills:
 Operation System: CentOS, RHEL,Ubuntu, Windows.
 Databases: MongoDB, Mysql, SQL Server
 Tools: Hunk/Splunk, Pentaho data Integration tools.
 Languages: Python
 Virtualization: VMware, Virtual Box
 Scripting/Automation: Bash, Shell Scripting
 Big Data: Ambari, HDFS, Hive HA, Spark
o Hue ,Oozie, YARN/Map reduce, NoSQL, Sqoop
o MongoDB, AWS
Process Oriented skills:
Worked for the Complete Lifecycle Requirement gathering, Analysis, Build, Unit testing,
Migration, Production deployment & Support activities.
CERTIFICATIONS
1. Certified on MongoDB Advanced Deployment and Operations August 2016
2. Initiation on Real time data analytics over Hadoop and MongoDB April 2015
3. Outstanding Team – “Cardtronics Big Data Team” April 2014-Sept 2014
4. Certified Courses
○ Hadoop 101 Sept 1, 2016
○ Introduction to Pig May 8, 2015
○ Using HBase for Real-time Access to your Big Data March 8, 2015
○ Spark Fundamentals I Jan 20, 2015
○ M102: MongoDB for DBAs Dec 1, 2014
○ Certified in Hadoop Fundamentals -1 Oct 27, 2014
○ M101J MongoDB for Java Developers Oct 7, 2014
AWA RDS/RECOGNITIONS
 Reward and Recognition obtained for outstanding Teamwork of Big Data Team
 Reward obtained for taking initiation on Real time data analytics over Hadoop and
MongoDB.
 Active participation in Opus Women Football Team and Women Cricket Team.
EXPERIENCE
Opus Consulting Solution 2014 – Present
1. Implemented a project required by the business users to analyze the number of
transactions taking place location and month wise across US, Jan 2016
○ Data migration and validation from MongoDB to Hadoop.
○ Installing and Configuring Splunk to connect Hadoop and MongoDB.
○ Created Providers and indexes in Splunk to connect hdfs.
○ Developed Dashboards in Splunk based on input parameters
○ Scripts written in Python to load data to Hadoop.
○ Developed, Configured and scheduled reports using Splunk
○ Developed SPL out of Stored Procedures and developed Reports.
○ Created users and roles in Splunk.
○ Query on Splunk to generate 20+ reports out of Hadoop.
Roles and Responsibility
○ Perform Requirement gathering to obtain business logic of the report
○ Provide estimates for development and testing of reports
○ Understand the business logic of the report.
○ Designed and Developed Reports in Hunk
○ Enhanced skills in SQL and SPL (Search Processing Language) while
developing reports
2. Worked on Implementing Spark on Real-Time transactional Data, Dec 2014
○ Installed and configured spark with MongoDB and Hadoop on CentOS 6.5.
MongoDB contained real-time data. Spark streaming was used to deal with live
streaming data. Spark was connected to MongoDB and various queries were
performed on the transactional data. The implementation helped in overcoming
the drawbacks of MongoDB that is joins. Queries were performed in faster and
efficient way on huge volume of data.
○ Apache Zeppelin was installed and configured with spark. The reports were
being developed in zeppelin and the transactional data was being analyzed in an
efficient way
Roles and Responsibility
○ Requirement gathering and provide design of how spark can be implemented on
Real time data.
○ Installation and configuration of MongoDB and Hadoop on Servers.
○ Developed queries in Spark and Zeppelin to connect and query over MongoDB
and Hadoop data.
○ The query was improvised to check the efficiency of the tool.
○ Reports generated out of Zeppelin
3. Created Hadoop and MongoDB as Data Warehouse, April 2015
○ Installed and configured MongoDB sharded cluster on AWS production. The
aim was to create Hadoop as a data warehouse.
○ Setup Hadoop cluster along with Ambari, Hue on Environment.
○ Performed data migration Activity using sqoop from SQL to Hadoop.
○ Familiar with Hbase, Pig, Hive by inserting, updating and performing query
operations
○ Developed Backup-scripts for taking MongoDB backup to prevent loss of data.
○ Used Pentaho-data integration tool to migrate data from SQL to MongoDB
○ Developed scripts in python to migrate data from MongoDB to Hadoop cluster.
○ Validate the data in MongoDB and Hadoop.
○ Hadoop consisted of 3 years of data.
Roles and Responsibility
○ Setting up Hadoop cluster and monitoring the cluster.
○ Perform Data Migration Activity from SQL to Hadoop or from MongoDB to
Hadoop
○ Create roles and Users in hive and grant permissions.
○ Setting up MongoDB sharded cluster with SSL authentication and
authorization.
○ Developed MongoDB Backup Scripts.
○ Installed and Configured Nagios
○ Monitored MongoDB cluster using Nagios.
○ Developed Aggregate queries in MongoDB to test the data in collections.
Candidate for Bachelors in Applied Science, Information Technology from College of
Engineering, Pune - 7.68 (Sept 2010-Aug2014)

More Related Content

What's hot

Data Analytics and Processing at Snap - Druid Meetup LA - September 2018
Data Analytics and Processing at Snap - Druid Meetup LA - September 2018Data Analytics and Processing at Snap - Druid Meetup LA - September 2018
Data Analytics and Processing at Snap - Druid Meetup LA - September 2018
Charles Allen
 
Talend spark meetup 03042017 - Paris Spark Meetup
Talend spark meetup 03042017 - Paris Spark MeetupTalend spark meetup 03042017 - Paris Spark Meetup
Talend spark meetup 03042017 - Paris Spark Meetup
Modern Data Stack France
 
963
963963
Druid: Under the Covers (Virtual Meetup)
Druid: Under the Covers (Virtual Meetup)Druid: Under the Covers (Virtual Meetup)
Druid: Under the Covers (Virtual Meetup)
Imply
 
Secured (Kerberos-based) Spark Notebook for Data Science: Spark Summit East t...
Secured (Kerberos-based) Spark Notebook for Data Science: Spark Summit East t...Secured (Kerberos-based) Spark Notebook for Data Science: Spark Summit East t...
Secured (Kerberos-based) Spark Notebook for Data Science: Spark Summit East t...
Spark Summit
 
How Netflix Uses Druid in Real-time to Ensure a High Quality Streaming Experi...
How Netflix Uses Druid in Real-time to Ensure a High Quality Streaming Experi...How Netflix Uses Druid in Real-time to Ensure a High Quality Streaming Experi...
How Netflix Uses Druid in Real-time to Ensure a High Quality Streaming Experi...
Imply
 
Big Data and Hadoop - key drivers, ecosystem and use cases
Big Data and Hadoop - key drivers, ecosystem and use casesBig Data and Hadoop - key drivers, ecosystem and use cases
Big Data and Hadoop - key drivers, ecosystem and use cases
Jeff Kelly
 
Apache Druid®: A Dance of Distributed Processes
 Apache Druid®: A Dance of Distributed Processes Apache Druid®: A Dance of Distributed Processes
Apache Druid®: A Dance of Distributed Processes
Imply
 
Splunk: Druid on Kubernetes with Druid-operator
Splunk: Druid on Kubernetes with Druid-operatorSplunk: Druid on Kubernetes with Druid-operator
Splunk: Druid on Kubernetes with Druid-operator
Imply
 
Sparkler Presentation for Spark Summit East 2017
Sparkler Presentation for Spark Summit East 2017Sparkler Presentation for Spark Summit East 2017
Sparkler Presentation for Spark Summit East 2017
Karanjeet Singh
 
Sponsored Talk @ PGConf APAC 2018 - Migrating Oracle to EDB Postgres Approach...
Sponsored Talk @ PGConf APAC 2018 - Migrating Oracle to EDB Postgres Approach...Sponsored Talk @ PGConf APAC 2018 - Migrating Oracle to EDB Postgres Approach...
Sponsored Talk @ PGConf APAC 2018 - Migrating Oracle to EDB Postgres Approach...
PGConf APAC
 
Apache Druid Vision and Roadmap
Apache Druid Vision and RoadmapApache Druid Vision and Roadmap
Apache Druid Vision and Roadmap
Imply
 
Next Generation Big Data Platform at Netflix 2014
Next Generation Big Data Platform at Netflix 2014Next Generation Big Data Platform at Netflix 2014
Next Generation Big Data Platform at Netflix 2014
Eva Tse
 
Big Telco - Yousun Jeong
Big Telco - Yousun JeongBig Telco - Yousun Jeong
Big Telco - Yousun Jeong
Spark Summit
 
Pig on spark
Pig on sparkPig on spark
Pig on spark
Sigmoid
 
Ignite Your Big Data With a Spark!
Ignite Your Big Data With a Spark!Ignite Your Big Data With a Spark!
Ignite Your Big Data With a Spark!
Progress
 
A Non-Standard use Case of Hadoop: High Scale Image Processing and Analytics
A Non-Standard use Case of Hadoop: High Scale Image Processing and AnalyticsA Non-Standard use Case of Hadoop: High Scale Image Processing and Analytics
A Non-Standard use Case of Hadoop: High Scale Image Processing and Analytics
DataWorks Summit
 
Big data hadoop
Big data  hadoopBig data  hadoop
Big data hadoop
Vikram Dey
 
10 Things About Spark
10 Things About Spark 10 Things About Spark
10 Things About Spark
Roger Brinkley
 
Beginner Apache Spark Presentation
Beginner Apache Spark PresentationBeginner Apache Spark Presentation
Beginner Apache Spark Presentation
Nidhin Pattaniyil
 

What's hot (20)

Data Analytics and Processing at Snap - Druid Meetup LA - September 2018
Data Analytics and Processing at Snap - Druid Meetup LA - September 2018Data Analytics and Processing at Snap - Druid Meetup LA - September 2018
Data Analytics and Processing at Snap - Druid Meetup LA - September 2018
 
Talend spark meetup 03042017 - Paris Spark Meetup
Talend spark meetup 03042017 - Paris Spark MeetupTalend spark meetup 03042017 - Paris Spark Meetup
Talend spark meetup 03042017 - Paris Spark Meetup
 
963
963963
963
 
Druid: Under the Covers (Virtual Meetup)
Druid: Under the Covers (Virtual Meetup)Druid: Under the Covers (Virtual Meetup)
Druid: Under the Covers (Virtual Meetup)
 
Secured (Kerberos-based) Spark Notebook for Data Science: Spark Summit East t...
Secured (Kerberos-based) Spark Notebook for Data Science: Spark Summit East t...Secured (Kerberos-based) Spark Notebook for Data Science: Spark Summit East t...
Secured (Kerberos-based) Spark Notebook for Data Science: Spark Summit East t...
 
How Netflix Uses Druid in Real-time to Ensure a High Quality Streaming Experi...
How Netflix Uses Druid in Real-time to Ensure a High Quality Streaming Experi...How Netflix Uses Druid in Real-time to Ensure a High Quality Streaming Experi...
How Netflix Uses Druid in Real-time to Ensure a High Quality Streaming Experi...
 
Big Data and Hadoop - key drivers, ecosystem and use cases
Big Data and Hadoop - key drivers, ecosystem and use casesBig Data and Hadoop - key drivers, ecosystem and use cases
Big Data and Hadoop - key drivers, ecosystem and use cases
 
Apache Druid®: A Dance of Distributed Processes
 Apache Druid®: A Dance of Distributed Processes Apache Druid®: A Dance of Distributed Processes
Apache Druid®: A Dance of Distributed Processes
 
Splunk: Druid on Kubernetes with Druid-operator
Splunk: Druid on Kubernetes with Druid-operatorSplunk: Druid on Kubernetes with Druid-operator
Splunk: Druid on Kubernetes with Druid-operator
 
Sparkler Presentation for Spark Summit East 2017
Sparkler Presentation for Spark Summit East 2017Sparkler Presentation for Spark Summit East 2017
Sparkler Presentation for Spark Summit East 2017
 
Sponsored Talk @ PGConf APAC 2018 - Migrating Oracle to EDB Postgres Approach...
Sponsored Talk @ PGConf APAC 2018 - Migrating Oracle to EDB Postgres Approach...Sponsored Talk @ PGConf APAC 2018 - Migrating Oracle to EDB Postgres Approach...
Sponsored Talk @ PGConf APAC 2018 - Migrating Oracle to EDB Postgres Approach...
 
Apache Druid Vision and Roadmap
Apache Druid Vision and RoadmapApache Druid Vision and Roadmap
Apache Druid Vision and Roadmap
 
Next Generation Big Data Platform at Netflix 2014
Next Generation Big Data Platform at Netflix 2014Next Generation Big Data Platform at Netflix 2014
Next Generation Big Data Platform at Netflix 2014
 
Big Telco - Yousun Jeong
Big Telco - Yousun JeongBig Telco - Yousun Jeong
Big Telco - Yousun Jeong
 
Pig on spark
Pig on sparkPig on spark
Pig on spark
 
Ignite Your Big Data With a Spark!
Ignite Your Big Data With a Spark!Ignite Your Big Data With a Spark!
Ignite Your Big Data With a Spark!
 
A Non-Standard use Case of Hadoop: High Scale Image Processing and Analytics
A Non-Standard use Case of Hadoop: High Scale Image Processing and AnalyticsA Non-Standard use Case of Hadoop: High Scale Image Processing and Analytics
A Non-Standard use Case of Hadoop: High Scale Image Processing and Analytics
 
Big data hadoop
Big data  hadoopBig data  hadoop
Big data hadoop
 
10 Things About Spark
10 Things About Spark 10 Things About Spark
10 Things About Spark
 
Beginner Apache Spark Presentation
Beginner Apache Spark PresentationBeginner Apache Spark Presentation
Beginner Apache Spark Presentation
 

Viewers also liked

Fresher_CV_BE_EC
Fresher_CV_BE_ECFresher_CV_BE_EC
Fresher_CV_BE_EC
Ashvini Kale
 
Ashu Rajdor seo_resume
Ashu Rajdor seo_resumeAshu Rajdor seo_resume
Ashu Rajdor seo_resume
Ashu Rajdor
 
Downloadmela.com mba-fresher-resume(1)
Downloadmela.com  mba-fresher-resume(1)Downloadmela.com  mba-fresher-resume(1)
Downloadmela.com mba-fresher-resume(1)
sumanta12345
 
Nirav Shukla CV fresher in foreign trade
Nirav Shukla CV fresher in foreign tradeNirav Shukla CV fresher in foreign trade
Nirav Shukla CV fresher in foreign trade
NIRAV SHUKLA
 
resume
resumeresume
Tasmiya_MBA_fresher CV
Tasmiya_MBA_fresher CVTasmiya_MBA_fresher CV
Tasmiya_MBA_fresher CV
Thasmiya Begum
 
Práctica i tecnología aplicada a la educación
Práctica i tecnología aplicada a la educaciónPráctica i tecnología aplicada a la educación
Práctica i tecnología aplicada a la educación
Margarita Acosta Vallejo
 
Ponchado de cable. luz
Ponchado  de  cable. luzPonchado  de  cable. luz
Ponchado de cable. luz
luzr131
 
Proyecto de vida
Proyecto de vida Proyecto de vida
Proyecto de vida
Nayelis Jaimes Garrido
 
앙트십유스 INTRODUCTION
앙트십유스 INTRODUCTION앙트십유스 INTRODUCTION
앙트십유스 INTRODUCTION
hosan Ryu
 
Arvind skylands
Arvind skylandsArvind skylands
Arvind skylands
propladder
 
8 3-8 el plástico, uno de los mayores contaminantes.Brenda Murillo
8 3-8   el plástico, uno de los mayores contaminantes.Brenda Murillo 8 3-8   el plástico, uno de los mayores contaminantes.Brenda Murillo
8 3-8 el plástico, uno de los mayores contaminantes.Brenda Murillo
katheryne14
 
Kiran Harrison
Kiran HarrisonKiran Harrison
Kiran Harrison
Kiran Harrison John
 
Generación y selección de ideas de negocio
Generación y selección de ideas de negocioGeneración y selección de ideas de negocio
Generación y selección de ideas de negocio
sergioyeyo1688
 

Viewers also liked (14)

Fresher_CV_BE_EC
Fresher_CV_BE_ECFresher_CV_BE_EC
Fresher_CV_BE_EC
 
Ashu Rajdor seo_resume
Ashu Rajdor seo_resumeAshu Rajdor seo_resume
Ashu Rajdor seo_resume
 
Downloadmela.com mba-fresher-resume(1)
Downloadmela.com  mba-fresher-resume(1)Downloadmela.com  mba-fresher-resume(1)
Downloadmela.com mba-fresher-resume(1)
 
Nirav Shukla CV fresher in foreign trade
Nirav Shukla CV fresher in foreign tradeNirav Shukla CV fresher in foreign trade
Nirav Shukla CV fresher in foreign trade
 
resume
resumeresume
resume
 
Tasmiya_MBA_fresher CV
Tasmiya_MBA_fresher CVTasmiya_MBA_fresher CV
Tasmiya_MBA_fresher CV
 
Práctica i tecnología aplicada a la educación
Práctica i tecnología aplicada a la educaciónPráctica i tecnología aplicada a la educación
Práctica i tecnología aplicada a la educación
 
Ponchado de cable. luz
Ponchado  de  cable. luzPonchado  de  cable. luz
Ponchado de cable. luz
 
Proyecto de vida
Proyecto de vida Proyecto de vida
Proyecto de vida
 
앙트십유스 INTRODUCTION
앙트십유스 INTRODUCTION앙트십유스 INTRODUCTION
앙트십유스 INTRODUCTION
 
Arvind skylands
Arvind skylandsArvind skylands
Arvind skylands
 
8 3-8 el plástico, uno de los mayores contaminantes.Brenda Murillo
8 3-8   el plástico, uno de los mayores contaminantes.Brenda Murillo 8 3-8   el plástico, uno de los mayores contaminantes.Brenda Murillo
8 3-8 el plástico, uno de los mayores contaminantes.Brenda Murillo
 
Kiran Harrison
Kiran HarrisonKiran Harrison
Kiran Harrison
 
Generación y selección de ideas de negocio
Generación y selección de ideas de negocioGeneración y selección de ideas de negocio
Generación y selección de ideas de negocio
 

Similar to Aditi_Wadekar_Resume1

Resume_VipinKP
Resume_VipinKPResume_VipinKP
Resume_VipinKP
indhuparvathy
 
Manikyam_Hadoop_5+Years
Manikyam_Hadoop_5+YearsManikyam_Hadoop_5+Years
Manikyam_Hadoop_5+Years
Manikyam M
 
hadoop exp
hadoop exphadoop exp
Kalyan Hadoop
Kalyan HadoopKalyan Hadoop
Kalyan Hadoop
Canarys
 
Nagarjuna_Damarla_Resume
Nagarjuna_Damarla_ResumeNagarjuna_Damarla_Resume
Nagarjuna_Damarla_Resume
Nag Arjun
 
Shubham, 7.5+ years exp, mcp, map r spark-hive-bi-etl-azure-dataengineer-ml
Shubham, 7.5+ years exp, mcp, map r spark-hive-bi-etl-azure-dataengineer-mlShubham, 7.5+ years exp, mcp, map r spark-hive-bi-etl-azure-dataengineer-ml
Shubham, 7.5+ years exp, mcp, map r spark-hive-bi-etl-azure-dataengineer-ml
Shubham Mallick
 
Nagarjuna_Damarla
Nagarjuna_DamarlaNagarjuna_Damarla
Nagarjuna_Damarla
Nag Arjun
 
Deepankar Sehdev- Resume2015
Deepankar Sehdev- Resume2015Deepankar Sehdev- Resume2015
Deepankar Sehdev- Resume2015
Deepankar Sehdev
 
DeepeshRehi
DeepeshRehiDeepeshRehi
DeepeshRehi
deepesh rehi
 
HimaBindu
HimaBinduHimaBindu
RESUME_N
RESUME_NRESUME_N
Hadoop Developer
Hadoop DeveloperHadoop Developer
Hadoop Developer
mallikarjunkoriindia
 
Poorna Hadoop
Poorna HadoopPoorna Hadoop
Resume
ResumeResume
Resume_Karthick
Resume_KarthickResume_Karthick
Resume_Karthick
Karthick Selvaraj
 
KOTI_RESUME_(1) (2)
KOTI_RESUME_(1) (2)KOTI_RESUME_(1) (2)
KOTI_RESUME_(1) (2)
ch koti
 
Prashanth Kumar_Hadoop_NEW
Prashanth Kumar_Hadoop_NEWPrashanth Kumar_Hadoop_NEW
Prashanth Kumar_Hadoop_NEW
Prashanth Shankar kumar
 
Continuuity Presents at Under the Radar 2013
Continuuity Presents at Under the Radar 2013Continuuity Presents at Under the Radar 2013
Continuuity Presents at Under the Radar 2013
Dealmaker Media
 
HariKrishna4+_cv
HariKrishna4+_cvHariKrishna4+_cv
HariKrishna4+_cv
revuri
 
Rajeev kumar apache_spark & scala developer
Rajeev kumar apache_spark & scala developerRajeev kumar apache_spark & scala developer
Rajeev kumar apache_spark & scala developer
Rajeev Kumar
 

Similar to Aditi_Wadekar_Resume1 (20)

Resume_VipinKP
Resume_VipinKPResume_VipinKP
Resume_VipinKP
 
Manikyam_Hadoop_5+Years
Manikyam_Hadoop_5+YearsManikyam_Hadoop_5+Years
Manikyam_Hadoop_5+Years
 
hadoop exp
hadoop exphadoop exp
hadoop exp
 
Kalyan Hadoop
Kalyan HadoopKalyan Hadoop
Kalyan Hadoop
 
Nagarjuna_Damarla_Resume
Nagarjuna_Damarla_ResumeNagarjuna_Damarla_Resume
Nagarjuna_Damarla_Resume
 
Shubham, 7.5+ years exp, mcp, map r spark-hive-bi-etl-azure-dataengineer-ml
Shubham, 7.5+ years exp, mcp, map r spark-hive-bi-etl-azure-dataengineer-mlShubham, 7.5+ years exp, mcp, map r spark-hive-bi-etl-azure-dataengineer-ml
Shubham, 7.5+ years exp, mcp, map r spark-hive-bi-etl-azure-dataengineer-ml
 
Nagarjuna_Damarla
Nagarjuna_DamarlaNagarjuna_Damarla
Nagarjuna_Damarla
 
Deepankar Sehdev- Resume2015
Deepankar Sehdev- Resume2015Deepankar Sehdev- Resume2015
Deepankar Sehdev- Resume2015
 
DeepeshRehi
DeepeshRehiDeepeshRehi
DeepeshRehi
 
HimaBindu
HimaBinduHimaBindu
HimaBindu
 
RESUME_N
RESUME_NRESUME_N
RESUME_N
 
Hadoop Developer
Hadoop DeveloperHadoop Developer
Hadoop Developer
 
Poorna Hadoop
Poorna HadoopPoorna Hadoop
Poorna Hadoop
 
Resume
ResumeResume
Resume
 
Resume_Karthick
Resume_KarthickResume_Karthick
Resume_Karthick
 
KOTI_RESUME_(1) (2)
KOTI_RESUME_(1) (2)KOTI_RESUME_(1) (2)
KOTI_RESUME_(1) (2)
 
Prashanth Kumar_Hadoop_NEW
Prashanth Kumar_Hadoop_NEWPrashanth Kumar_Hadoop_NEW
Prashanth Kumar_Hadoop_NEW
 
Continuuity Presents at Under the Radar 2013
Continuuity Presents at Under the Radar 2013Continuuity Presents at Under the Radar 2013
Continuuity Presents at Under the Radar 2013
 
HariKrishna4+_cv
HariKrishna4+_cvHariKrishna4+_cv
HariKrishna4+_cv
 
Rajeev kumar apache_spark & scala developer
Rajeev kumar apache_spark & scala developerRajeev kumar apache_spark & scala developer
Rajeev kumar apache_spark & scala developer
 

Aditi_Wadekar_Resume1

  • 1. Email: aditiwadekar10@gmail.com| Mobile: +91-87933-49212 BIG DATA –PROFESSIONAL Working in Big Data with a two plus years of IT experience. I am part of Opus Consulting Solutions Analytics Team. My responsibilities are to setup, configure & manage Hadoop and MongoDB cluster. Perform Data Migration Activities. I am looking for an organization that offers professional growth and provides an environment conducive to excellence and enhance my ability to be an achiever. Big Data Skills: MongoDB  Created Sharded cluster with replica Set on AWS production cluster  Worked with Authentication and Authorization on cluster  Installed and configured Nagios for monitoring shard cluster  Tested Node migration, Cluster upgradation and Backup-recovery on MongoDB cluster  Worked with Aggregate queries to analyze data Hadoop  Hands on experience with Apache Hadoop  Experimented tools like Hbase, Pig, Hive by inserting, updating and performing query operations  Implemented Spark on Real time data  Installed and configured cloudera-manager for testing purpose BI-Tool  Data Migration using Pentaho data Integration tools.  Used Hunk/Splunk to generate and analyze reports Hands on Experience with AWS Environment  EC2, volumes, snapshot, S3, IAM Other Technical Skills:  Operation System: CentOS, RHEL,Ubuntu, Windows.  Databases: MongoDB, Mysql, SQL Server  Tools: Hunk/Splunk, Pentaho data Integration tools.  Languages: Python  Virtualization: VMware, Virtual Box  Scripting/Automation: Bash, Shell Scripting  Big Data: Ambari, HDFS, Hive HA, Spark o Hue ,Oozie, YARN/Map reduce, NoSQL, Sqoop o MongoDB, AWS Process Oriented skills: Worked for the Complete Lifecycle Requirement gathering, Analysis, Build, Unit testing, Migration, Production deployment & Support activities.
  • 2. CERTIFICATIONS 1. Certified on MongoDB Advanced Deployment and Operations August 2016 2. Initiation on Real time data analytics over Hadoop and MongoDB April 2015 3. Outstanding Team – “Cardtronics Big Data Team” April 2014-Sept 2014 4. Certified Courses ○ Hadoop 101 Sept 1, 2016 ○ Introduction to Pig May 8, 2015 ○ Using HBase for Real-time Access to your Big Data March 8, 2015 ○ Spark Fundamentals I Jan 20, 2015 ○ M102: MongoDB for DBAs Dec 1, 2014 ○ Certified in Hadoop Fundamentals -1 Oct 27, 2014 ○ M101J MongoDB for Java Developers Oct 7, 2014 AWA RDS/RECOGNITIONS  Reward and Recognition obtained for outstanding Teamwork of Big Data Team  Reward obtained for taking initiation on Real time data analytics over Hadoop and MongoDB.  Active participation in Opus Women Football Team and Women Cricket Team. EXPERIENCE Opus Consulting Solution 2014 – Present 1. Implemented a project required by the business users to analyze the number of transactions taking place location and month wise across US, Jan 2016 ○ Data migration and validation from MongoDB to Hadoop. ○ Installing and Configuring Splunk to connect Hadoop and MongoDB. ○ Created Providers and indexes in Splunk to connect hdfs. ○ Developed Dashboards in Splunk based on input parameters ○ Scripts written in Python to load data to Hadoop. ○ Developed, Configured and scheduled reports using Splunk ○ Developed SPL out of Stored Procedures and developed Reports. ○ Created users and roles in Splunk. ○ Query on Splunk to generate 20+ reports out of Hadoop. Roles and Responsibility ○ Perform Requirement gathering to obtain business logic of the report ○ Provide estimates for development and testing of reports ○ Understand the business logic of the report. ○ Designed and Developed Reports in Hunk ○ Enhanced skills in SQL and SPL (Search Processing Language) while developing reports
  • 3. 2. Worked on Implementing Spark on Real-Time transactional Data, Dec 2014 ○ Installed and configured spark with MongoDB and Hadoop on CentOS 6.5. MongoDB contained real-time data. Spark streaming was used to deal with live streaming data. Spark was connected to MongoDB and various queries were performed on the transactional data. The implementation helped in overcoming the drawbacks of MongoDB that is joins. Queries were performed in faster and efficient way on huge volume of data. ○ Apache Zeppelin was installed and configured with spark. The reports were being developed in zeppelin and the transactional data was being analyzed in an efficient way Roles and Responsibility ○ Requirement gathering and provide design of how spark can be implemented on Real time data. ○ Installation and configuration of MongoDB and Hadoop on Servers. ○ Developed queries in Spark and Zeppelin to connect and query over MongoDB and Hadoop data. ○ The query was improvised to check the efficiency of the tool. ○ Reports generated out of Zeppelin 3. Created Hadoop and MongoDB as Data Warehouse, April 2015 ○ Installed and configured MongoDB sharded cluster on AWS production. The aim was to create Hadoop as a data warehouse. ○ Setup Hadoop cluster along with Ambari, Hue on Environment. ○ Performed data migration Activity using sqoop from SQL to Hadoop. ○ Familiar with Hbase, Pig, Hive by inserting, updating and performing query operations ○ Developed Backup-scripts for taking MongoDB backup to prevent loss of data. ○ Used Pentaho-data integration tool to migrate data from SQL to MongoDB ○ Developed scripts in python to migrate data from MongoDB to Hadoop cluster. ○ Validate the data in MongoDB and Hadoop. ○ Hadoop consisted of 3 years of data. Roles and Responsibility ○ Setting up Hadoop cluster and monitoring the cluster. ○ Perform Data Migration Activity from SQL to Hadoop or from MongoDB to Hadoop ○ Create roles and Users in hive and grant permissions. ○ Setting up MongoDB sharded cluster with SSL authentication and authorization. ○ Developed MongoDB Backup Scripts. ○ Installed and Configured Nagios ○ Monitored MongoDB cluster using Nagios. ○ Developed Aggregate queries in MongoDB to test the data in collections.
  • 4. Candidate for Bachelors in Applied Science, Information Technology from College of Engineering, Pune - 7.68 (Sept 2010-Aug2014)