Cloudera's open-source Apache Hadoop distribution, CDH (Cloudera Distribution Including Apache Hadoop), targets enterprise-class deployments of that technology. Cloudera says that more than 50% of its engineering output is donated upstream to the various Apache-licensed open source projects.
https://www.pass4sureexam.com/ccD-410.html
Webinar | From Zero to Big Data Answers in Less Than an Hour – Live Demo SlidesCloudera, Inc.
Slides describing Cloudera and Karmasphere, and how combined their products can install a Hadoop cluster, import data, run queries and generate results.
Cloudera's open-source Apache Hadoop distribution, CDH (Cloudera Distribution Including Apache Hadoop), targets enterprise-class deployments of that technology. Cloudera says that more than 50% of its engineering output is donated upstream to the various Apache-licensed open source projects.
https://www.pass4sureexam.com/ccD-410.html
Webinar | From Zero to Big Data Answers in Less Than an Hour – Live Demo SlidesCloudera, Inc.
Slides describing Cloudera and Karmasphere, and how combined their products can install a Hadoop cluster, import data, run queries and generate results.
Big Data Analytics - Is Your Elephant Enterprise Ready?Hortonworks
Hadoop’s cost effective scalability and flexibility to analyze all data types is driving organizations everywhere to embrace big data analytics. From proof of concept to deployment across the enterprise, join Datameer and Hortonworks as we answer the ‘now what?’ when rolling out your Hadoop big data analytics project. This webinar will address critical project components such as data security, data privacy, high availability, user training and use case development.
Strata + Hadoop World 2012: Data Science on Hadoop: How Cloudera Impala Unloc...Cloudera, Inc.
This talk will cover what tools and techniques work and don’t work well for data scientists working on Hadoop today and how to leverage the lessons learned by the experts to increase your productivity as well as what to expect for the future of data science on Hadoop. We will leverage insights derived from the top data scientists working on big data systems at Cloudera as well as experiences from running big data systems at Facebook, Google, and Yahoo.
Oracle Unified Information Architeture + Analytics by ExampleHarald Erb
Der Vortrag gibt zunächst einen Architektur-Überblick zu den UIA-Komponenten und deren Zusammenspiel. Anhand eines Use Cases wird vorgestellt, wie im "UIA Data Reservoir" einerseits kostengünstig aktuelle Daten "as is" in einem Hadoop File System (HDFS) und andererseits veredelte Daten in einem Oracle 12c Data Warehouse miteinander kombiniert oder auch per Direktzugriff in Oracle Business Intelligence ausgewertet bzw. mit Endeca Information Discovery auf neue Zusammenhänge untersucht werden.
With the advent of Hadoop, there comes the need for professionals skilled in Hadoop Administration making it imperative to be skilled as a Hadoop Admin for better career, salary and job opportunities.
Upgrade Without the Headache: Best Practices for Upgrading Hadoop in ProductionCloudera, Inc.
Walk through some of the best practices to keep in mind when it comes to upgrading your cluster, and learn how to leverage new Upgrade Wizard features in Cloudera Enterprise 5.3.
For most mission critical workloads, downtime is never an option. Any downtime can have a direct impact on revenue and lead to frantic calls in the middle of the night. For this reason, upgrading the software that powers these workloads can often be a daunting task. It can cause unpredictable issues without access to support. That’s why an enterprise-grade administration tool is crucial for running Hadoop in production. Hadoop consists of dozens of components, running across multiple machines, all with their own configurations. That can lead to a lot of complexity and uncertainty - especially when taking the upgrade plunge.
Cloudera Manager makes it easy and is the only production-ready administration tool for Hadoop. Not only does Cloudera Manager feature zero-downtime rolling upgrades, but it also has a built in Upgrade Wizard to make upgrades simple and predictable.
What it takes to bring Hadoop to a production-ready stateClouderaUserGroups
While Hadoop may be a hot topic and is probably the buzziest big data term, the fact is that many Hadoop projects get stuck in pilot mode. We hear a number of reasons for this.
• “It’s too complicated.”
• “I don’t have the right resources.”
• “Security and compliance are never going to approve this.”
This session digs deep into why certain projects seem destined to remain in development. We’ll also cover what it takes to bring Hadoop to a production-ready state and convince management that it’s time to start using Hadoop to store and analyze real business data.
Overview of Big data, Hadoop and Microsoft BI - version1Thanh Nguyen
Big Data and advanced analytics are critical topics for executives today. But many still aren't sure how to turn that promise into value. This presentation provides an overview of 16 examples and use cases that lay out the different ways companies have approached the issue and found value: everything from pricing flexibility to customer preference management to credit risk analysis to fraud protection and discount targeting. For the latest on Big Data & Advanced Analytics: http://mckinseyonmarketingandsales.com/topics/big-data
What the Enterprise Requires - Business Continuity and VisibilityCloudera, Inc.
Cloudera Enterprise BDR delivers centralized disaster recovery for data and metadata, enabling you to prepare for disaster by moving data to your secondary site automatically. Cloudera Navigator 1.0 provides data governance capabilities such as verifying access privileges and auditing access to all data stored in Hadoop, which are critical for customers that are in highly regulated industries and have stringent compliance requirements.
This presentation will teach you how to:
- Centrally configure and manage replication workflows for files (HDFS) and metadata (Hive)
- Consistently meet or exceed SLAs and RTOs through simplified management and process automation
- Track access permissions and actual accesses to all data objects in Hive, HBase, and HDFS
- Answer the questions:
- Who has access to which data object(s)
- Which data objects were accessed by a user
- When was a data object accessed and by whom
- What data assets were accessed using a service
- Which device was used to access
• Capable of processing large sets of structured, semi-structured and unstructured data and supporting system architecture
• Implemented Proof of concepts on Hadoop stack and different big data analytic tools, migration from different databases to Hadoop.
• Developed multiple Map Reduce jobs in java for data cleaning and pre-processing according to the business requirements, Importing and exporting data into HDFS and Hive using Sqoop.
Having Experience in writing HIVE queries & Pig scripts.
Big Data Analytics - Is Your Elephant Enterprise Ready?Hortonworks
Hadoop’s cost effective scalability and flexibility to analyze all data types is driving organizations everywhere to embrace big data analytics. From proof of concept to deployment across the enterprise, join Datameer and Hortonworks as we answer the ‘now what?’ when rolling out your Hadoop big data analytics project. This webinar will address critical project components such as data security, data privacy, high availability, user training and use case development.
Strata + Hadoop World 2012: Data Science on Hadoop: How Cloudera Impala Unloc...Cloudera, Inc.
This talk will cover what tools and techniques work and don’t work well for data scientists working on Hadoop today and how to leverage the lessons learned by the experts to increase your productivity as well as what to expect for the future of data science on Hadoop. We will leverage insights derived from the top data scientists working on big data systems at Cloudera as well as experiences from running big data systems at Facebook, Google, and Yahoo.
Oracle Unified Information Architeture + Analytics by ExampleHarald Erb
Der Vortrag gibt zunächst einen Architektur-Überblick zu den UIA-Komponenten und deren Zusammenspiel. Anhand eines Use Cases wird vorgestellt, wie im "UIA Data Reservoir" einerseits kostengünstig aktuelle Daten "as is" in einem Hadoop File System (HDFS) und andererseits veredelte Daten in einem Oracle 12c Data Warehouse miteinander kombiniert oder auch per Direktzugriff in Oracle Business Intelligence ausgewertet bzw. mit Endeca Information Discovery auf neue Zusammenhänge untersucht werden.
With the advent of Hadoop, there comes the need for professionals skilled in Hadoop Administration making it imperative to be skilled as a Hadoop Admin for better career, salary and job opportunities.
Upgrade Without the Headache: Best Practices for Upgrading Hadoop in ProductionCloudera, Inc.
Walk through some of the best practices to keep in mind when it comes to upgrading your cluster, and learn how to leverage new Upgrade Wizard features in Cloudera Enterprise 5.3.
For most mission critical workloads, downtime is never an option. Any downtime can have a direct impact on revenue and lead to frantic calls in the middle of the night. For this reason, upgrading the software that powers these workloads can often be a daunting task. It can cause unpredictable issues without access to support. That’s why an enterprise-grade administration tool is crucial for running Hadoop in production. Hadoop consists of dozens of components, running across multiple machines, all with their own configurations. That can lead to a lot of complexity and uncertainty - especially when taking the upgrade plunge.
Cloudera Manager makes it easy and is the only production-ready administration tool for Hadoop. Not only does Cloudera Manager feature zero-downtime rolling upgrades, but it also has a built in Upgrade Wizard to make upgrades simple and predictable.
What it takes to bring Hadoop to a production-ready stateClouderaUserGroups
While Hadoop may be a hot topic and is probably the buzziest big data term, the fact is that many Hadoop projects get stuck in pilot mode. We hear a number of reasons for this.
• “It’s too complicated.”
• “I don’t have the right resources.”
• “Security and compliance are never going to approve this.”
This session digs deep into why certain projects seem destined to remain in development. We’ll also cover what it takes to bring Hadoop to a production-ready state and convince management that it’s time to start using Hadoop to store and analyze real business data.
Overview of Big data, Hadoop and Microsoft BI - version1Thanh Nguyen
Big Data and advanced analytics are critical topics for executives today. But many still aren't sure how to turn that promise into value. This presentation provides an overview of 16 examples and use cases that lay out the different ways companies have approached the issue and found value: everything from pricing flexibility to customer preference management to credit risk analysis to fraud protection and discount targeting. For the latest on Big Data & Advanced Analytics: http://mckinseyonmarketingandsales.com/topics/big-data
What the Enterprise Requires - Business Continuity and VisibilityCloudera, Inc.
Cloudera Enterprise BDR delivers centralized disaster recovery for data and metadata, enabling you to prepare for disaster by moving data to your secondary site automatically. Cloudera Navigator 1.0 provides data governance capabilities such as verifying access privileges and auditing access to all data stored in Hadoop, which are critical for customers that are in highly regulated industries and have stringent compliance requirements.
This presentation will teach you how to:
- Centrally configure and manage replication workflows for files (HDFS) and metadata (Hive)
- Consistently meet or exceed SLAs and RTOs through simplified management and process automation
- Track access permissions and actual accesses to all data objects in Hive, HBase, and HDFS
- Answer the questions:
- Who has access to which data object(s)
- Which data objects were accessed by a user
- When was a data object accessed and by whom
- What data assets were accessed using a service
- Which device was used to access
• Capable of processing large sets of structured, semi-structured and unstructured data and supporting system architecture
• Implemented Proof of concepts on Hadoop stack and different big data analytic tools, migration from different databases to Hadoop.
• Developed multiple Map Reduce jobs in java for data cleaning and pre-processing according to the business requirements, Importing and exporting data into HDFS and Hive using Sqoop.
Having Experience in writing HIVE queries & Pig scripts.
• Excellent analytical and problem solving skills.
• Excellent communication skills.
• Quick Learner, Self-Motivated and team player traits.
• Ability to mentor and educate peers whenever needed for the greater good of the team as a whole.
1. MONIKA RAGHUVANSHI
Hadoop Administrator
“Transforminglarge, unruly data sets into competitive advantages”
+1 2407083996
moni.raghuvansh@gmail.com
CAREER OBJECTIVE
To pursue a growth orientedcareer witha progressive companythat provides a scope to applymyknowledge andskills that wouldhelp me contribute
my best to the organization.
To demonstrate myexpertise as HadoopAdministrator to ensure Hadoop cluster administration and technical support in large scale IT industry.
PROFILESUMMARY
Purveyorof competitive intelligence andholistic, timelyanalysis of BigDatamade possible by the successful installation, configuration, administration
and maintenance of Hadoop ecosystem components and architecture.
7 Years of total IT experience with strong exposure in Hadoop and Unix administration.
Excellent understanding of Hadoop architecture and Map Reduce (MR1) & YARN Frameworks.
Hands on experiencein installing, configuringandusingHadoop ecosystem components like MapReduce, HDFS, Hive, Hbase, Zoo Keeper, Oozie,
Sqoop, Flume,impala, spark, storm, hue .
Practical knowledge on functionalities ofall Hadoop daemons, interaction between them, resource utilizations and dynamic tuning to make cluster
available and efficient.
Ensuring hadoop cluster security by implementing SSL and setting up Kerberos enviornment.
Experience with troubleshooting Hadoop performance issues as job failures, slow running queries and cluster issues.
Proficient in project management activities which includes planning, design, scope management, estimation, resource administr ation and project
completion as per the quality parameters and best practice guidelines.
Able to assess business rules, collaborate with stakeholders and perform source-to-target data mapping, design and review.
Expertise with full Project Life CycleandProject Management, Standards and Configuration Management, QA Management, Team Management,
documentation, Implementation, project deployment etc.
Hands on experience in all aspects of software development life cycle (SDLC) in Waterfall & Agile enviornments.
Knowledge of ISO, Capability Maturity Model Integration (CMMI), and IEEE processes.
Experienced in analysis, design of manual and automated testing for Client/Server and Web-based applications.
Strong interpersonal skills with the experience of working in a multi-cultural environment.
Excellent written and oral Communication skills.
TECHNICAL EXPERTISE
Hadoop Distributions :
Apache Hadoop
Cloudera
Excellent installationandadministration skills of Hadoopcomponents:
HDFS
MapReduce
Hive
HBase
Flume
Sqoop
Zookeeper
Proficient in supporting Linux operatingsystems:
CentOS
Ubuntu
RHEL
Solid understandingof open source monitoringtools:
Nagios
Ganglia
Cloudera Manager
Hue
Familiaritywith networks:
TCP/IP
Firewall
DNS
Exceptional in overseeingsystemadministrationoperations:
2. Performance tuning
Storage capacity management
Skilled in programming/scriptinglanguages:
Sell Scripting
Core Java
Well-versedin databases:
Oracle
MS SQL
MY-SQL
PROFESSIONAL EXPERIENCESUMMARY
PROJECTS
Oct 2015 – Till date
1) Client : Barclays Financial
Location: Delaware
Role: HadoopAdmin.
Responsibilities:
Involved in Hadoop Cluster administration that includes commissioning & decommissioning of datanode, capacity planning,
performance tunning, cluster monitoring and troubleshooting.
Involved in Kerberos and SSL setup to ensure cluster security.
Working with data delivery teams to setup new Hadoop users which includes setting up Linux users (AD), setting up Kerberos
principals and testing HDFS, Hive, Pig and MapReduce access for the new users.
Installed and configured MapReduce, HIVE and the HDFS, Assisted with performance tuning and monitoring.
Workedon settingup high availabilityfor major production cluster and designed automatic failover control using zookeeper and
quorum journal nodes.
Performance Tuningandoptimizingclusters, toget best throughput usingtools like HIVE, Impala, Hbase, Spark.
Assemble newly bought hardware into racks with switches, assign IP addresses, firewalling, enable/disable ports etc.
Importing and exporting data into HDFS and Hive using Sqoop.
Configured Oozie for workflow automation and coordination.
Configured ZooKeeper to implement node coordination, in clustering support.
Modified cluster nodes and analyzed Hadoop log files.
Workingonthe issues anddefects reportedon CDH platform, anddrive them to closure in coordination withCloudera Support Team
where needed.
Jan 2013 – May 2015
2) Client : GE Healthcare
Location : Mumbai, India
Role:Hadoop Consultant.
Responsibilities:
Hands on experience with Apache Hadoop and CDH distributions.
Installed and configured multiple Apache and CDH Hadoop clusters.
Involvedin hadoop cluster planning, finalizing the architecture, onboarding the projects to the cluster, cluster monitoring and
maintenance.
Assisted with data capacity planning and node forecasting.
Translation of functional and technical requirements into detailed architecture and design.
Installation, monitoring, managing, troubleshooting, applyingpatches in different environments such as Development Cluster, Test
Cluster and Production environments.
Monitoring and controlling local file system disk space usage, local log files, cleaning log files with automated scripts.
As a Hadoop admin, monitoringcluster health status on daily basis, tuning system performance related configuration parameters.
Assuring cluster security by Kerberos and SSL/TSL setup on development cluster
Undertakingdesign & development,testing, debuggingandtroubleshootingof the application; administeringsmoothimplementation
of the application
Managingcode deployment, change management andqualityassurance; providingtechnical support toApplicationDevelopment
Teams andsetting/maintainingstandards.
Supportedtechnical teammembers forautomation, installation andconfigurationtasks.
Experience in setup, configurationandmanagement of Apache Sentry for Role-basedauthorizationandprivilege validation for Hive
and Impala Services.
Setting up the machines with Network Control, Static IP, Disabled Firewalls, Swap memory.
Suggested latest upgrades and patches for operating systems and Hadoop.
AutomatingManual and repetitivetasks usingshell scripts
Aligningwith the systems engineeringteamtopropose anddeploynewhardware andsoftware environments requiredfor Hadoopand
to expandexistingenvironments
Implemented rack aware topology on the Hadoop cluster.
Implemented Fair scheduler on the job tracker to allocate fair amount of resources to small jobs.
Implemented Kerberos Security Authentication protocol for existing cluster.
Good experience in troubleshooting performance and security level issues in the cluster and its functionality.
Regular Commissioning and Decommissioning of nodes depending upon the amount of data.
3. ManagingOperational Acceptance Testingof projects/applications developedon Hadoopplatform andtheir deployment toproduction
environment
Live deployment ofApplicationbuildon hadoopplatform andsupport for thefirst-run in production environment andlongterm
solution for incidents andcontributingtowards buildingstable andreliable systems.
Composingdevelopment planandeffort estimations for newproject/change requests anddistributingwork amongteammembers,
managingassigneddeliveries andprovidingend-to-endconsultancyon project life cycle
Executingcontinuous integration andcode deployment automationusingGit, Jenkins,Nexus andshell scripts.
Formulatedprocedures for planningandexecutionof systemupgrades forall existingHadoopclusters.
Participated in regular production support services to Hadoop infrastructure components.
April 2011 – Jan 2013
3) Client : OntarioMinistryof Transportation
Location : Mumbai, India
Role:UnixAdministrator.
Responsibilities:
Involved in support and monitoring production Linux Systems.
Managed client’s Unix environments by installing, supporting, monitoring and automating the latest unix servers.
Installation SQL and DB Backup.
Expertise in Archive logs and Monitoring the jobs.
Monitoring Linux daily jobs and monitoring log management system.
Expertise in troubleshooting and able to work with a team to fix large production issues.
Expertise in creating and managing DB tables, Index and Views.
User creation and managing user accounts and permissions on Linux level and DB level.
Expertise in Security in OS level and DB level.
Nov2008 - April 2011
4) Client : Nortel,Canada
Location : Mumbai,India
Role:UnixAdministrator.
Responsibilities:
Involved in Development and monitoring Application.
Prod patch and config implementation on production unix servers
Good Experience in Estimating work effort.
Performed backup and regression performance testing to decide the benchmark.
Expertise with Developing SQL scripts and Performance Tuning.
Expertise in Analyzing data Quality checks using shell scripts.
Expertise in Loading data into Data Base using Linux OS
Developing MapReduce Program to format the data.
Expertise in handling with Large Data Warehouses for pulling reports.
Expertise in preparing the HLDs and LDS and preparing Unit Test Cases based on functionality.
Expertise in Linux OS health Checks and debugging the issues.
Expertise in installing prm packages on Linux.
Expertise in Security Layer in OS permission level and DB table level.
Expertise in Alert System Nagios
TRAINING
Administrator Training for Apache Hadoop.
Agile Training
Software Testing Foundation and Advance Level.
Goal SettingandTime management
ACHIEVEMENTS
Rolta Star Award.
TCS Gems and Great Team Spirit Award
Great Individual performer award in GE Healthcare.
EDUCATIONAL QUALIFICATION
MBA (Information Systems) from ICFAI, India
Bachelor of Technology (BTech) from MIT Ujjain, India