This document is a curriculum vitae for Amith Rayappa that summarizes his work experience and qualifications. It outlines his 6 years of experience in IT and over 2 years working as a Hadoop Administrator, along with expertise in installing and managing Hadoop clusters. It also provides details of his past roles as a Siebel Administrator and his educational background, including a master's degree from BITS Pilani.
SUSE, Hadoop and Big Data Update. Stephen Mogg, SUSE UKhuguk
This session will give you an update on what SUSE is up to in the Big Data arena. We will take a brief look at SUSE Linux Enterprise Server and why it makes the perfect foundation for your Hadoop Deployment.
Hadoop Security and Compliance - StampedeCon 2016StampedeCon
As Hadoop becomes a mainstream data platform across organizations, securing a vast and growing volume of critical information, especially financial and healthcare data, is more essential than ever. In this presentation, Derek will elaborate how to leverage Big Data technologies without sacrificing security and compliance, and will focus specially on how comprehensive security mechanisms should be put in place to secure a production ready Hadoop environment. The presentation will also highlight technologies, such as encryption in-motion and at-rest for Hadoop services, as well as the complicated compliant processes to meet strictest regulatory requirements and standards.
What Is Hadoop | Hadoop Tutorial For Beginners | EdurekaEdureka!
( Hadoop Training: https://www.edureka.co/hadoop )
This Edureka "What is Hadoop" tutorial ( Hadoop Blog series: https://goo.gl/LFesy8 ) helps you to understand how Big Data emerged as a problem and how Hadoop solved that problem. This tutorial will be discussing about Hadoop Architecture, HDFS & it's architecture, YARN and MapReduce in detail. Below are the topics covered in this tutorial:
1) 5 V’s of Big Data
2) Problems with Big Data
3) Hadoop-as-a solution
4) What is Hadoop?
5) HDFS
6) YARN
7) MapReduce
8) Hadoop Ecosystem
Big Data Warehousing Meetup: Securing the Hadoop Ecosystem by ClouderaCaserta
In our recent Big Data Warehousing Meetup, we discussed Data Governance, Compliance and Security in Hadoop.
As the Big Data paradigm becomes more commonplace, we must apply enterprise-grade governance capabilities for critical data that is highly regulated and adhere to stringent compliance requirements. Caserta and Cloudera shared techniques and tools that enables data governance, compliance and security on Big Data.
For more information, visit www.casertaconcepts.com
Hadoop Security in Big-Data-as-a-Service Deployments - Presented at Hadoop Su...Abhiraj Butala
The talk covers limitations of current Hadoop eco-system components in handling security (Authentication, Authorization, Auditing) in multi-tenant, multi-application environments. Then it proposes how we can use Apache Ranger and HDFS super-user connections to enforce correct HDFS authorization policies and achieve the required auditing.
Hadoop 3.0 has been years in the making, and now it's finally arriving. Andrew Wang and Daniel Templeton offer an overview of new features, including HDFS erasure coding, YARN Timeline Service v2, YARN federation, and much more, and discuss current release management status and community testing efforts dedicated to making Hadoop 3.0 the best Hadoop major release yet.
PASS Summit - SQL Server 2017 Deep DiveTravis Wright
Deep dive into SQL Server 2017 covering SQL Server on Linux, containers, HA improvements, SQL graph, machine learning, python, adaptive query processing, and much much more.
Introduction To Hadoop Administration - SpringPeopleSpringPeople
The Hadoop framework is used by major players including Google, Yahoo and IBM, largely for applications involving search engines and advertising. The popularity of Hadoop is juts increasing exponentially.
This presentation discusses the follow topics
What is Hadoop?
Need for Hadoop
History of Hadoop
Hadoop Overview
Advantages and Disadvantages of Hadoop
Hadoop Distributed File System
Comparing: RDBMS vs. Hadoop
Advantages and Disadvantages of HDFS
Hadoop frameworks
Modules of Hadoop frameworks
Features of 'Hadoop‘
Hadoop Analytics Tools
Securing your Big Data Environments in the CloudDataWorks Summit
Big Data tools are becoming a critical part of enterprise architectures and as such securing the data, at rest, and in motion is a necessity. More so, when you’re implementing these solutions in the cloud and the data doesn't reside within the confines of your trusted data center. Also, there is a fine balance between implementing enterprise-grade security and negotiating utmost performance given the overheads of encryption and/or identity management.
This session is designed to tackle these challenges head on and explain the various options available in the cloud. The focal points are the implementation of tools like Ranger and Knox for cloud deployments, but we also pay attention to the security features offered in the cloud that complement this process and secure the data in unprecedented ways.
Cloud Security + OSS Security tools are a deadly combination, when it comes to securing your Data Lake.
SUSE, Hadoop and Big Data Update. Stephen Mogg, SUSE UKhuguk
This session will give you an update on what SUSE is up to in the Big Data arena. We will take a brief look at SUSE Linux Enterprise Server and why it makes the perfect foundation for your Hadoop Deployment.
Hadoop Security and Compliance - StampedeCon 2016StampedeCon
As Hadoop becomes a mainstream data platform across organizations, securing a vast and growing volume of critical information, especially financial and healthcare data, is more essential than ever. In this presentation, Derek will elaborate how to leverage Big Data technologies without sacrificing security and compliance, and will focus specially on how comprehensive security mechanisms should be put in place to secure a production ready Hadoop environment. The presentation will also highlight technologies, such as encryption in-motion and at-rest for Hadoop services, as well as the complicated compliant processes to meet strictest regulatory requirements and standards.
What Is Hadoop | Hadoop Tutorial For Beginners | EdurekaEdureka!
( Hadoop Training: https://www.edureka.co/hadoop )
This Edureka "What is Hadoop" tutorial ( Hadoop Blog series: https://goo.gl/LFesy8 ) helps you to understand how Big Data emerged as a problem and how Hadoop solved that problem. This tutorial will be discussing about Hadoop Architecture, HDFS & it's architecture, YARN and MapReduce in detail. Below are the topics covered in this tutorial:
1) 5 V’s of Big Data
2) Problems with Big Data
3) Hadoop-as-a solution
4) What is Hadoop?
5) HDFS
6) YARN
7) MapReduce
8) Hadoop Ecosystem
Big Data Warehousing Meetup: Securing the Hadoop Ecosystem by ClouderaCaserta
In our recent Big Data Warehousing Meetup, we discussed Data Governance, Compliance and Security in Hadoop.
As the Big Data paradigm becomes more commonplace, we must apply enterprise-grade governance capabilities for critical data that is highly regulated and adhere to stringent compliance requirements. Caserta and Cloudera shared techniques and tools that enables data governance, compliance and security on Big Data.
For more information, visit www.casertaconcepts.com
Hadoop Security in Big-Data-as-a-Service Deployments - Presented at Hadoop Su...Abhiraj Butala
The talk covers limitations of current Hadoop eco-system components in handling security (Authentication, Authorization, Auditing) in multi-tenant, multi-application environments. Then it proposes how we can use Apache Ranger and HDFS super-user connections to enforce correct HDFS authorization policies and achieve the required auditing.
Hadoop 3.0 has been years in the making, and now it's finally arriving. Andrew Wang and Daniel Templeton offer an overview of new features, including HDFS erasure coding, YARN Timeline Service v2, YARN federation, and much more, and discuss current release management status and community testing efforts dedicated to making Hadoop 3.0 the best Hadoop major release yet.
PASS Summit - SQL Server 2017 Deep DiveTravis Wright
Deep dive into SQL Server 2017 covering SQL Server on Linux, containers, HA improvements, SQL graph, machine learning, python, adaptive query processing, and much much more.
Introduction To Hadoop Administration - SpringPeopleSpringPeople
The Hadoop framework is used by major players including Google, Yahoo and IBM, largely for applications involving search engines and advertising. The popularity of Hadoop is juts increasing exponentially.
This presentation discusses the follow topics
What is Hadoop?
Need for Hadoop
History of Hadoop
Hadoop Overview
Advantages and Disadvantages of Hadoop
Hadoop Distributed File System
Comparing: RDBMS vs. Hadoop
Advantages and Disadvantages of HDFS
Hadoop frameworks
Modules of Hadoop frameworks
Features of 'Hadoop‘
Hadoop Analytics Tools
Securing your Big Data Environments in the CloudDataWorks Summit
Big Data tools are becoming a critical part of enterprise architectures and as such securing the data, at rest, and in motion is a necessity. More so, when you’re implementing these solutions in the cloud and the data doesn't reside within the confines of your trusted data center. Also, there is a fine balance between implementing enterprise-grade security and negotiating utmost performance given the overheads of encryption and/or identity management.
This session is designed to tackle these challenges head on and explain the various options available in the cloud. The focal points are the implementation of tools like Ranger and Knox for cloud deployments, but we also pay attention to the security features offered in the cloud that complement this process and secure the data in unprecedented ways.
Cloud Security + OSS Security tools are a deadly combination, when it comes to securing your Data Lake.
Evolution of Data Analytics: the past, the present and the futureVarun Nemmani
This paper delves into the topic of advanced analytics, the current industry demands to utilize and analyze huge/diverse amounts of data, how big data analytics is becoming a part of the decision making process and to anticipate trends. This paper takes the reader from Analytics era 1.0 to the current Analytics era 3.0; shows the future projections of big data analytics and also the current leaders of the Big Data Analytics market.
Dynamic & result oriented Professional with Overall 2.5+ years of professional experience in Software Development/Maintenance using Core Java/Servlet/JSP/PHP and MySQL.
Completed Hadoop training and looking a job change in big data technologies as Hadoop Developer.
Big Data is one of the most prominent disruptive technologies available today. The potential it offers for business is truly astounding.
But what is it? Time for a crashcourse!
• Capable of processing large sets of structured, semi-structured and unstructured data and supporting system architecture
• Implemented Proof of concepts on Hadoop stack and different big data analytic tools, migration from different databases to Hadoop.
• Developed multiple Map Reduce jobs in java for data cleaning and pre-processing according to the business requirements, Importing and exporting data into HDFS and Hive using Sqoop.
Having Experience in writing HIVE queries & Pig scripts.
1. Curriculum vitae
Amith Rayappa eMail: amitmce@gmail.com
DOB: 07-11-1988 Contact: +91(0) 9738282577
SUMMARY:
Overall 6 years of experience in IT and 2+ years in Hadoop as a Hadoop Administrator
Solid Knowledge of Hadoop framework and its Eco-System.
Expertise in Hadoop architecture, design and development of Big Data platform including
large clusters, Hadoop ecosystem projects
Expertise to install Hadoop and its related components in a multi-node cluster
environment.
Expertise in Implement new Hadoop hardware infrastructure
Expertise in HDFS Architecture and Cluster concepts.
Worked in 24x7 environment for production support in an on-call rotation.
Expertise in Cluster Installation for Lab, Staging and Production environment
Expertise in Hadoop job schedulers such as Fair scheduler and Capacity scheduler
Knowledge of Cluster coordination services using Zoo Keeper.
Expertise to install Hadoop and its related components in a multi-node cluster
environment.
Having working Knowledge and experience on Cloudera Hadoop (CDH5)
Experience on installing Ecosystem Components such as Hive 0. 13.0, HBase 0.98,sqoop
1.4.5, MySQL 5.5, Oozie 4.0
Should manage and clearly articulate in detail on managing HDP upgrades with rollback
options
Experience in configuring other Hadoop data distribution services like HBase, Hive, Impala
etc. on a need basis
Experience in configuring on demand Hadoop clusters
4 Yrs.of the Experience in Oracle SIEBEL 7.8, 8.1 and 8.2 on UNIX and Windows
implementation and troubleshooting
Excellent planning, analytical, written Communication and Interpersonal skills.
Proven ability to work efficiently in both independent and team environment.
Educational Profile:
Qualification University /Board Year Percentage
Master of Science BITS, Pilani 2011-2013 7.4(CGPA)
Bachelor of Engineering(E.C.) MCE, Hassan
(Vishveshwarya Uni.)
2006-2010 81.85%
PUC P.U. Board Karnataka 2004-2006 86.00%
SSC KSEEB 2003-2004 84.00%
2. Career Profile:
Big Data Apache Hadoop, Hive, Sqoop, Pig, HBase, Flume, Cloudera,
Operating Systems Windows, Linux, Centos,
Database MS-SQL, MySQL
Monitoring tools Cloudera Manager , Nagios , Ganglia
Working as Associate IT Consultant (Hadoop Administrator) with ITC Infotech India Ltd,
Bangalore from October2014 to till date.
PROJECT DETAILS:
Project #1
Client : Landmark
Role : Hadoop Administrator
Duration : Oct 2014– Till date
Project Description:
Landmark group is Dubai based group working in Retail business, It has 20 territories
with multiple concepts across each territory. The Production Cluster is of 200+ nodes
with 180 GB daily inflow data and around 380 TB of total data .Analyze hive, SQL queries
and refactor for better performance which includes backup of data to S3 , purging of
data in Hadoop cluster based on the retention period.
Roles & Responsibilities:
Installed and configured hadoop ecosystem tools Sqoop , Flume,
HBase,Zookeeper,oozie.
Configured various property files like core-site.xml, hdfs-site.xml,
mapred-site.xml based upon the job requirement.
Managing alerts from cluster monitoring tools like Ganglia and Nagios.
Managing and reviewing Hadoop log files.
Performance tuning of Hadoop cluster and Hadoop jobs.
Disk space management and monitoring.
Importing and exporting data into HDFS using Sqoop.
Performing data balancing on clusters.
Managing HDFS cluster users, permissions.
3. Involving in Analyzing system failures, identifying root causes, and recommended
course of actions. Documented the systems processes and procedures for future
references.
Taking backup of hadoop metadata using snapshots.
Working together with infrastructure,network and application teams to
guarantee high data quality and availability.
Install Hadoop patches and version upgrades when required.
Decommissioning and commissioning the Node on running hadoop cluster.
Work along with the Service Providers to resolve the tickets that were raised by
various business teams
data copy from one cluster to another or cluster using distcp utility
Install and configure different Hadoop ecosystem components such as HBase ,
Hive, Pig, Sqoop, Flume etc. as per requirement of specific PoCs and solutions.
Troubleshooting,diagnosing,solving and managing hdfs and mapReduce job
issues.
2. Worked as Siebel Administrator in IPSoft Global Services Bangalore from Sept 2013 to till
Oct 2014.
Technical Skills:
CRM Siebel8.1
Siebel Proficiency Siebel Server and Application Administration,User
Administration, Repository management.
Databases Oracle 11g,
Platforms AIX
Tools used Informatica 8.6, OBIEE 10.1.3.4, DAC 10g
4. PROJECT PROFILE:
Project #2
Client : MasterCard
Role : Siebel Administrator
Duration : Oct 2013– Oct 2014
ProjectDescription:
MasterCard isa NewYork basedmultinational financial servicescorporation,whichprocesses
paymentsbetweenthe banksof merchantsandthe card issuingbankswhouse the
"MasterCard"brand debitandcreditcards to make purchases. We Provide 24*7 monitoring,L1
& L2 supportprovidedforthe MasterCardSiebel environment.
Roles & Responsibilities:
■ Provide Dev, Stage, Prod and KSC environment support for Siebel Enterprise
Marketing/Call Center applications by making sure that the response & resolution times for
issue is within acceptable SLA’s.
■ Provide support on Siebel Releases activities on Dev/Stage/Prod/KSC environments
– Responsibilities include Repository Migrations, SRF Pushes, Incremental/Full
Compilations, DDL sync, Browser Scripts, file deployment as per implementation
document
■ Provide support on OBIEE, DAC, Informatica Releases activities on
Dev/Stage/Prod/KSC environments – Responsibilities include RPD Migration,
Connection pool updation, OBIEE/DAC/Informatica Service bounces, DAC
Repository Migration, Informatica Repository Migration, Informatica Workflow
migration, files deployment as per implementation document.
■ Responsible for coordination with multiple Infrastructure teams like critical
Upgrade/Patch activities with OS groups, DBA group, Onsite/Offshore Siebel
development teams to support production and release activities.
■ Configuration and Tunings of Siebel Servers and Components
■ We use IBM GSM change Management tool to track all changes implemented across all the
environments.
■ Coordinate with oracle effectively for issue resolution. Use support web effectively
and follow up on service requests for high priority issues.
5. ■ MasterCard use Serena PVCS version control tool to Manage Siebel release
activities across all the Siebel environments.
■ User provisioning request processed for Call Center agents on MAC Siebel Call
Center application as per telecom team request.
■ 24x7 monitoring - to monitor Siebel services,components, tasks etc
3. Worked as Siebel Administrator in Techmahindra Bangalore from Oct 2010 to till Sep
2013.
Technical Skills:
CRM Siebel 7.8, Siebel8.1
Siebel Proficiency Siebel Server and Application Administration,User
Administration, Repository management.
Databases Oracle 10g, MS SQL Server 2000
Platforms Sun Solaris , MS Windows XP 2003 Server.
PROJECT PROFILE:
Project #2
Client : British Telecom
Role : Siebel Administrator
Duration : Oct 2010– Sep 2013
ProjectDescription:
BritishTelecom isone of the largesttelecommunicationsservicescompaniesinthe worldand
has operationsinover170 countries.Toprovide 24x7ServerAdministrationandReleaseMgmt
supportto the existingOne Siebel application.
6. Roles & Responsibilities:
■ Environment support, Environment Maintenance Activities.
■ Coordinate with various stakeholders and ensure timely closure of open action items
(includes Oracle, Infrastructure vendors, Application Development and Maintenance teams)
■ Release Deployments - Deployment of release on UT,CST and other Testing environments
■ Configuration and Tunings of Siebel Servers and Components
■ Building new Environments , creation and assigning of components.
■ Creation and Monitoring of IBM Message Queues based on the requirement.
■ 24x7 monitoring - to monitor Siebel services,components, tasks etc
■ Using unix script developed automations to monitor the applications and automated the
process to reduce manual work.
Personal Details
Date of Birth : 07 Nov 1988
Marital Status : Unmarried
Nationality : Indian
Language Known : English ,Hindi, Kannada.
Current location : Bangalore, India
(Amith R)