The document provides a summary of Debtanu Chatterjee's skills and experience. It includes their contact information, objective to secure a position offering career growth, and lists over 20 technical skills including experience with Hadoop, Hive, Pig, HDFS, Sqoop, and SQL. It details two years of experience as a UI developer and Hadoop developer at Clinzen Pvt. Ltd. Their academic qualifications include an MCA and B.Sc. It also summarizes two projects, including developing a school management system using PHP, MySQL, and a Hadoop cluster installation project developing MapReduce jobs and solving Hive and Pig performance issues.
Companies around the world today find it increasingly difficult to organize and
manage large volumes of data. Hadoop has emerged as the most efficient data
platform for companies working with big data, and is an integral part of storing,
handling and retrieving enormous amounts of data in a variety of applications.
Hadoop helps to run deep analytics which cannot be effectively handled by a
database engine.
Big enterprises around the world have found Hadoop to be a game changer in their
Big Data management, and as more companies embrace this powerful technology
the demand for Hadoop Developers is also growing. By learning how to harness the
power of Hadoop 2.0 to manipulate, analyse and perform computations on Big
Data, you will be paving the way for an enriching and financially rewarding career as
an expert Hadoop developer.
MapReduce Best Practices and Lessons Learned Applied to Enterprise Datasets -...StampedeCon
At StampedeCon 2012 in St. Louis, Erich Hochmuth of Monsanto presents: Hadoop is quickly becoming the preferable platform for performing analysis over large datasets. We will explore opportunities for utilizing MapReduce to process genomic data in an enterprise system. We will discuss lessons learned introducing Hadoop into an existing enterprise and cover topics such as security, network architecture, and backups.
Big Data and Hadoop training course is designed to provide knowledge and skills to become a successful Hadoop Developer. In-depth knowledge of concepts such as Hadoop Distributed File System, Setting up the Hadoop Cluster, Map-Reduce,PIG, HIVE, HBase, Zookeeper, SQOOP etc. will be covered in the course.
Companies around the world today find it increasingly difficult to organize and
manage large volumes of data. Hadoop has emerged as the most efficient data
platform for companies working with big data, and is an integral part of storing,
handling and retrieving enormous amounts of data in a variety of applications.
Hadoop helps to run deep analytics which cannot be effectively handled by a
database engine.
Big enterprises around the world have found Hadoop to be a game changer in their
Big Data management, and as more companies embrace this powerful technology
the demand for Hadoop Developers is also growing. By learning how to harness the
power of Hadoop 2.0 to manipulate, analyse and perform computations on Big
Data, you will be paving the way for an enriching and financially rewarding career as
an expert Hadoop developer.
MapReduce Best Practices and Lessons Learned Applied to Enterprise Datasets -...StampedeCon
At StampedeCon 2012 in St. Louis, Erich Hochmuth of Monsanto presents: Hadoop is quickly becoming the preferable platform for performing analysis over large datasets. We will explore opportunities for utilizing MapReduce to process genomic data in an enterprise system. We will discuss lessons learned introducing Hadoop into an existing enterprise and cover topics such as security, network architecture, and backups.
Big Data and Hadoop training course is designed to provide knowledge and skills to become a successful Hadoop Developer. In-depth knowledge of concepts such as Hadoop Distributed File System, Setting up the Hadoop Cluster, Map-Reduce,PIG, HIVE, HBase, Zookeeper, SQOOP etc. will be covered in the course.
Big Data Analytics involves examining or processing large amounts of data (unstructured and structured) to create useful information which can help organizations to critically fine tune their business plans and increase profitability.
Apache HadoopTM is the most efficient data platform that simplifies and allows for
the distributed processing of large data sets. The latest revolution in big data
technology, Hadoop forms the core of an open source software framework, supporting the processing of large data across clustered systems. Using Hadoop, deep analytics that cannot be handled by a database engine can be run effectively.
Hadoop Online Training : kelly technologies is the bestHadoop online Training Institutes in Bangalore. ProvidingHadoop online Training by real time faculty in Bangalore.
What are Hadoop Components? Hadoop Ecosystem and Architecture | EdurekaEdureka!
YouTube Link: https://youtu.be/ll_O9JsjwT4
** Big Data Hadoop Certification Training - https://www.edureka.co/big-data-hadoop-training-certification **
This Edureka PPT on "Hadoop components" will provide you with detailed knowledge about the top Hadoop Components and it will help you understand the different categories of Hadoop Components. This PPT covers the following topics:
What is Hadoop?
Core Components of Hadoop
Hadoop Architecture
Hadoop EcoSystem
Hadoop Components in Data Storage
General Purpose Execution Engines
Hadoop Components in Database Management
Hadoop Components in Data Abstraction
Hadoop Components in Real-time Data Streaming
Hadoop Components in Graph Processing
Hadoop Components in Machine Learning
Hadoop Cluster Management tools
Follow us to never miss an update in the future.
YouTube: https://www.youtube.com/user/edurekaIN
Instagram: https://www.instagram.com/edureka_learning/
Facebook: https://www.facebook.com/edurekaIN/
Twitter: https://twitter.com/edurekain
LinkedIn: https://www.linkedin.com/company/edureka
Castbox: https://castbox.fm/networks/505?country=in
Big-Data Hadoop Tutorials - MindScripts Technologies, Pune amrutupre
MindScripts Technologies, is the leading Big-Data Hadoop Training institutes in Pune, providing a complete Big-Data Hadoop Course with Cloud-Era certification.
Hadoop Online Training : kelly technologies is the bestHadoop online Training Institutes in Bangalore. ProvidingHadoop online Training by real time faculty in Bangalore.
Forrester predicts, CIOs who are late to the Hadoop game will finally make the platform a priority in 2015. Hadoop has evolved as a must-to-know technology and has been a reason for better career, salary and job opportunities for many professionals.
#Bigdata #hadoop LIVE FREE DEMO on 16th JUNE 2017 at 07:30AM.
Interested candidates can register here: https://goo.gl/za6kI5.
We have an outstanding real time trainers to provide excellent growth in career....
Acute Soft Solutions India Pvt.Ltd. is a global leader in providing online training services which are a part of our wide area of services.
Big Data Analytics involves examining or processing large amounts of data (unstructured and structured) to create useful information which can help organizations to critically fine tune their business plans and increase profitability.
Apache HadoopTM is the most efficient data platform that simplifies and allows for
the distributed processing of large data sets. The latest revolution in big data
technology, Hadoop forms the core of an open source software framework, supporting the processing of large data across clustered systems. Using Hadoop, deep analytics that cannot be handled by a database engine can be run effectively.
Hadoop Online Training : kelly technologies is the bestHadoop online Training Institutes in Bangalore. ProvidingHadoop online Training by real time faculty in Bangalore.
What are Hadoop Components? Hadoop Ecosystem and Architecture | EdurekaEdureka!
YouTube Link: https://youtu.be/ll_O9JsjwT4
** Big Data Hadoop Certification Training - https://www.edureka.co/big-data-hadoop-training-certification **
This Edureka PPT on "Hadoop components" will provide you with detailed knowledge about the top Hadoop Components and it will help you understand the different categories of Hadoop Components. This PPT covers the following topics:
What is Hadoop?
Core Components of Hadoop
Hadoop Architecture
Hadoop EcoSystem
Hadoop Components in Data Storage
General Purpose Execution Engines
Hadoop Components in Database Management
Hadoop Components in Data Abstraction
Hadoop Components in Real-time Data Streaming
Hadoop Components in Graph Processing
Hadoop Components in Machine Learning
Hadoop Cluster Management tools
Follow us to never miss an update in the future.
YouTube: https://www.youtube.com/user/edurekaIN
Instagram: https://www.instagram.com/edureka_learning/
Facebook: https://www.facebook.com/edurekaIN/
Twitter: https://twitter.com/edurekain
LinkedIn: https://www.linkedin.com/company/edureka
Castbox: https://castbox.fm/networks/505?country=in
Big-Data Hadoop Tutorials - MindScripts Technologies, Pune amrutupre
MindScripts Technologies, is the leading Big-Data Hadoop Training institutes in Pune, providing a complete Big-Data Hadoop Course with Cloud-Era certification.
Hadoop Online Training : kelly technologies is the bestHadoop online Training Institutes in Bangalore. ProvidingHadoop online Training by real time faculty in Bangalore.
Forrester predicts, CIOs who are late to the Hadoop game will finally make the platform a priority in 2015. Hadoop has evolved as a must-to-know technology and has been a reason for better career, salary and job opportunities for many professionals.
#Bigdata #hadoop LIVE FREE DEMO on 16th JUNE 2017 at 07:30AM.
Interested candidates can register here: https://goo.gl/za6kI5.
We have an outstanding real time trainers to provide excellent growth in career....
Acute Soft Solutions India Pvt.Ltd. is a global leader in providing online training services which are a part of our wide area of services.
This presentations gives high level overview of : networking, security, storage, end point devices (PC), Servers, Enterprise System Management and Development and SOA\integration domains. It includes staffing ratios, vendor and integrator positioning in Israel.
Ci sono prodotti ed attrezzature in commercio che ci aiutano a tenere lontane o combattere le zanzare.
Ma qualcosa possiamo fare
anche noi nel nostro piccolo
per ridurre i focolai di diffusione
e prevenire le fastidiose punture
nanopub-java: A Java Library for NanopublicationsTobias Kuhn
The concept of nanopublications was first proposed about six years ago, but it lacked openly available implementations. The library presented here is the first one that has become an official implementation of the nanopublication community. Its core features are stable, but it also contains unofficial and experimental extensions: for publishing to a decentralized server network, for defining sets of nanopublications with indexes, for informal assertions, and for digitally signing nanopublications. Most of the features of the library can also be accessed via an online validator interface.
• Capable of processing large sets of structured, semi-structured and unstructured data and supporting system architecture
• Implemented Proof of concepts on Hadoop stack and different big data analytic tools, migration from different databases to Hadoop.
• Developed multiple Map Reduce jobs in java for data cleaning and pre-processing according to the business requirements, Importing and exporting data into HDFS and Hive using Sqoop.
Having Experience in writing HIVE queries & Pig scripts.
IGEEKS is a Bangalore based Training & Recruitment company.
We offers osoftware training courses starting from absolute beginner level to advanced levels.Providing high quality training at affordable fees is our core value. All our trainers are working IT professionals with rich experience. We work with our students in developing the right skills they need to build their career in present competitive environment. We have flexible batch times to suit the timings of graduating students and working professionals.
1. Debtanu Chatterjee Sri Kalki Chamber
debtanu.good@gmail.com Block – B, Flat no. : 105
phone no. :7501111325 Allwyn X Road, Madinaguda
Hyderabad , 500050.
OBJECTIVE
To secure a position in an organization that offers career growth and a chance to achieve goals
through persistence and hard work where and also to give my best in whatever I do for
future development of organization.
Skills Elaboration
1. 2 years of Work Experience in technologies like HTML, CSS, JS, PHP, Mysql.
2. 2 years of Work Experience in technologies like Hadoop Eco System, HDFS, map reduce,
hive, pig, sqoop, flume, oozie, HBASE.
3. Attended Training on Big Data Technologies like Hadoop, PIG, HIVE, and Have the hands
on, on public data sets.
4. Experience in using XML, JavaScript, JSON, Ajax, CSS, HTML, and PHP.
5. Good knowledge of Hadoop ecosystem, HDFS, java , RDBMS(oracle 11g.
6. Experienced on working with Big Data and Hadoop File System (HDFS).
7. Hands on Experience in working with ecosystems like Hive, Pig, Sqoop, Map Reduce, Flume,
OoZie.
8. good Knowledge of Hadoop and Hive and Hive's analytical functions.
9. Very good understanding of Partitions, Bucketing concepts in Hive and designed both
Managed and External tables in Hive to optimize performance
10. Capturing data from existing databases that provide SQL interfaces using Sqoop Import.
11. Efficient in building hive, pig and map Reduce scripts.
12. Experienced in managing and reviewing Hadoop log files
13. Implemented Proofs of Concept on Hadoop stack and different big data analytic tools,
migration from different databases (i.e. Oracle,MYSQL ) to Hadoop.
14. Successfully loaded files to Hive and HDFS from MongoDB, HBase
15. Loaded the data set into Hive for ETL Operation.
16. Good knowledge on Hadoop Cluster architecture and monitoring the cluster.
17. Experience in using Zoo keeper and cloudera Manager.
18. Hands on experience in IDE tools like Eclipse.
2. 19. Experience in using Sequence files, RCFile, AVRO file formats.
20. Developed Oozie work flow for scheduling and orchestrating the ETL process
21. tools used for cluster management like Cloudera Manager and Apache Ambari
Professional Experience
Worked in Clinzen Pvt.Ltd as a UI developer and SQL Developer .
Currently Working in Clinzen Pvt.Ltd, Hyderabad as a Hadoop developer till Date.
Academic Profile
MCA (Master of Computer Applications) From WBUT with 78.50%.
B.Sc. (Mathematics, Physics & Chemistry) from Calcutta UNIVERSITY with 50.01%.
Project Experience
Project # 1 School Management System
Responsibilities:
DB design, coding, development, implementation.
Skills Used:
PHP, MySQL, JavaScript, HTML, CSS,AJAX.
Team size : 3 members
Description:
storing students data and staff data
generating Mark sheet and yearly report card for each students.
Generating ID card and unique id for students and staffs.
Full library management of the school.
Keeping records Staff' salary records and leave recommendation
Generating service book and pension book for each staff.
Project # 2
3. Install raw Hadoop and NoSQL applications on cluster mode (3 node) and
develop programs for sorting and analyzing data.
Responsibilities:
Replaced default Derby meta-data storage system for Hive with MySQL system.
Executed queries using Hive and developed Map-Reduce jobs to analyze data.
Solved performance issues in Hive and Pig scripts with understanding of Joins, Group and
aggregation and how does it translate to MapReduce jobs.
Developed Pig Latin scripts to extract the data from the web server output files to load into
HDFS.
Developed the Pig UDF's to preprocess the data for analysis.
• DevelopedHive queries for the analysts.
• Utilized Apache Hadoop environment.
• Involved in loading data from LINUX and UNIX file system to HDFS.
• Supported in setting up QA environment and updating configurations for implementing scripts with
Pig.
Environment: Core Java, Apache Hadoop, HDFS, Pig, Hive, Shell Scripting, My Sql, Linux.
Areas of Expertise:
Big Data Ecosystems: Hadoop, MapReduce, HDFS, HBase, Zookeeper, Hive, Pig, Sqoop,
Oozie,
Programming Languages: Java, C,php.
Scripting Languages: JavaScript, XML, HTML,pig Latin.
Databases: NoSQL (HBASE), Oracle(11g).
Server: Apache.
Tools: Eclipse.
Platforms: Windows, Linux.
Methodologies: UML.