Sam has over 8 years of experience in big data technologies including Hadoop, Spark, Hive, Pig, Kafka and Cassandra. He has expertise in data ingestion, processing, ETL and analytics. Some of his responsibilities include developing MapReduce programs, Spark applications, data pipelines, streaming jobs and visualizations. He is proficient in Java, Python, Scala and various big data tools.
This document provides a detailed summary of Poorna Chandra Rao Kommana's professional experience and technical skills. It outlines his 8 years of experience in big data technologies including Hadoop, Hive, Pig, Spark, Kafka and AWS services. It details his roles and responsibilities in building scalable big data solutions, developing ETL pipelines, performing data analysis, and optimizing performance. His skills include Java, Python, SQL, Pig Latin, HiveQL, and tools like Eclipse and PyCharm.
Sanath Pabba has over 5 years of experience working with big data technologies like Hadoop, Spark, Hive, Pig, Kafka and NoSQL databases. He has expertise in data extraction, transformation and loading processes. Some of his responsibilities include writing Sqoop and Spark jobs to load and prepare data, developing automation scripts to monitor cluster utilization, and implementing validation rules for data quality. He has worked on various projects involving data warehousing, reporting, stream processing and analytics using technologies like SQL Server, Hive and Spark.
Hadoop World 2011: Hadoop and RDBMS with Sqoop and Other Tools - Guy Harrison...Cloudera, Inc.
The document discusses integrating Hadoop with relational databases. It describes scenarios where reference data is stored in an RDBMS and used in Hadoop, Hadoop is used for offline analytics on data stored in an RDBMS, and exporting MapReduce outputs to an RDBMS. It then presents a case study on extending SQOOP for optimized Oracle integration and compares performance with and without the extension. Other tools for Hadoop-RDBMS integration are also briefly outlined.
This document provides an overview of big data architecture, the Hadoop ecosystem, and NoSQL databases. It discusses common big data use cases, characteristics, and tools. It describes the typical 3-tier traditional architecture compared to the big data architecture using Hadoop. Key components of Hadoop like HDFS, MapReduce, Hive, Pig, Avro/Thrift, HBase are explained. The document also discusses stream processing tools like Storm, Spark and real-time query with Impala. It notes how NoSQL databases can integrate with Hadoop/MapReduce for both batch and real-time processing.
The document provides a summary of a senior big data consultant with over 4 years of experience working with technologies such as Apache Spark, Hadoop, Hive, Pig, Kafka and databases including HBase, Cassandra. The consultant has strong skills in building real-time streaming solutions, data pipelines, and implementing Hadoop-based data warehouses. Areas of expertise include Spark, Scala, Java, machine learning, and cloud platforms like AWS.
• Capable of processing large sets of structured, semi-structured and unstructured data and supporting system architecture
• Implemented Proof of concepts on Hadoop stack and different big data analytic tools, migration from different databases to Hadoop.
• Developed multiple Map Reduce jobs in java for data cleaning and pre-processing according to the business requirements, Importing and exporting data into HDFS and Hive using Sqoop.
Having Experience in writing HIVE queries & Pig scripts.
M.V. Rama Kumar has 3 years of experience in application development using Java and big data technologies like Hadoop. He has 1.6 years of experience using Hadoop components such as HDFS, MapReduce, Pig, Hive, Sqoop, HBase and Oozie. He has extensive experience setting up Hadoop clusters and processing large, structured and unstructured data.
Apache Hadoop is a framework for distributed storage and processing of large datasets across clusters of commodity hardware. It provides HDFS for distributed file storage and MapReduce as a programming model for distributed computations. Hadoop includes other technologies like YARN for resource management, Spark for fast computation, HBase for NoSQL database, and tools for data analysis, transfer, and security. Hadoop can run on-premise or in cloud environments and supports analytics workloads.
This document provides a detailed summary of Poorna Chandra Rao Kommana's professional experience and technical skills. It outlines his 8 years of experience in big data technologies including Hadoop, Hive, Pig, Spark, Kafka and AWS services. It details his roles and responsibilities in building scalable big data solutions, developing ETL pipelines, performing data analysis, and optimizing performance. His skills include Java, Python, SQL, Pig Latin, HiveQL, and tools like Eclipse and PyCharm.
Sanath Pabba has over 5 years of experience working with big data technologies like Hadoop, Spark, Hive, Pig, Kafka and NoSQL databases. He has expertise in data extraction, transformation and loading processes. Some of his responsibilities include writing Sqoop and Spark jobs to load and prepare data, developing automation scripts to monitor cluster utilization, and implementing validation rules for data quality. He has worked on various projects involving data warehousing, reporting, stream processing and analytics using technologies like SQL Server, Hive and Spark.
Hadoop World 2011: Hadoop and RDBMS with Sqoop and Other Tools - Guy Harrison...Cloudera, Inc.
The document discusses integrating Hadoop with relational databases. It describes scenarios where reference data is stored in an RDBMS and used in Hadoop, Hadoop is used for offline analytics on data stored in an RDBMS, and exporting MapReduce outputs to an RDBMS. It then presents a case study on extending SQOOP for optimized Oracle integration and compares performance with and without the extension. Other tools for Hadoop-RDBMS integration are also briefly outlined.
This document provides an overview of big data architecture, the Hadoop ecosystem, and NoSQL databases. It discusses common big data use cases, characteristics, and tools. It describes the typical 3-tier traditional architecture compared to the big data architecture using Hadoop. Key components of Hadoop like HDFS, MapReduce, Hive, Pig, Avro/Thrift, HBase are explained. The document also discusses stream processing tools like Storm, Spark and real-time query with Impala. It notes how NoSQL databases can integrate with Hadoop/MapReduce for both batch and real-time processing.
The document provides a summary of a senior big data consultant with over 4 years of experience working with technologies such as Apache Spark, Hadoop, Hive, Pig, Kafka and databases including HBase, Cassandra. The consultant has strong skills in building real-time streaming solutions, data pipelines, and implementing Hadoop-based data warehouses. Areas of expertise include Spark, Scala, Java, machine learning, and cloud platforms like AWS.
• Capable of processing large sets of structured, semi-structured and unstructured data and supporting system architecture
• Implemented Proof of concepts on Hadoop stack and different big data analytic tools, migration from different databases to Hadoop.
• Developed multiple Map Reduce jobs in java for data cleaning and pre-processing according to the business requirements, Importing and exporting data into HDFS and Hive using Sqoop.
Having Experience in writing HIVE queries & Pig scripts.
M.V. Rama Kumar has 3 years of experience in application development using Java and big data technologies like Hadoop. He has 1.6 years of experience using Hadoop components such as HDFS, MapReduce, Pig, Hive, Sqoop, HBase and Oozie. He has extensive experience setting up Hadoop clusters and processing large, structured and unstructured data.
Apache Hadoop is a framework for distributed storage and processing of large datasets across clusters of commodity hardware. It provides HDFS for distributed file storage and MapReduce as a programming model for distributed computations. Hadoop includes other technologies like YARN for resource management, Spark for fast computation, HBase for NoSQL database, and tools for data analysis, transfer, and security. Hadoop can run on-premise or in cloud environments and supports analytics workloads.
The document provides an overview of Apache Hadoop and related big data technologies. It discusses Hadoop components like HDFS for storage, MapReduce for processing, and HBase for columnar storage. It also covers related projects like Hive for SQL queries, ZooKeeper for coordination, and Hortonworks and Cloudera distributions.
The document provides a professional summary and details for Madhusudhn Reddy.Gujja including 3 years of experience in big data tools like Hadoop, Hive, Pig and Spark. He has extensive experience developing Pig Latin scripts, writing MapReduce programs in Java, and loading/transforming large datasets. He is proficient in technologies such as HDFS, HBase, Kafka, Flume, Impala and has worked on projects involving data analytics, ETL processes and clustering Hadoop.
This document contains a resume for Ramez Rangrez, a Hadoop Administrator. It includes his contact information, skills, work experience configuring and deploying Hadoop clusters for CloudAge in Pune, India, education including a BE in Electronics and Telecommunication, and personal details. His responsibilities at CloudAge involved configuring, deploying, maintaining, and troubleshooting Hadoop clusters on AWS and performing tasks like capacity planning, performance tuning, and documentation. He contributed to a customer churn analysis project using a cloud-based system combining data mining, social network analysis, and statistics analysis on telecom data.
The document provides an overview of big data technologies including Hadoop, MapReduce, HDFS, Hive, Pig, Sqoop, HBase, MongoDB, and Cassandra. It discusses how these technologies enable processing and analyzing very large datasets across commodity hardware. It also outlines the growth and market potential of the big data sector, which is expected to reach $48 billion by 2018.
Prashanth Shankar Kumar has over 8 years of experience in data analytics, Hadoop, Teradata, and mainframes. He currently works as a Hadoop Developer/Tech Lead at Bank of America where he develops Hive queries, Impala queries, MapReduce programs, and Oozie workflows. Previously he worked as a Hadoop Developer at State Farm Insurance where he installed and managed Hadoop clusters and developed solutions using Hive, Pig, Sqoop, and HBase. He has expertise in Teradata, SQL, Java, Linux, and agile methodologies.
The document discusses and compares MapReduce and relational database management systems (RDBMS) for large-scale data processing. It describes several hybrid approaches that attempt to combine the scalability of MapReduce with the query optimization and efficiency of parallel RDBMS. HadoopDB is highlighted as a system that uses Hadoop for communication and data distribution across nodes running PostgreSQL for query execution. Performance evaluations show hybrid systems can outperform pure MapReduce but may still lag specialized parallel databases.
Hadoop is a framework that allows businesses to analyze vast amounts of data quickly and at low cost by distributing processing across commodity servers. It consists of two main components: HDFS for data storage and MapReduce for processing. Learning Hadoop requires familiarity with Java, Linux, and object-oriented programming principles. The document recommends getting hands-on experience by installing a Cloudera Distribution of Hadoop virtual machine or package to become comfortable with the framework.
Mopuru Babu has over 9 years of experience in software development using Java technologies and 3 years experience in Hadoop development. He has extensive experience designing, developing, and deploying multi-tier and enterprise-level distributed applications. He has expertise in technologies like Hadoop, Hive, Pig, Spark, and frameworks like Spring and Struts. He has worked on both small and large projects for clients in various industries.
Introduction to Hadoop, HBase, and NoSQLNick Dimiduk
The document is a presentation on NoSQL databases given by Nick Dimiduk. It begins with an introduction of the speaker and their background. The presentation then covers what NoSQL is not, the motivations for NoSQL databases, an overview of Hadoop and its components, and a description of HBase as a structured, distributed database built on Hadoop.
This document contains a resume for Ramez Rangrez, who is a Hadoop Administrator with over 5 years of experience in configuring, deploying, maintaining, and monitoring Hadoop clusters. He has worked at Persistent Systems Limited since 2016 and CloudAge from 2013 to 2016. His skills include SQL, Hadoop ecosystem tools, Linux, AWS, and he is working towards Cloudera certification. His responsibilities have involved Hadoop administration, troubleshooting issues, capacity planning, and performance tuning. He also has experience with a customer churn analysis project using Hadoop and cloud systems.
This document provides an overview and comparison of RDBMS, Hadoop, and Spark. It introduces RDBMS and describes its use cases such as online transaction processing and data warehouses. It then introduces Hadoop and describes its ecosystem including HDFS, YARN, MapReduce, and related sub-modules. Common use cases for Hadoop are also outlined. Spark is then introduced along with its modules like Spark Core, SQL, and MLlib. Use cases for Spark include data enrichment, trigger event detection, and machine learning. The document concludes by comparing RDBMS and Hadoop, as well as Hadoop and Spark, and addressing common misconceptions about Hadoop and Spark.
Fundamentals of Big Data, Hadoop project design and case study or Use case
General planning consideration and most necessaries in Hadoop ecosystem and Hadoop projects
This will provide the basis for choosing the right Hadoop implementation, Hadoop technologies integration, adoption and creating an infrastructure.
Building applications using Apache Hadoop with a use-case of WI-FI log analysis has real life example.
TWDI Accelerate Seattle, Oct 16, 2017: Distributed and In-Database Analytics ...Debraj GuhaThakurta
Event: TDWI Accelerate Seattle, October 16, 2017
Topic: Distributed and In-Database Analytics with R
Presenter: Debraj GuhaThakurta
Description: How to develop scalable and in-DB analytics using R in Spark and SQL-Server
Overview of Big data, Hadoop and Microsoft BI - version1Thanh Nguyen
Big Data and advanced analytics are critical topics for executives today. But many still aren't sure how to turn that promise into value. This presentation provides an overview of 16 examples and use cases that lay out the different ways companies have approached the issue and found value: everything from pricing flexibility to customer preference management to credit risk analysis to fraud protection and discount targeting. For the latest on Big Data & Advanced Analytics: http://mckinseyonmarketingandsales.com/topics/big-data
Quick Brief about " What is Hadoop"
I didn't explain in detail about hadoop, but reading this slides will give you insight of Hadoop and core product usage. This document will be more useful for PM, Newbies, Technical Architect entering into Cloud Computing.
This document provides an overview of an advanced Big Data hands-on course covering Hadoop, Sqoop, Pig, Hive and enterprise applications. It introduces key concepts like Hadoop and large data processing, demonstrates tools like Sqoop, Pig and Hive for data integration, querying and analysis on Hadoop. It also discusses challenges for enterprises adopting Hadoop technologies and bridging the skills gap.
This document summarizes Andrew Brust's presentation on using the Microsoft platform for big data. It discusses Hadoop and HDInsight, MapReduce, using Hive with ODBC and the BI stack. It also covers Hekaton, NoSQL, SQL Server Parallel Data Warehouse, and PolyBase. The presentation includes demos of HDInsight, MapReduce, and using Hive with the BI stack.
http://bit.ly/1BTaXZP – Hadoop has been a huge success in the data world. It’s disrupted decades of data management practices and technologies by introducing a massively parallel processing framework. The community and the development of all the Open Source components pushed Hadoop to where it is now.
That's why the Hadoop community is excited about Apache Spark. The Spark software stack includes a core data-processing engine, an interface for interactive querying, Sparkstreaming for streaming data analysis, and growing libraries for machine-learning and graph analysis. Spark is quickly establishing itself as a leading environment for doing fast, iterative in-memory and streaming analysis.
This talk will give an introduction the Spark stack, explain how Spark has lighting fast results, and how it complements Apache Hadoop.
Keys Botzum - Senior Principal Technologist with MapR Technologies
Keys is Senior Principal Technologist with MapR Technologies, where he wears many hats. His primary responsibility is interacting with customers in the field, but he also teaches classes, contributes to documentation, and works with engineering teams. He has over 15 years of experience in large scale distributed system design. Previously, he was a Senior Technical Staff Member with IBM, and a respected author of many articles on the WebSphere Application Server as well as a book.
Brief Introduction about Hadoop and Core Services.Muthu Natarajan
I have given quick introduction about Hadoop, Big Data, Business Intelligence and other core services and program involved to use Hadoop as a successful tool for Big Data analysis.
My true understanding in Big-Data:
“Data” become “information” but now big data bring information to “Knowledge” and ‘knowledge” becomes “Wisdom” and “Wisdom” turn into “Business” or “Revenue”, All if you use promptly & timely manner
The report discusses the key components and objectives of HDFS, including data replication for fault tolerance, HDFS architecture with a NameNode and DataNodes, and HDFS properties like large data sets, write once read many model, and commodity hardware. It provides an overview of HDFS and its design to reliably store and retrieve large volumes of distributed data.
Rajesh R is seeking a position as an efficient developer to contribute to organizational progress while enhancing his technical skills. He has over 20 years of experience in roles such as service engineer, customer support engineer, desktop support engineer, hardware and networking engineer, and sales manager. He has skills in hardware installation, operating system installation, networking, and security systems. He holds an I.T.I in Electronics, a Diploma in Hardware & Networking, and a B.C.A.
The document is a resume for a technical writer with over 10 years of experience writing technical documentation for software and hardware products. It summarizes two roles, the current one as a lead engineer at HCL Technologies since 2011 where responsibilities include analyzing requirements, writing documentation, and reviewing documents, and a prior role as a technical writer from 2009-2010 where duties involved documenting engineering patents. It also lists education, training, software skills, and personal details.
The document provides an overview of Apache Hadoop and related big data technologies. It discusses Hadoop components like HDFS for storage, MapReduce for processing, and HBase for columnar storage. It also covers related projects like Hive for SQL queries, ZooKeeper for coordination, and Hortonworks and Cloudera distributions.
The document provides a professional summary and details for Madhusudhn Reddy.Gujja including 3 years of experience in big data tools like Hadoop, Hive, Pig and Spark. He has extensive experience developing Pig Latin scripts, writing MapReduce programs in Java, and loading/transforming large datasets. He is proficient in technologies such as HDFS, HBase, Kafka, Flume, Impala and has worked on projects involving data analytics, ETL processes and clustering Hadoop.
This document contains a resume for Ramez Rangrez, a Hadoop Administrator. It includes his contact information, skills, work experience configuring and deploying Hadoop clusters for CloudAge in Pune, India, education including a BE in Electronics and Telecommunication, and personal details. His responsibilities at CloudAge involved configuring, deploying, maintaining, and troubleshooting Hadoop clusters on AWS and performing tasks like capacity planning, performance tuning, and documentation. He contributed to a customer churn analysis project using a cloud-based system combining data mining, social network analysis, and statistics analysis on telecom data.
The document provides an overview of big data technologies including Hadoop, MapReduce, HDFS, Hive, Pig, Sqoop, HBase, MongoDB, and Cassandra. It discusses how these technologies enable processing and analyzing very large datasets across commodity hardware. It also outlines the growth and market potential of the big data sector, which is expected to reach $48 billion by 2018.
Prashanth Shankar Kumar has over 8 years of experience in data analytics, Hadoop, Teradata, and mainframes. He currently works as a Hadoop Developer/Tech Lead at Bank of America where he develops Hive queries, Impala queries, MapReduce programs, and Oozie workflows. Previously he worked as a Hadoop Developer at State Farm Insurance where he installed and managed Hadoop clusters and developed solutions using Hive, Pig, Sqoop, and HBase. He has expertise in Teradata, SQL, Java, Linux, and agile methodologies.
The document discusses and compares MapReduce and relational database management systems (RDBMS) for large-scale data processing. It describes several hybrid approaches that attempt to combine the scalability of MapReduce with the query optimization and efficiency of parallel RDBMS. HadoopDB is highlighted as a system that uses Hadoop for communication and data distribution across nodes running PostgreSQL for query execution. Performance evaluations show hybrid systems can outperform pure MapReduce but may still lag specialized parallel databases.
Hadoop is a framework that allows businesses to analyze vast amounts of data quickly and at low cost by distributing processing across commodity servers. It consists of two main components: HDFS for data storage and MapReduce for processing. Learning Hadoop requires familiarity with Java, Linux, and object-oriented programming principles. The document recommends getting hands-on experience by installing a Cloudera Distribution of Hadoop virtual machine or package to become comfortable with the framework.
Mopuru Babu has over 9 years of experience in software development using Java technologies and 3 years experience in Hadoop development. He has extensive experience designing, developing, and deploying multi-tier and enterprise-level distributed applications. He has expertise in technologies like Hadoop, Hive, Pig, Spark, and frameworks like Spring and Struts. He has worked on both small and large projects for clients in various industries.
Introduction to Hadoop, HBase, and NoSQLNick Dimiduk
The document is a presentation on NoSQL databases given by Nick Dimiduk. It begins with an introduction of the speaker and their background. The presentation then covers what NoSQL is not, the motivations for NoSQL databases, an overview of Hadoop and its components, and a description of HBase as a structured, distributed database built on Hadoop.
This document contains a resume for Ramez Rangrez, who is a Hadoop Administrator with over 5 years of experience in configuring, deploying, maintaining, and monitoring Hadoop clusters. He has worked at Persistent Systems Limited since 2016 and CloudAge from 2013 to 2016. His skills include SQL, Hadoop ecosystem tools, Linux, AWS, and he is working towards Cloudera certification. His responsibilities have involved Hadoop administration, troubleshooting issues, capacity planning, and performance tuning. He also has experience with a customer churn analysis project using Hadoop and cloud systems.
This document provides an overview and comparison of RDBMS, Hadoop, and Spark. It introduces RDBMS and describes its use cases such as online transaction processing and data warehouses. It then introduces Hadoop and describes its ecosystem including HDFS, YARN, MapReduce, and related sub-modules. Common use cases for Hadoop are also outlined. Spark is then introduced along with its modules like Spark Core, SQL, and MLlib. Use cases for Spark include data enrichment, trigger event detection, and machine learning. The document concludes by comparing RDBMS and Hadoop, as well as Hadoop and Spark, and addressing common misconceptions about Hadoop and Spark.
Fundamentals of Big Data, Hadoop project design and case study or Use case
General planning consideration and most necessaries in Hadoop ecosystem and Hadoop projects
This will provide the basis for choosing the right Hadoop implementation, Hadoop technologies integration, adoption and creating an infrastructure.
Building applications using Apache Hadoop with a use-case of WI-FI log analysis has real life example.
TWDI Accelerate Seattle, Oct 16, 2017: Distributed and In-Database Analytics ...Debraj GuhaThakurta
Event: TDWI Accelerate Seattle, October 16, 2017
Topic: Distributed and In-Database Analytics with R
Presenter: Debraj GuhaThakurta
Description: How to develop scalable and in-DB analytics using R in Spark and SQL-Server
Overview of Big data, Hadoop and Microsoft BI - version1Thanh Nguyen
Big Data and advanced analytics are critical topics for executives today. But many still aren't sure how to turn that promise into value. This presentation provides an overview of 16 examples and use cases that lay out the different ways companies have approached the issue and found value: everything from pricing flexibility to customer preference management to credit risk analysis to fraud protection and discount targeting. For the latest on Big Data & Advanced Analytics: http://mckinseyonmarketingandsales.com/topics/big-data
Quick Brief about " What is Hadoop"
I didn't explain in detail about hadoop, but reading this slides will give you insight of Hadoop and core product usage. This document will be more useful for PM, Newbies, Technical Architect entering into Cloud Computing.
This document provides an overview of an advanced Big Data hands-on course covering Hadoop, Sqoop, Pig, Hive and enterprise applications. It introduces key concepts like Hadoop and large data processing, demonstrates tools like Sqoop, Pig and Hive for data integration, querying and analysis on Hadoop. It also discusses challenges for enterprises adopting Hadoop technologies and bridging the skills gap.
This document summarizes Andrew Brust's presentation on using the Microsoft platform for big data. It discusses Hadoop and HDInsight, MapReduce, using Hive with ODBC and the BI stack. It also covers Hekaton, NoSQL, SQL Server Parallel Data Warehouse, and PolyBase. The presentation includes demos of HDInsight, MapReduce, and using Hive with the BI stack.
http://bit.ly/1BTaXZP – Hadoop has been a huge success in the data world. It’s disrupted decades of data management practices and technologies by introducing a massively parallel processing framework. The community and the development of all the Open Source components pushed Hadoop to where it is now.
That's why the Hadoop community is excited about Apache Spark. The Spark software stack includes a core data-processing engine, an interface for interactive querying, Sparkstreaming for streaming data analysis, and growing libraries for machine-learning and graph analysis. Spark is quickly establishing itself as a leading environment for doing fast, iterative in-memory and streaming analysis.
This talk will give an introduction the Spark stack, explain how Spark has lighting fast results, and how it complements Apache Hadoop.
Keys Botzum - Senior Principal Technologist with MapR Technologies
Keys is Senior Principal Technologist with MapR Technologies, where he wears many hats. His primary responsibility is interacting with customers in the field, but he also teaches classes, contributes to documentation, and works with engineering teams. He has over 15 years of experience in large scale distributed system design. Previously, he was a Senior Technical Staff Member with IBM, and a respected author of many articles on the WebSphere Application Server as well as a book.
Brief Introduction about Hadoop and Core Services.Muthu Natarajan
I have given quick introduction about Hadoop, Big Data, Business Intelligence and other core services and program involved to use Hadoop as a successful tool for Big Data analysis.
My true understanding in Big-Data:
“Data” become “information” but now big data bring information to “Knowledge” and ‘knowledge” becomes “Wisdom” and “Wisdom” turn into “Business” or “Revenue”, All if you use promptly & timely manner
The report discusses the key components and objectives of HDFS, including data replication for fault tolerance, HDFS architecture with a NameNode and DataNodes, and HDFS properties like large data sets, write once read many model, and commodity hardware. It provides an overview of HDFS and its design to reliably store and retrieve large volumes of distributed data.
Rajesh R is seeking a position as an efficient developer to contribute to organizational progress while enhancing his technical skills. He has over 20 years of experience in roles such as service engineer, customer support engineer, desktop support engineer, hardware and networking engineer, and sales manager. He has skills in hardware installation, operating system installation, networking, and security systems. He holds an I.T.I in Electronics, a Diploma in Hardware & Networking, and a B.C.A.
The document is a resume for a technical writer with over 10 years of experience writing technical documentation for software and hardware products. It summarizes two roles, the current one as a lead engineer at HCL Technologies since 2011 where responsibilities include analyzing requirements, writing documentation, and reviewing documents, and a prior role as a technical writer from 2009-2010 where duties involved documenting engineering patents. It also lists education, training, software skills, and personal details.
Leslie Diaz is a Service Desk Manager with over 10 years of experience in customer service and IT support. She is proficient in ServiceDesk, MS Office, Microsoft Exchange, Lync, Windows Server 2003, and networked applications. As Service Desk Manager since 2012, she has raised service standards, empowered her team, and assisted with two acquisitions. Previously she was an IT Project Manager and Senior Analyst at Dell where she troubleshot issues and identified ways to improve the customer experience.
Santhosh Kumar S.J. is seeking a progressive and responsible position where he can utilize his MBA and experience in HR recruiting. He has over 4 months of experience as an HR recruiter sourcing candidates through online job portals, social media, and references. His skills include end-to-end recruitment processes, types of recruitment, payroll processes, and administrative activities. He is proficient in MS Office and has completed academic projects in customer satisfaction, employee training, and industrial training.
Shivakumara D R has over 7 years of experience as a Senior Software Engineer developing embedded systems and applications using C, C++, and other programming languages. He has expertise in areas such as PCB design, IoT product development, and working with tools like KEIL, MPLAB, and Eclipse. His background includes projects in CNC application development, IoT hotel automation, and power interface unit development for telecom sites.
Sindhujha Gopi has over 7 years of experience working with Ariba Spend Management Suite, including extensive experience with Ariba Buyer 8.2 and 9r1. She has worked as a Production Support Lead and Offshore Support Lead on projects for Cummins Inc., providing level 3 technical support. Her responsibilities have included requirement gathering, design, development, testing, and production implementation. She is proficient in Ariba modules, languages like AML and SQL, and tools like Remedy and Appworx.
This document provides a summary of R.HariKrishna's professional experience and skills. He has over 4 years of experience developing software using technologies like Java, Scala, Hadoop and NoSQL databases. Some of his key projects involved developing real-time analytics platforms using Spark Streaming, Kafka and Cassandra to analyze sensor data, and using Hadoop, Hive and Pig to perform predictive analytics on server logs and calculate production credit reports by analyzing banking transactions. He is proficient in MapReduce, Pig, Hive, HDFS and has skills in machine learning technologies like Mahout.
The document provides a summary of an individual's professional experience working with big data technologies like Hadoop, Spark, Scala, Java, and AWS. It details over 9 years of experience in areas including data engineering, ETL processes, batch and stream data processing, working with technologies such as HDFS, YARN, Hive, Impala, Kafka, and databases like Oracle, MySQL, and MongoDB. Specific experiences are listed from roles at Morgan Stanley, ECA, and BMO Harris Bank involving data ingestion, transformation, analytics and reporting using Hadoop ecosystems.
This document contains Anil Kumar's resume. It summarizes his contact information, professional experience working with Hadoop and related technologies like MapReduce, Pig, and Hive. It also lists his technical skills and qualifications, including being a MapR certified Hadoop Professional. His work experience includes developing MapReduce algorithms, installing and configuring MapR Hadoop clusters, and working on projects for clients like Pfizer and American Express involving data analytics using Hadoop, Spark, and Hive.
Monika Raghuvanshi is seeking a position as a Hadoop Administrator where she can apply her 7 years of experience in Hadoop and Unix administration. She has expertise in installing, configuring, and maintaining Hadoop clusters as well as ensuring security through Kerberos and SSL. She is proficient in Linux, networking, programming languages, and databases. Her experience includes projects with Barclays, GE Healthcare, Ontario Ministry of Transportation, and Nortel where she administered Hadoop and Unix systems.
Rajeev kumar apache_spark & scala developerRajeev Kumar
Rajeev Kumar is an experienced Apache Spark and Scala developer based in Amsterdam, NL. He has over 8 years of experience working with big data technologies like Apache Spark, Scala, Java, Hadoop, and data integration tools. He is proficient in processing large structured and unstructured datasets to identify patterns and gain insights. His experience includes designing and developing Spark applications using Scala, ETL processes, data warehousing, and working with technologies like Hive, HDFS, MapReduce, Sqoop, Kafka and more.
Arindam Sengupta has over 17 years of experience in architecting, developing, implementing, and customizing client-server and web-based applications. He has extensive experience with technologies like Hadoop, HDFS, MapReduce, Pig, Hive, HBase, Spark, Java, Oracle, SQL Server, and IBM DB2. Some of his recent projects involve designing Hadoop-based solutions for data ingestion, analytics, and visualization using technologies like Flume, Sqoop, HBase, MapReduce, Spark, and REST services.
This document provides a curriculum vitae for Mr. Yuvaraj Mani, who has over 13 years of experience in big data, SQL Server, Oracle, and data warehousing technologies. It lists his professional qualifications and certifications, industry experience, and highlights of several projects between 2014-2016 where he designed architectures and implemented solutions using technologies like Hadoop, HDFS, Hive, Pig, Spark, and Sqoop. Responsibilities included ETL processing, distributed storage and computing, performance optimization, and using Agile methodologies.
Spark is a big data processing framework built in Scala that runs on the JVM. It provides speed, generality, ease of use, and accessibility for processing large datasets. Spark features include working directly on memory for speed, supporting MapReduce, lazy evaluation of queries for optimization, and APIs for Scala, R and Python. It includes Spark Streaming for real-time data, Spark SQL for SQL queries, and MLlib for machine learning. Resilient Distributed Datasets (RDDs) are Spark's fundamental data structure, and MapReduce is a programming model used for processing large amounts of data in parallel.
M. Manikyam is a software engineer with over 5 years of experience developing solutions for big data problems using Apache Spark, Hadoop, and related technologies. He has extensive hands-on experience building real-time data streaming pipelines and analytics applications on large datasets. Some of his responsibilities have included developing Spark Streaming applications, integrating Hive and HBase, and administering Hadoop clusters. He is looking for new opportunities to innovate and improve software products.
Owez Mujawar is a Hadoop Administrator with over 5 years of experience in configuring, deploying, and maintaining Hadoop clusters on AWS and private clouds. He has expertise in Apache Hadoop, Cloudera Enterprise, HDFS, YARN, Pig, Hive, Oozie, and Sqoop. Some of his responsibilities include setting up and customizing Hadoop clusters, high availability, performance tuning, and troubleshooting issues. He has worked with Vodafone and CloudAge BigData on large-scale data processing and analytics projects involving customer data.
Senior systems engineer at Infosys with 2.4yrs of experience on Bigdata & hadoopabinash bindhani
Abinash Bindhani is seeking a position as a Hadoop developer where he can utilize over 2 years of experience with Hadoop and Java technologies. He currently works as a senior systems engineer at Infosys where he has gained experience migrating data from Oracle to Hadoop platforms and collecting/analyzing log data using tools like Flume, Pig, and Hive. His technical skills include MapReduce, HBase, HDFS, Java, Spring, MySQL, and Apache Tomcat. He has expertise in Hadoop architecture, cluster concepts, and each phase of the software development life cycle.
Shubham, 7.5+ years exp, mcp, map r spark-hive-bi-etl-azure-dataengineer-mlShubham Mallick
Shubham has over 7 years of experience in data analytics and engineering. He has extensive experience with technologies like MapR-Hadoop, Spark, Python, Hive, Kafka and machine learning algorithms. He is currently a senior data analyst where he builds data pipelines and analytics solutions. Previously he has led teams and taken on roles with responsibilities like requirement gathering, data modeling, ETL development, database administration and cloud migrations. He is pursuing an M.Tech in data science and has received several awards and certifications for his work.
Vijay Muralidharan has over 5 years of experience as a Big Data Engineer working with Hadoop, Spark, Hive, Pig and other big data tools. He has a Master's in Cloud Computing and is a certified Hadoop Administrator. The document provides details of his skills, work experience implementing and managing Hadoop clusters, and education background working on projects related to performance evaluation of distributed file systems.
Cassandra Lunch #89: Semi-Structured Data in CassandraAnant Corporation
In Cassandra Lunch #89, we will discuss how to store and parse semi-structured data in Cassandra using Spark
Accompanying Blog: Coming Soon!
Accompanying YouTube: https://youtu.be/ZhNnn51BRUc
Sign Up For Our Newsletter: http://eepurl.com/grdMkn
Join Cassandra Lunch Weekly at 12 PM EST Every Wednesday: https://www.meetup.com/Cassandra-DataStax-DC/events/
Cassandra.Link:
https://cassandra.link/
Follow Us and Reach Us At:
Anant:
https://www.anant.us/
Awesome Cassandra:
https://github.com/Anant/awesome-cassandra
Cassandra.Lunch:
https://github.com/Anant/Cassandra.Lunch
Email:
solutions@anant.us
LinkedIn:
https://www.linkedin.com/company/anant/
Twitter:
https://twitter.com/anantcorp
Eventbrite:
https://www.eventbrite.com/o/anant-1072927283
Facebook:
https://www.facebook.com/AnantCorp/
Join The Anant Team:
https://www.careers.anant.us
Big Data Hoopla Simplified - TDWI Memphis 2014Rajan Kanitkar
The document provides an overview and quick reference guide to big data concepts including Hadoop, MapReduce, HDFS, YARN, Spark, Storm, Hive, Pig, HBase and NoSQL databases. It discusses the evolution of Hadoop from versions 1 to 2, and new frameworks like Tez and YARN that allow different types of processing beyond MapReduce. The document also summarizes common big data challenges around skills, integration and analytics.
This document contains the resume of Hassan Qureshi. He has over 9 years of experience as a Hadoop Lead Developer with expertise in technologies like Hadoop, HDFS, Hive, Pig and HBase. Currently he works as the technical lead of a data engineering team developing insights from data. He has extensive hands-on experience installing, configuring and maintaining Hadoop clusters in different environments.
- Gubendran Lakshmanan has over 13 years of experience in IT with expertise in Java/J2EE, AWS cloud, big data, and distributed systems.
- He has extensive experience designing, developing, and implementing applications using technologies like Java, Spring, Hibernate, AWS, Hadoop, HDFS, HBase, and NoSQL databases.
- He has worked as a senior developer and technical lead on several projects in domains like online advertising, energy, and food automation.
Vishnu has over 5 years of experience in application development using Java and big data technologies like Hadoop. He has worked on projects involving web application development, data analytics using Hadoop components like HDFS, MapReduce, Pig and Hive. His skills include Java, J2EE, databases, version control and he has experience developing applications for both web and mobile. He is currently working as a Hadoop developer at Capgemini.
This document summarizes HadoopDB, a system for building real-world applications on Hadoop. It discusses HadoopDB's architecture and components like the database connector, data loader, and catalog. It then provides two example applications - a semantic web application for biological data analysis and a business data warehousing application. The document demonstrates how to load sample datasets for each application into HadoopDB and execute sample queries on the data, including visualizing the query execution flow and demonstrating fault tolerance.
1. SAM
Samk.bigdata@gmail.com
614-664-6543
Professional Summary:
8+ yearsof overall experience in Financial, Marketing and Enterprise Application Development in
diverse industries whichincludes hands on experience in Big data ecosystem related technologies.
3+ years of Data Analyticsexperience in ApacheHadoopClouderaandHortonworksDistributions
Expertise in core Hadoop and Hadoop technology stackwhich includes HDFS, MapReduce,Oozie,
Hive,Sqoop,Pig, Flume,HBase, Spark,Storm, Kafka and Zookeeper.
Experience in AWS cloudenvironment and on s3 storage and ec2 instances and deploying in it.
In-depth knowledge of Statistics, MachineLearning,Datamining.
Well versed in installation, configuration, supporting and managing of Big Data and underlying
infrastructure of HadoopCluster.
Experienced in implementing complex algorithms on semi/unstructured data using Mapreduce
programs.
Experienced in working withstructureddata using HiveQL, joinoperations,HiveUDFs,partitions,
bucketing and internal/external tables.
Experienced in migrating ETL kind of operations using Pig transformations, operations and UDF's.
Good knowledge on Python.
Designed and developed ETLprocesses to extract and load data from Legacy System by using Talend
data integration tool.
SparkStreaming collectsthis data from Kafkain near-real-time and performs necessary
transformations and aggregation on the fly to build the common learner data model and persists the
data in NoSQL store(Hbase).
Specialization in Data Ingestion, Processing, Development from Various RDBMS data sources into a
Hadoop Cluster using MapReduce/Pig/Hive/Sqoop
Configured different topologies forStormclusterand deployed them on regular basis.
Experienced in implementing unified data platform to get data from different data sources using Apache
Kafka brokers, cluster, Javaproducers and Consumers.
Excellent WorkingKnowledge in Spark Core,SparkSQL, SparkStreaming using Scala.
Experienced in working within-memory processing frame worklike Sparktransformations,SprakQL
and Sparkstreaming.
Experienced in proving User based recommendation by implementing collaborativefiltering and matrix
factorizationand different classificationtechniques like random forest, SVM, K-NN using Spark
Mliblibrary.
Excellent understanding and knowledge of NOSQL databases like HBase,Cassandra,Mongo DB,
Teradataand onData warehouse.
Installed and configured Cassandraandgood knowledgeabout Cassandra architecture, read, write
paths and query.
Implemented Frameworks using java and python to automate the ingestion flow.
Involvedin NoSQL(Datastax Cassandra) database design, integration and implementation and written
scripts and invoked them using CQLSH.
Involvedin data modeling in Cassandraand Involvedin implementing sharding and replication
strategies in MongoDB.
2. Developed fan-out workflow using flumeforingesting data fromvarious data sources like
Webservers,RestAPI by using different sources and ingested data into Hadoop with HDFS sink.
Experienced in implementing custom interceptors and sterilizers in flumeforspecific customer
requirements.
Toolmonitored log input from several datacenters, via SparkStream, was analyzed in Apache Storm
and data was parsed and saved into Cassandra.
Experience in importing and exporting data using SqoopfromHDFS to Relational DatabaseSystems
MYSQL,SQL SERVER and vice versa.
Involvedin Database Designing including ERDiagram and Database Normalization (3NF).
Experience in developing strategies forExtraction, Transformation and Loading (ETL) data from
various sources into Data Warehouse and Data Marts using informatica.
Excellent understanding / knowledge of Hadoop architecture and various components such as HDFS,
JobTracker,TaskTracker,Name Node,Data Nodeand MapReduce programming paradigm.
Good Exposure on ApacheHadoopMapReduceprogramming,PIG ScriptingandDistribute
ApplicationandHDFS.
Experience in managing HadoopclustersusingClouderaManagerTool.
Very good experience in complete project life cycle(design, development, testing and implementation)
of Client Server and Web applications.
Developed Tableau dashboards with combination charts for clear understanding.
Improved Reports performance in Tableau using extracts and context filters
Worked on Cluster co-ordination services through Zookeeper.
Actively involvedin coding using CoreJavaand collectionAPI'ssuch as Lists, Sets and Maps.
Hands on experience in application development using Java, RDBMS,andLinuxshell scripting.
Experience on different operating systems like UNIX, Linuxand Windows.
Experience on Java Multi-Threading,Collection,Interfaces,Synchronization,andException
Handling.
Involvedin writing PL/SQLstored procedures,triggersandcomplexqueries.
Worked in Agileenvironmentwithactivescrum participation.
Technical Skills:
Hadoop/BigData HDFS, Map reduce, HBase, Pig, Hive, Sqoop, MongoDB, Cassandra, Flume,
Oozie, Zookeeper, AWS, Spark, Kafka, Teradata, Storm, ETL,Informatica,
Tableau, Talend, Scala.
Java & J2EE Technologies Core Java,Servlets, JSP,JDBC, JavaBeans, Maven, Gradle, JUnit,TestNG.
IDE’s Eclipse, Net beans, Intellij Idea.
Frameworks MVC, Struts, Hibernate, Spring.
Programminglanguages C,C++, Java,Python,Ant scripts, Linux shell scripts
Databases Oracle 11g/10g/9i, MYSQL, DB2, MS-SQL SERVER
Web Servers Web Logic, Web Sphere, Apache Tomcat,
Web Technologies HTML, XML,JavaScript, AJAX,SOAP, WSDL,JAX-RS, Restful, JAX-WS.
NetworkProtocols TCP/IP,UDP,HTTP,DNS, DHCP
VersionControls CVS, SVN, GIT.
Work Experience:
3. Client: CVS,Greensboro,NC Feb’ 15 to Till Date
Role:HadoopDeveloper
Responsibilities:
Worked on analyzing Hadoopclusterand different big data analytic tools including Pig,Hbase
databaseand Sqoop,Cassandra,zookeeper,AWS.
Evaluated business requirements and prepared detailed specifications that follow projectguidelines
required to develop written programs.
Responsible forbuilding scalabledistributed data solutions using Hadoop.
Analyzed large amounts of data sets to determine optimal way to aggregate and report on it using Map
Reduceprograms.
Implemented Map reduceprograms toretrieve Top-K results from unstructured data set.
Migrating various hive UDF’sand queries into SparkSQLfor faster requests as part of POC
implementation using Scala.
Involvedin Database Designing including ERDiagram and Database Normalization (3NF).
Optimized Map ReduceJobs touse HDFS efficiently by using various compression mechanisms.
Handled importing of data from various data sources, performed transformations using Hive,
MapReduce,loadeddatainto HDFS and Extractedthe data from HDFS to MYSQL,SQL SERVER using
Sqoop.
FollowedAgilemethodology (ScrumStandups, Sprint Planning, Sprint Review,Sprint Showcase and
Sprint Retrospective meetings).
Exported the analyzed data to the relational databases such as oracle, mysql using Sqoopfor
visualization and to generate reports for the BI team.
Experience in AWS cloudenvironment and on s3 storage and ec2 instances
Developed fan-out workflow using flumeforingesting data fromvarious data sources like
Webservers,RestAPI by using different sources and ingested data into Hadoop with HDFS sink.
Involvedin migrating MongoDBversion2.4to 2.6 and implementing new security features and
designing more efficient groups.
Installed and configured Cassandraandgood knowledgeabout Cassandra architecture, read, write
paths and query.
Developed Tableau dashboards with combination charts for clear understanding.
Improved Reports performance in Tableau using extracts and context filters
Implemented various ETL solutions as per the business requirement using informatica
Experience with creating ETLjobs to load JSON data and serverdata into MongoDB andtransformed
MongoDBinto theDataWarehouse.
Designed and developed ETLprocesses to extract and load data from Legacy System by using talend
data integration tool.
Extensively used components like tWaitForFile, tIterateToFlow,tFlowToIterate,tHashoutput,
tHashInput, tMap, tRunjob, tJava,tNormalize and tfile components to create talend jobs.
Implemented Frameworks using java and python to automate the ingestion flow.
Involvedin data modeling in CassandraandMongoDB andinvolvedin choosing indexes and primary
keys based on the client requirement.
Configured Spark streaming to receive real time data fromthe Kafka and store the stream data to HDFS
using Scala.
Used Sparkfor Parallel data processing and better performances using Scala.
Extensively used Pig fordata cleansing and extract the data from the web server output files to load
into HDFS.
Developed a data pipeline using Kafka and Stormto store data into HDFS.
4. Implemented Kafka Javaproducers, create custom partitions, configured brokers and implemented
High level consumers to implement data platform.
Implemented Storm topologies topreprocess data, implemented custom grouping to configure
partitions.
Managed and reviewed Hadooplogfiles.
Involvedin creating Hivetables, loading with data and writing hivequeries which willrun internally in
MapReduce way.
Used Hiveto analyze the partitioned and bucketed data and compute various metrics forreporting.
Installed and configured Pigand also written Pig Latin scripts.
Responsible tomanage data coming fromdifferent sources such as SQL SERVER.
Environment:Hadoop,MapReduce, Agile, HDFS, Hive, Pig, Java,SQL SERVER, Sqoop, Java (jdk1.6),
Spark, kafka,AWS, MongoDB,Storm, Cassandra, ETL,Informatica, Python,Tableau, Talend, scala.
Client:BB&T Bank-Charlotte,NC Feb’ 14 to Jan’ 15
Role:HadoopDeveloper
Responsibilities:
Installed and configured Cassandraandgood knowledgeabout Cassandra architecture, read, write
paths and quering using Cassandrashell.
Worked on writing MapReducejobs to discover trends in data usage by customers.
Worked on and designed Big Data analytics platform forprocessing customer interface preferences and
comments using Java, Hadoop,HiveandPig.
Involvedin hive-Hbaseintegrationby creating hiveexternal tables and specifying storage as Hbase
format.
Importing and exporting data into HDFS and Hiveusing Sqoopfromoracle and vice versa.
Exported the analyzed data to the relational databases such as oracle using Sqoopforvisualization and
to generate reports for the BI team.
Designed and developed ETLprocesses to extract and load data from Legacy System by using talend
data integration tool.
Implemented Frameworks using java and python to automate the ingestion flow.
Experienced in defining job flowsto run multiple MapReduceand Pigjobsusing Oozie.
Installed and configured Hiveand also written HiveQL scripts.
Experience with loading the data into relational database for reporting, dash boarding and ad-hoc
analyses, whichrevealed waysto lower operating costs and offsetthe rising costof programming.
Experience with creating ETLjobs to load JSON data and server data into MongoDB andtransformed
MongoDB intotheData Warehouse.
Involvedin ETLcode deployment, PerformanceTuning of mappings in Informatica.
Created reports and dashboards using structuredand unstructureddata.
Experienced with performing analytics on Time Series data using HBase.
Implemented HBaseco-processors, Observers to workas event based analysis.
Hands on Installing and configuring nodes CDH4HadoopCluster onCentOS.
Implemented HiveGeneric UDF'sto implement business logic.
Experienced with accessing Hivetables to perform analytics from java applications using JDBC.
Experienced in running batch processes using Pig Scripts and developed Pig UDFsfor data
manipulation according to Business Requirements.
Experience with streaming workflow operations and Hadoop jobs using Oozieworkflow and
scheduled through AUTOSYSon a regular basis.
5. FollowedAgilemethodology (ScrumStandups, Sprint Planning, Sprint Review,Sprint Showcase and
Sprint Retrospective meetings).
Performed operation using Partitioning pattern in MapReduceto move records into different
categories.
Developed SparkSQL scripts and involvedin converting hiveUDF’s to SparkSQL UDF’s.
Responsible forbatch processing and real time processing in HDFS and NOSQL Databases.
Responsible forretrieval of Data from Casandra and ingestion to PIG.
Experience in customizing map reduce frameworkat various levels by generating Custom Input
formats, Record Readers, Partitioner and Data types.
Experienced with multiple file in HIVE,AVRO,Sequence file formats.
Created and maintained Technical documentation for launching HADOOPClusters and forexecuting
PigScript.
Implemented business logic by writing PigUDF's in Javaand used various UDFs fromPiggybanks and
other sources.
Environment:Casandra, Map jobs, Spark SQL, Agile, ETL,Pig Scripts, Flume, Hadoop BI, Pig UDF’s,
Oozie, AVRO,Hive, Map Reduce, Java, Eclipse, Zookeeper, Informatica,oracle, Python,Talend.
Client: Epsilon,Danbury,CT Aug’ 12to Dec’ 13
Role:HadoopDeveloper
Responsibilities:
• Involvedin the Complete Softwaredevelopment life cycle(SDLC) to develop the application.
• Worked on analyzing Hadoop cluster and different big data analytic tools including Pig, Hbase
databaseand Sqoop,Cassandra,zookeeper.
• Involvedin loading data from LINUX filesystem to HDFS.
• Exported the analyzed data to the relational databases using Sqoopforvisualization and to generate
reports forthe BI team.
• FollowedAgilemethodology (ScrumStandups, Sprint Planning, Sprint Review,Sprint Showcase and
Sprint Retrospective meetings).
• Importing and exporting data into HDFS and Hiveusing SqoopfromOracleand viceversa.
• Implemented test scripts to support test driven development and continuous integration.
• Developed multiple MapReducejobs in java fordata cleaning.
• Installed and configured HadoopMapReduce,HDFS,DevelopedmultipleMapReducejobs injava
for data cleaning and preprocessing.
• Created Pig Latin scripts to sort, group, join and filter the enterprise wise data.
• Involvedin creating Hivetables, loadingwithdataand writinghivequeries that willrun internally
in Map Reduce way.
• Supported MapReducePrograms thoseare running on the cluster.
• Analyzed large data sets by running HivequeriesandPigscripts.
• Implemented Frameworks using java and python to automate the ingestion flow.
• Worked on tuning the performance Pig queries.
• Mentored analyst and test team for writing HiveQueries.
• Installed Oozieworkflow engine to run multiple Mapreducejobs.
• Worked withapplication teams to install operating system, Hadoop updates, patches, version upgrades
as required.
• Worked on zookeeper forcoordinatingbetween differentmaster node and datanodes
6. Environment:Hadoop,HDFS, Map Reduce, Agile, Hive, Pig, Sqoop, Linux, Java, Oozie,Hbase, zookeeper,
SQL SERVER,python.
Client:MeridianEnterprise,StLouis,MO Nov’ 10to Jul’ 12
Role:Java/J2EE Developer
Responsibilities:
Workwith business users to determine requirements and technical solutions.
FollowedAgilemethodology (ScrumStandups, Sprint Planning, Sprint Review,Sprint Showcase and
Sprint Retrospective meetings).
Developed business components using core java concepts and classes like Inheritance,
Polymorphism,Collections,SerializationandMultithreadingetc.
Used SPRING framework that handles application logic and makes calls to business make them as
SpringBeans.
Implemented, configured data sources, session factory and used Hibernate Template to integrate
SpringwithHibernate.
Developed web services to allow communication between applications through SOAP over HTTP with
JMS and mule ESB.
Actively involved in coding using Core Java and collection API's such as Lists, Sets and Maps
Developed a Web Service(SOAP,WSDL) that is shared between front end and cable bill review
system.
Implemented Rest based web service using JAX-RS annotations, Jersey implementation for data
retrieval with JSON.
Developed MAVENscriptsto build and deploy the application onto Web logic Application Server and
ran UNIX shell scripts and implemented auto deployment process.
Used Mavenas the build tool and is scheduled/triggered by Jenkins(build tool).
Develop JUNIT test cases for application unit testing.
Implement Hibernatefordata persistence and management.
Used SOAP UItool for testing web services connectivity.
Used SVNas version control to checkin the code, Created branches and tagged the code in SVN.
Used RESTFULServices to interact with the Client by providing the RESTFULURLmapping.
Used Log4j frameworkto log/track application and debugging.
Environment:JDK1.6, Eclipse IDE,Core Java,J2EE,Spring, Hibernate, Unix, Web Services, SOAP UI,
Maven, Web logic Application Server, SQL Developer,Camel, Junit, SVN, Agile, SONAR, Log4j, REST,
Log4j, JSON, JBPM,Agile.
Client:MOBILINK – DSS MOBILE COMMUNICATIONS,INDIA Aug’ 07to Sep’ 10
Role:JuniorJavaDeveloper
Responsibilities:
Involved in analysis, design and development of Expense Processing system.
Created used interfaces using JSP.
Developed the Web Interface using Servlets, Java Server Pages, HTML and CSS.
Developed the DAO objects using JDBC.
Business Services using the Servlets and Java.
Design and development of User Interfaces and menus using HTML 5, JSP, Java Script, client side and
server side validations.
Developed GUI using JSP, Struts frame work.
Involvedin developing the presentation layer using SpringMVC/AngularJS/JQuery.
Involved in designing the user interfaces using Struts Tiles Framework.
Used Spring 2.0 Framework for Dependency injection and integrated with the Struts Framework and
Hibernate.
Used Hibernate 3.0 in data access layer to access and update information in the database.
Experience in SOA (Service Oriented Architecture) by creating the web services with SOAP and WSDL.
7. Developed JUnit test cases for all the developed modules.
Used Log4J to capture the log that includes runtime exceptions, monitored error logs and fixed the
problems.
Used RESTFULServices to interact with the Client by providing the RESTFULURLmapping.
Used CVS for version control across common source code used by developers.
Used ANT scripts to build the application and deployed on Web logic Application Server 10.0.
Environment:-Struts1.2, Hibernate3.0, Spring2.5 ,JSP,Servlets, XML,SOAP,WSDL,JDBC, JavaScript,
HTML, CVS, Log4J, JUNIT,Web logic App server, Eclipse, Oracle,Restful.
Education:
Bachelors in Technology – Computer Science And Engineering
Jawaharlal Nehru Technological University
References:
Provided upon request