This document provides instructions for installing Hadoop for development and production using open source distributions. It discusses installing Hadoop sandboxes like Cloudera Quickstart and Hortonworks Sandbox for development on local machines or clouds. It also covers installing Cloudera Express clusters for production on AWS, including launching EC2 instances and configuring security groups.
This workshop is for a "Big Data using Hadoop course" at IMC Institute in March 2015. The workshop is based on Apache Hadoop and using an EC2 server on AWS.
This workshop is for a "Big Data using Hadoop course" at IMC Institute in March 2015. The workshop is based on Apache Hadoop and using an EC2 server on AWS.
Real-Time Big Data at In-Memory Speed, Using StormNati Shalom
Storm, a popular framework from Twitter, is used for real-time event processing. The challenge presented is how to manage the state of your real-time data processing at all times. In addition, you need Storm to integrate with your batch processing system (such as Hadoop) in a consistent manner.
This session will demonstrate how to integrate Storm with an in-memory database/grid, and explore various strategies for integrating the data grid with Hadoop and Cassandra, seamlessly. By achieving smooth integration with consistent management, you will be able to easily manage all the tiers of you Big Data stack in a consistent and effective way.
- See more at: http://nosql2013.dataversity.net/sessionPop.cfm?confid=74&proposalid=5526#sthash.FWIdqRHh.dpuf
Real time big data analytics with Storm by Ron Bodkin of Think Big AnalyticsData Con LA
This talk provides an overview of the open source Storm system for processing Big Data in realtime. The talk starts with an overview of the technology, including key components: Nimbus, Zookeeper, Topology, Tuple, Trident. It looks at integration with Hadoop through YARN and recent improvements. The presentation then dives into the complex Big Data architecture in which Storm can be integrated . The result is a compelling stack of technologies including integrated Hadoop clusters, MPP, and NoSQL databases.
After this, we look at example use cases for Storm: real-time advertising statistics, updating a Machine Learned model for content popularity predictions, and financial compliance monitoring.
API analytics with Redis and Google Bigquery. NoSQL matters editionjavier ramirez
At teowaki we have a system for API use analytics using Redis as a fast intermediate store and bigquery as a big data backend. As a result, we can launch aggregated queries on our traffic/usage data in a few seconds and we can try and find for usage patterns that wouldn’t be obvious otherwise. In this session I will speak of the alternatives we evaluated and how we are using Redis and Bigquery to solve our problem.
How to test infrastructure code: automated testing for Terraform, Kubernetes,...Yevgeniy Brikman
This talk is a step-by-step, live-coding class on how to write automated tests for infrastructure code, including the code you write for use with tools such as Terraform, Kubernetes, Docker, and Packer. Topics covered include unit tests, integration tests, end-to-end tests, test parallelism, retries, error handling, static analysis, and more.
Real Time Graph Computations in Storm, Neo4J, Python - PyCon India 2013Sonal Raj
This talk briefly outlines the Storm framework and Neo4J graph database, and how to compositely use them to perform computations on complex graphs in Python using the Petrel and Py2neo packages. This talk was given at PyCon India 2013.
A comprehensive walkthrough of how to manage infrastructure-as-code using Terraform. This presentation includes an introduction to Terraform, a discussion of how to manage Terraform state, how to use Terraform modules, an overview of best practices (e.g. isolation, versioning, loops, if-statements), and a list of gotchas to look out for.
For a written and more in-depth version of this presentation, check out the "Comprehensive Guide to Terraform" blog post series: https://blog.gruntwork.io/a-comprehensive-guide-to-terraform-b3d32832baca
My talk at FullStackFest, 4.9.2017. Become more familiar with managing infrastructure using Terraform, Packer and deployment pipeline. Code repository - https://github.com/antonbabenko/terraform-deployment-pipeline-talk
In this Introduction to Apache Sqoop the following topics are covered:
1. Why Sqoop
2. What is Sqoop
3. How Sqoop Works
4. Importing and Exporting Data using Sqoop
5. Data Import in Hive and HBase with Sqoop
6. Sqoop and NoSql data store i.e. MongoDB
7. Resources
Real-Time Big Data at In-Memory Speed, Using StormNati Shalom
Storm, a popular framework from Twitter, is used for real-time event processing. The challenge presented is how to manage the state of your real-time data processing at all times. In addition, you need Storm to integrate with your batch processing system (such as Hadoop) in a consistent manner.
This session will demonstrate how to integrate Storm with an in-memory database/grid, and explore various strategies for integrating the data grid with Hadoop and Cassandra, seamlessly. By achieving smooth integration with consistent management, you will be able to easily manage all the tiers of you Big Data stack in a consistent and effective way.
- See more at: http://nosql2013.dataversity.net/sessionPop.cfm?confid=74&proposalid=5526#sthash.FWIdqRHh.dpuf
Real time big data analytics with Storm by Ron Bodkin of Think Big AnalyticsData Con LA
This talk provides an overview of the open source Storm system for processing Big Data in realtime. The talk starts with an overview of the technology, including key components: Nimbus, Zookeeper, Topology, Tuple, Trident. It looks at integration with Hadoop through YARN and recent improvements. The presentation then dives into the complex Big Data architecture in which Storm can be integrated . The result is a compelling stack of technologies including integrated Hadoop clusters, MPP, and NoSQL databases.
After this, we look at example use cases for Storm: real-time advertising statistics, updating a Machine Learned model for content popularity predictions, and financial compliance monitoring.
API analytics with Redis and Google Bigquery. NoSQL matters editionjavier ramirez
At teowaki we have a system for API use analytics using Redis as a fast intermediate store and bigquery as a big data backend. As a result, we can launch aggregated queries on our traffic/usage data in a few seconds and we can try and find for usage patterns that wouldn’t be obvious otherwise. In this session I will speak of the alternatives we evaluated and how we are using Redis and Bigquery to solve our problem.
How to test infrastructure code: automated testing for Terraform, Kubernetes,...Yevgeniy Brikman
This talk is a step-by-step, live-coding class on how to write automated tests for infrastructure code, including the code you write for use with tools such as Terraform, Kubernetes, Docker, and Packer. Topics covered include unit tests, integration tests, end-to-end tests, test parallelism, retries, error handling, static analysis, and more.
Real Time Graph Computations in Storm, Neo4J, Python - PyCon India 2013Sonal Raj
This talk briefly outlines the Storm framework and Neo4J graph database, and how to compositely use them to perform computations on complex graphs in Python using the Petrel and Py2neo packages. This talk was given at PyCon India 2013.
A comprehensive walkthrough of how to manage infrastructure-as-code using Terraform. This presentation includes an introduction to Terraform, a discussion of how to manage Terraform state, how to use Terraform modules, an overview of best practices (e.g. isolation, versioning, loops, if-statements), and a list of gotchas to look out for.
For a written and more in-depth version of this presentation, check out the "Comprehensive Guide to Terraform" blog post series: https://blog.gruntwork.io/a-comprehensive-guide-to-terraform-b3d32832baca
My talk at FullStackFest, 4.9.2017. Become more familiar with managing infrastructure using Terraform, Packer and deployment pipeline. Code repository - https://github.com/antonbabenko/terraform-deployment-pipeline-talk
In this Introduction to Apache Sqoop the following topics are covered:
1. Why Sqoop
2. What is Sqoop
3. How Sqoop Works
4. Importing and Exporting Data using Sqoop
5. Data Import in Hive and HBase with Sqoop
6. Sqoop and NoSql data store i.e. MongoDB
7. Resources
Apache Sqoop: A Data Transfer Tool for HadoopCloudera, Inc.
Apache Sqoop is a tool designed for efficiently transferring bulk data between Hadoop and structured datastores such as relational databases. This slide deck aims at familiarizing the user with Sqoop and how to effectively use it in real deployments.
Updated version at https://www.slideshare.net/GiulioRoggero/kanban-board-82363781
Do you have a team that works on both project and maintenance? Do you need to organize your team activities? Do you have a lot of activities in parallel and the time to market it's a problem? With a Kanban board and an Agile approach you can solve your problems!
Take a look of the animation of the slides to discover how it works.
This article will help you fetch details about the Ubuntu based AWS EC2 instance. You need to deploy the Python (2.7) based REST Services in Apache webserver. The core of application is Python DJango framework, which uses a custom virtual environment (vitualenv). The Apache uses mod_wsgi for connecting the WSGI application and mod_sec for security purposes.
Deploying Django with Apache and mod_wsgi is a method to get Django into production. mod_wsgi is an Apache module which is supposed to host any Python WSGI application, which includes Django. Django can work with any version of Apache that supports mod_wsgi.
Read the article further, to understand the step-by-step deployment process.
Ubuntu Server is lean, fast and powerful. Its services are reliable, predictable and economical. It is the perfect base on which you can build your instances. Django is a web framework which is written in Python. One can easily guess that everything, in Django, is also done in Python. Django was developed to simplify the creation of database driven sites. The best feature in Django is that it, probably, is the fastest website framework to create a fully functioning website.
WordPress + Amazon Web Services Hands-on WARSAWMatt Pilarski
Let's learn how to launch WordPress Powered by AMIMOTO HHVM on AWS: Hands-on Workshop (where you can learn how to self-host your WordPress site & many more!)
AMIMOTO delivers High Performance WordPress Cloud Hosting on Amazon Web Services.
https://amimoto-ami.com/
AMIMOTO: WordPress + Amazon Web Services Hands-on PARISKel
WordPress, AWS & AMIMOTO introduction, community, first steps, what you can do, resources, use cases, hands-on, SFTP, managing database, cloudformation.
Hammr Project Update: Machine Images and Docker Containers for your Cloud, OW...OW2
Hammr is an OW2 open source, command-line tool for creating consistent and repeatable machine images for different cloud or virtual environments, or migrating live systems from one environment to another. Designed for cloud era environments, where agility and automation are key, hammr helps organizations automate the creation of machine images for hybrid environments. This presentation will focus on the hybrid capabilities of hammr for any virtual or cloud target environment. It will also include a focus on DevOps and Docker integration, and show how hammr can be used to quickly build and run Docker images, helping accelerate development and test processes among other benefits. Finally, we will present the latest hammr features, including the ability for cloud providers to customize target platforms and expose their own IaaS infrastructure as top-level branded objects accessible via hammr, thus easing the path from user images to their cloud infrastructure.
It Works On My Machine: Vagrant for Software DevelopmentCarlos Perez
Vagrant is a command-line interface for simplifying the use of virtual machines (VM's). Vagrant allows teams to standardize their software development workflows by offering a uniform and portable interface to provision and run VM's on different operating platforms such as Mac OS X, Windows, and Linux and achieve identical results. It supports all the major virtualization solutions such as VirtualBox, VMWare, and Hyper-V and supports configuration tools that range from simple shell scripts to powerful Chef and Puppet recipes. Developers simply invoke “vagrant up” and immediately enjoy a standard, consistent, and reproducible VM for software development and testing.
Install and manage windows nano server 2016 step by stepAhmed Abdelwahed
In this lab, we'll provide complete information to create and configure Windows Server 2016 Nano Server also we'll add, configure and test DNS and file server on Nano Server.
Table of Contents
Lab Objective
Windows Nano Server 2016 overview
Current infrastructure environment
Create Nano Server
Step 1: Download and install Windows Assessment and Deployment Kit (ADK)
Step 2: Download and install Nano Server Image Builder
Step 3: Create Windows Nano Server 2016 VHD image file format
Step 4: Connect to the created Nano Server VHD through Hyper-V console
Manage Windows Nano Server 2016 (locally)
Add static IP address
Allow ICMP (Ping)
Manage Windows Nano Server 2016 (Remotely)
Check Nano server installed roles
Manage Nano server remotely through PowerShell
Add file server role service
Testing
Test 1: Create folder share on Nano Server
Test 2: add DNS Server Role
Connect to DNS on Nano Server and add DNS secondary zone from ITPROLEARN.COM DNS
Test Name Resolution and Internet Connectivity through Nano Server DNS
Choosing a right cloud provider to host your mission-critical application on the cloud is a multi-dimensional challenge faced by most of the companies. This presentation provides you information on the various benefits, challenges, and insights while evaluating a cloud provider and helps you quickly make a decision to choose a right cloud partner for your business to host your application.
Docker released 1.13. As usual full of features and bug fixes. During the Dublin Meetup I presented news and I run a demo about image squash, stack and compose v3, secrets and so on.
The answer is no, but it will help! There's not much commentary on those slides actually, you had to be there. The actual model is at https://community.apache.org/apache-way/apache-project-maturity-model.html
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
2. Thanachart Numnonda, thanachart@imcinstitute.com Oct 2016Hadoop Installation
Agenda
●
Introduction
●
Part 1: Hadoop Development Installation
●
Part 1A: Cloudera Quickstart on Cloud
●
Part 1B: Hortonworks Sandbox on Azure
●
Part 2: Hadoop Production
●
Install Cloudera Express Cluster on AWS
●
Part 3: Hadoop as a Service
●
Launch EMR cluster on AWS
Hive.apache.org
3. Thanachart Numnonda, thanachart@imcinstitute.com Oct 2016Hadoop Installation
This hands-on lab aims to guide
a student to install Hadoop
for development using Sandbox
and for production on Clusters
4. Thanachart Numnonda, thanachart@imcinstitute.com Oct 2016Hadoop Installation
Hadoop Distribution
●
On-Premise
●
Pure Apache Hadoop
●
Cloudera
●
Hortonworks
●
MapR
●
Pivotal
●
IBM Infosphere BigInsight
●
On-Cloud (Hadoop as a Service)
●
Amazon EMR
●
Microsoft Azure HDInsight
●
Google Cloud Platform
Hive.apache.org
8. Thanachart Numnonda, thanachart@imcinstitute.com Oct 2016Hadoop Installation
Hadoop Development: Sandbox
●
Recommended
●
Cloudera Quickstart
●
Hortonworks Sandbox
●
On PC/Mac
●
Require to install VMWare Player or VirtualBox
●
On Linux Virtual Server
●
Need to install Docker Engine
●
Microsoft Azure has a Hortonworks Sandbox Image
which can easily launched.
Hive.apache.org
12. Thanachart Numnonda, thanachart@imcinstitute.com Oct 2016Hadoop Installation
Hands-On
Launch a virtual server (EC2)
on Amazon Web Services
(Note: You can skip this session if you use your own
computer or another cloud service)
15. Thanachart Numnonda, thanachart@imcinstitute.com Oct 2016Hadoop Installation
Virtual Server
This lab will use a EC2 virtual server to install a
Cloudera Cluster using the following features:
Ubuntu Server 14.04 LTS
One m3.xLarge 4vCPU, 15 GB memory,80 GB
SSD
Security group: default
Keypair: imchadoop
24. Thanachart Numnonda, thanachart@imcinstitute.com Oct 2016Hadoop Installation
Review an instance and rename one instance as a
master / click Connect for an instruction to connect
to the instance
40. Thanachart Numnonda, thanachart@imcinstitute.com Oct 2016Hadoop Installation
Option: Launch a virtual server
on Microsoft Azure for Cloudera
Sandbox
(Note: You do not need to do this if you use your
own computer or another cloud service)
44. Thanachart Numnonda, thanachart@imcinstitute.com Oct 2016Hadoop Installation
Virtual Server
This lab will use an Azure virtual server to install a
Cloudera Quickstart Docker using the following
features:
Ubuntu Server 14.04 LTS
DS3_V2 Standard 4 Core, 14 GB memory,28
GB SSD
46. Thanachart Numnonda, thanachart@imcinstitute.com Oct 2016Hadoop Installation
On the Basics page, enter:
1. a name for the VM
2. a username for the Admin User
3. the Authentication Type set to password
4. a password
5. a resource group name
62. Thanachart Numnonda, thanachart@imcinstitute.com Oct 2016Hadoop Installation
Uncompress the docker zip file
$ gunzip HDP_2.5_docker.tar.gz
$ sudo docker load < HDP_2.5_docker.tar
(This may take several minutes)
Load the Docker image
64. Thanachart Numnonda, thanachart@imcinstitute.com Oct 2016Hadoop Installation
Virtual Server
This lab will use an Azure virtual server to install a
Hortonworks Sandbox using the following features:
Hortonworks Sandbox with HDP 2.4
DS3_V2 Standard 4 Core, 14 GB memory,28
GB SSD
67. Thanachart Numnonda, thanachart@imcinstitute.com Oct 2016Hadoop Installation
On the Basics page, enter:
1. a name for the VM
2. a username for the Admin User
3. the Authentication Type set to password
4. a password
5. a resource group name
86. Thanachart Numnonda, thanachart@imcinstitute.com Oct 2016Hadoop Installation
Virtual Server
This lab will use a EC2 virtual server to install a
Cloudera Cluster using the following features:
Ubuntu Server 14.04 LTS
Four m3.xLarge 4vCPU, 15 GB memory,80 GB
SSD
Security group: default
Keypair: imchadoop
95. Thanachart Numnonda, thanachart@imcinstitute.com Oct 2016Hadoop Installation
Review an instance and rename one instance as a
master / click Connect for an instruction to connect
to the instance
104. Thanachart Numnonda, thanachart@imcinstitute.com Oct 2016Hadoop Installation
Login to Cloudera Manager
Wait several minutes for the Cloudera Manager Server to complete its startup.
Then running web browser: http:// public-dns: 7180
113. Thanachart Numnonda, thanachart@imcinstitute.com Oct 2016Hadoop Installation
Define user as ubuntu & Browse the private key (imchadoop.pem) file which we
have downloaded in the previous part. Keep Passphrase as blank