The document provides an overview of application lifecycle management (ALM) in a serverless world. It discusses key concepts like continuous integration/delivery and testing practices for serverless applications. Serverless architectures using AWS Lambda and API Gateway are highlighted, along with how to manage deployments, configurations, and monitor applications.
Today’s cutting edge companies have software release cycles measured in days instead of months. This agility is enabled by the DevOps practice of continuous integration and delivery, which automates building, testing, and deploying all code changes. This automation helps you catch bugs sooner and accelerates developer productivity. In this session, we’ll share the processes followed by Amazon engineers and discuss how you can bring them to your company by using a set of application lifecycle management tools from AWS: the newly announced AWS CodeBuild service, AWS CodePipeline, and AWS CodeDeploy.
AWS re:Invent 2016: Chalk Talk: Succeeding at Infrastructure-as-Code (GPSCT312)Amazon Web Services
- Infrastructure as code is the practice of provisioning and managing infrastructure using code and software development techniques like version control. This allows infrastructure changes to be tested and deployed in a consistent, repeatable way.
- AWS services like CloudFormation, OpsWorks, and CodeDeploy allow defining infrastructure as code templates and automating the deployment of applications and infrastructure changes across environments like development, testing, and production.
- CloudFormation templates define AWS resources and their dependencies and can be used to create matching environments in different stages. OpsWorks and CodeDeploy help manage application deployments and ongoing configuration of running systems.
Automating Software Deployments with AWS CodeDeploy by Matthew Trescot, Manag...Amazon Web Services
This document discusses AWS CodeDeploy, a service that automates software deployments to EC2 instances and on-premises servers. It provides an overview of CodeDeploy's key concepts including applications, deployment groups, deployment configurations, and hooks. It also shows examples of how CodeDeploy can be used for automated deployments across development, test, and production environments. The document suggests additional features like CloudFormation support and integration with CI/CD tools.
AWS January 2016 Webinar Series - Introduction to Deploying Applications on AWSAmazon Web Services
Based on your specific needs and the nature of your application, AWS offers a variety of services for getting your application up and running. You may want to launch and scale a web application or you may want to host a microservices application using Docker containers. How do you decide which service to use and when?
In this webinar, we will provide an overview of the AWS services that help simplify launching and running your application in the cloud. We will discuss the strengths of each service and provide a framework for understanding when to use them.
Learning Objectives:
Understand the primary services for deploying your application on AWS
Learn the basics of AWS Elastic Beanstalk, AWS CodeDeploy, and Amazon EC2 Container Service
Gain an understanding of the strengths of each service and when to use them
Who Should Attend:
Developers, DevOps Engineers, IT Professionals
DevOps on AWS: Deep Dive on Continuous Delivery and the AWS Developer ToolsAmazon Web Services
Today’s cutting-edge companies have software release cycles measured in days instead of months. This agility is enabled by the DevOps practice of continuous delivery, which automates building, testing, and deploying all code changes. This automation helps you catch bugs sooner and accelerates developer productivity. In this session, we’ll share the processes that Amazon’s engineers use to practice DevOps and discuss how you can bring these processes to your company by using a new set of AWS tools (AWS CodePipeline and AWS CodeDeploy). These services were inspired by Amazon's own internal developer tools and DevOps culture.
(DVO306) AWS CodeDeploy: Automating Your Software DeploymentsAmazon Web Services
So you’ve written some code. Now what? How do you make it available to your customers in an efficient and reliable manner? Learn how you can use AWS CodeDeploy to easily and quickly push your application updates. This talk will introduce you to the basics of CodeDeploy: key concepts, how it works, where it fits in your release process, and some deployment strategies to get you started on the right foot. We’ll walk through several demos, going from a basic sample deployment to a live update of a large multi-instance fleet, giving you a sense for how CodeDeploy can grow with your needs.
Releasing Software Quickly and Reliably With AWS CodePipeline by Mark Mansour...Amazon Web Services
The document discusses AWS CodePipeline and continuous delivery. It covers how Amazon evolved their own release processes, an overview of common release process terminology and phases, a tour of AWS CodePipeline features and concepts, how to extend AWS CodePipeline using custom actions or Lambda functions, and examples of building custom actions and job workers.
The document provides an overview of application lifecycle management (ALM) in a serverless world. It discusses key concepts like continuous integration/delivery and testing practices for serverless applications. Serverless architectures using AWS Lambda and API Gateway are highlighted, along with how to manage deployments, configurations, and monitor applications.
Today’s cutting edge companies have software release cycles measured in days instead of months. This agility is enabled by the DevOps practice of continuous integration and delivery, which automates building, testing, and deploying all code changes. This automation helps you catch bugs sooner and accelerates developer productivity. In this session, we’ll share the processes followed by Amazon engineers and discuss how you can bring them to your company by using a set of application lifecycle management tools from AWS: the newly announced AWS CodeBuild service, AWS CodePipeline, and AWS CodeDeploy.
AWS re:Invent 2016: Chalk Talk: Succeeding at Infrastructure-as-Code (GPSCT312)Amazon Web Services
- Infrastructure as code is the practice of provisioning and managing infrastructure using code and software development techniques like version control. This allows infrastructure changes to be tested and deployed in a consistent, repeatable way.
- AWS services like CloudFormation, OpsWorks, and CodeDeploy allow defining infrastructure as code templates and automating the deployment of applications and infrastructure changes across environments like development, testing, and production.
- CloudFormation templates define AWS resources and their dependencies and can be used to create matching environments in different stages. OpsWorks and CodeDeploy help manage application deployments and ongoing configuration of running systems.
Automating Software Deployments with AWS CodeDeploy by Matthew Trescot, Manag...Amazon Web Services
This document discusses AWS CodeDeploy, a service that automates software deployments to EC2 instances and on-premises servers. It provides an overview of CodeDeploy's key concepts including applications, deployment groups, deployment configurations, and hooks. It also shows examples of how CodeDeploy can be used for automated deployments across development, test, and production environments. The document suggests additional features like CloudFormation support and integration with CI/CD tools.
AWS January 2016 Webinar Series - Introduction to Deploying Applications on AWSAmazon Web Services
Based on your specific needs and the nature of your application, AWS offers a variety of services for getting your application up and running. You may want to launch and scale a web application or you may want to host a microservices application using Docker containers. How do you decide which service to use and when?
In this webinar, we will provide an overview of the AWS services that help simplify launching and running your application in the cloud. We will discuss the strengths of each service and provide a framework for understanding when to use them.
Learning Objectives:
Understand the primary services for deploying your application on AWS
Learn the basics of AWS Elastic Beanstalk, AWS CodeDeploy, and Amazon EC2 Container Service
Gain an understanding of the strengths of each service and when to use them
Who Should Attend:
Developers, DevOps Engineers, IT Professionals
DevOps on AWS: Deep Dive on Continuous Delivery and the AWS Developer ToolsAmazon Web Services
Today’s cutting-edge companies have software release cycles measured in days instead of months. This agility is enabled by the DevOps practice of continuous delivery, which automates building, testing, and deploying all code changes. This automation helps you catch bugs sooner and accelerates developer productivity. In this session, we’ll share the processes that Amazon’s engineers use to practice DevOps and discuss how you can bring these processes to your company by using a new set of AWS tools (AWS CodePipeline and AWS CodeDeploy). These services were inspired by Amazon's own internal developer tools and DevOps culture.
(DVO306) AWS CodeDeploy: Automating Your Software DeploymentsAmazon Web Services
So you’ve written some code. Now what? How do you make it available to your customers in an efficient and reliable manner? Learn how you can use AWS CodeDeploy to easily and quickly push your application updates. This talk will introduce you to the basics of CodeDeploy: key concepts, how it works, where it fits in your release process, and some deployment strategies to get you started on the right foot. We’ll walk through several demos, going from a basic sample deployment to a live update of a large multi-instance fleet, giving you a sense for how CodeDeploy can grow with your needs.
Releasing Software Quickly and Reliably With AWS CodePipeline by Mark Mansour...Amazon Web Services
The document discusses AWS CodePipeline and continuous delivery. It covers how Amazon evolved their own release processes, an overview of common release process terminology and phases, a tour of AWS CodePipeline features and concepts, how to extend AWS CodePipeline using custom actions or Lambda functions, and examples of building custom actions and job workers.
The document discusses AWS Code services that can be used to automate the software release process. It describes CodeCommit for source control, CodeBuild for building and testing code, and CodeDeploy for deploying builds to EC2/on-premises servers. CodePipeline allows orchestrating builds and deployments across different environments through a visual workflow.
Configuration Management with AWS OpsWorks - November 2016 Webinar SeriesAmazon Web Services
To compete in today’s cloud-driven market, you need tools to help you automate bootstrapping, configuration, deployment, monitoring, scaling, recovery, and more. AWS OpsWorks helps you define, deploy, and operate your applications on AWS using Chef. AWS OpsWorks handles the heavy lifting for you, and it does it by using a single console. In this session, we will demonstrate how to bake automation and predictability into your application’s lifecycle with AWS OpsWorks, and showcase one of our latest collaborations with AWS CodePipeline that streamlines CI/CD processes.
Learning Objectives:
• Learn about AWS OpsWorks
• Learn about configuration management
• Learn about AWS CodePipeline
• Learn how to deploy applications using AWS CodePipeline and AWS OpsWorks
"Amazon Inspector is a new service from AWS that identifies security issues in your application deployments. Use Inspector with your applications to assess your security posture and identify areas that can be improved. Inspector works with your Amazon EC2 instances to monitor activity in your applications and system.
This session will cover getting started with Inspector, how to automate the process, how to manage and act on findings, and additional ways you can enhance your development and release lifecycle using Inspector."
Getting Started With Continuous Delivery on AWS - AWS April 2016 Webinar SeriesAmazon Web Services
Today’s cutting-edge companies have software release cycles measured in days instead of months. This agility is enabled by the DevOps practice of continuous delivery, which automates building, testing, and deploying code changes. This automation helps you catch bugs sooner and increases developer productivity.
In this webinar, we’ll share the processes that Amazon engineers use to practice DevOps and discuss how you can bring these processes to your company by using a new set of AWS tools (AWS CodePipeline and AWS CodeDeploy). These services were inspired by Amazon's own internal developer tools and DevOps culture.
Learning Objectives:
• Learn what is continuous delivery, its benefits, and how to implement it
• Learn how to increase the frequency and reliability of your application updates
• Learn to create an automated software release workflow on AWS
• Understand the basics of AWS CodePipeline and AWS CodeDeploy
Continuous Delivery with AWS Lambda - AWS April 2016 Webinar SeriesAmazon Web Services
Managing the deployment of code to multiple AWS Lambda functions and updating your API Gateway methods can be manual and time consuming.
In this webinar, we will show you how to build a deployment pipeline to AWS Lambda using AWS CodePipeline. We will discuss how to use versioning, allowing you to better manage the different variations of your Lambda function and API Gateway methods in your development workflow, such as development, staging, and production. We will walk through how to automate the entire release process of your application from development to staging and finally to production, performing automated integration tests at each stage.
Learning Objectives:
Understand the basics of AWS CodePipeline
Learn how to version AWS Lambda functions and API Gateway methods
Build a deployment pipeline to AWS Lambda
- AWS OpsWorks for Chef Automate provides a fully managed Chef Automate server on AWS to help with infrastructure configuration management.
- It allows users to easily create an AWS managed Chef server in about 10 minutes to define infrastructure using code.
- The service handles backups, security updates, and Chef software updates automatically so users can focus on writing cookbooks and recipes.
AWS offers a number of services that help you easily deploy and run applications in the cloud. Come to this session to learn how to choose among these options. Through interactive demonstrations, this session will show you how to get an application running using AWS OpsWorks and AWS Elastic Beanstalk application management services. You will also learn how to use AWS CloudFormation templates to document, version control, and share your application configuration. This session will cover topics like application updates, customization and working with resources such as load balancers and databases.
This session is recommended for people who understand AWS and want to know more about deployment options for their applications.
DevOps combines cultural philosophies, practices, and tools to increase collaboration between development and operations teams. It aims to improve reliability, speed, and scale through practices like microservices, continuous integration and delivery, infrastructure as code, and monitoring. AWS provides services like CodeCommit, CodePipeline, CodeDeploy, CloudFormation, OpsWorks, Config, CloudWatch, and CloudTrail to help customers implement DevOps practices on AWS.
In this session, we will discuss the difficulties of running Docker in production and how Amazon EC2 Container Service (ECS) can be used to reduce the operational burdens. We will give an overview of the core architectural principles underlying Amazon ECS, and we will walk through a number of patterns used by our customers to run their microservices platforms and batch applications.
(DVO313) Building Next-Generation Applications with Amazon ECSAmazon Web Services
Two trends are driving app development: The shift from the server-based web to rich applications that run on a diverse set of mobile devices and modern browsers, and the growth of microservices running in the cloud that serve these clients. The results are “connected clients” - apps with the processing power of the device that are statefully connected and scaled to the cloud. In this session, you will learn about the architecture for Meteor's JavaScript app platform, Galaxy, which uses Amazon ECS, Elastic Load Balancing, and AWS CloudFormation to provide highly available, scalable, isolated environments for stateful apps across browsers and devices. We will discuss the essential characteristics of the platform, how those are provided for, and why we decided to use Amazon ECS instead of alternatives, such as Kubernetes. We will also demonstrate the Galaxy system in production.
This document provides an overview of automating software deployments with AWS CodeDeploy. It discusses the value of automation in reducing human errors during deployments. AWS CodeDeploy is a managed service that allows for deploying code to EC2 instances or on-premises servers in an efficient, scalable, and reliable manner. It works by having an agent installed on instances that coordinates lifecycle events defined in an AppSpec file during deployments. Deployments can be configured for different deployment groups using options like one-at-a-time or all-at-once. AWS CodeDeploy integrates with services like AWS CodeCommit, CodePipeline, and Auto Scaling to enable fully automated deployments.
(DVO305) Turbocharge YContinuous Deployment Pipeline with ContainersAmazon Web Services
This document outlines best practices for using containers in a continuous delivery pipeline. It recommends using containers with tools like Docker, Docker Compose, Amazon ECS, Jenkins, and AWS CodePipeline to build, test, and deploy applications. The workflow involves developing code in a source code repository, building Docker images, running tests inside containers, and deploying containers to production using Amazon ECS and AWS services for automation and orchestration of the pipeline. Demo applications and architectures are presented to illustrate container-based continuous delivery.
Best Practices with IoT Security - February Online Tech TalksAmazon Web Services
AWS IoT is a managed cloud platform that lets connected devices easily and securely interact with cloud applications and other devices. This tech talk will introduce the best practices for IoT Security in the cloud and the access control mechanisms used by AWS IoT. These mechanisms can be used to not only securely build and provision devices, but also to integrate devices with other AWS services to create secure solutions.
Learning Objectives:
• Learn common Internet of Things security issues
• Learn about AWS IoT security and access control mechanisms
• Learn how to build secure interactions with the AWS Cloud
AWS re:Invent 2016: Configuration Management in the Cloud (DEV305)Amazon Web Services
To ensure that your application operates in a predictable manner in both your test and production environments, you must vigilantly maintain the configuration of your resources. By leveraging configuration management solutions, Dev and Ops engineers can define the state of their resources across their entire lifecycle. In this session, we will show you how to use AWS OpsWorks, AWS CodeDeploy, and AWS CodePipeline to build a reliable and consistent development pipeline that assures your production workloads behave in a predictable manner.
(DVO308) Docker & ECS in Production: How We Migrated Our Infrastructure from ...Amazon Web Services
This session will introduce you to Empire, a new self-hosted PaaS built on top of Amazon’s EC2 Container Service (ECS). Empire is a recently open-sourced project that provides a mostly Heroku-compatible API. It allows engineering teams to deploy and manage applications in a method similar to Heroku, but with the added flexibility and control of running your own ECS container instances. We'll talk about why Remind decided to move its infrastructure from Heroku to AWS, introduce you to ECS and the open source platform we built on top of it to make migration easier, and then we'll demo Empire to show you how you can try it today.
Just as serverless application development is rapidly becoming the most popular way to bring highly scalable applications to the cloud, .NET has undergone radical changes with .NET Core to become a premier development platform for the cloud. In this session, you will learn how to use the newly launched C# support for .NET Core with AWS Lambda to create highly scalable serverless applications that target platforms from the traditional desktop to mobile devices. We will demonstrate how to write, test, and deploy C# code to AWS Lambda and see how we can leverage our serverless back end from mobile applications.
AWS is an elastic, secure, flexible, and developer-centric ecosystem that serves as an ideal platform for Docker deployments. AWS offers the scalable infrastructure, APIs, and SDKs that integrate tightly into a development lifecycle and accentuate the benefits of the lightweight and portable containers that Docker offers to its users. This session familiarizes you with the benefits of containers, introduce Amazon EC2 Container Service, and demonstrates how to use Amazon ECS to run containerized applications at scale in production.
In this session we'll discuss and demonstrate key concepts and design patterns for continuous deployment and integration using technologies like AWS OpsWorks and Chef to enable better control of applications and infrastructures.
This document provides an overview of CloudFormation best practices:
- It discusses organizing infrastructure using CloudFormation stacks by layers, environments, and services to promote reuse and decoupling.
- It recommends starting with existing templates, validation tools, parameter types, and IAM roles to prevent errors.
- Debugging tips include viewing stack events, using wait conditions, and logging to CloudWatch.
- Safe stack updates involve change sets to review impacts and choosing update styles for minimal disruption.
Amazon EC2 Systems Manager for Hybrid Cloud Management at ScaleAmazon Web Services
Amazon EC2 Systems Manager provides capabilities that enable automated configuration and ongoing management of systems at scale across Windows and Linux workloads running in Amazon EC2 or on-premises at no additional charge. It offers components like Run Command, State Manager, Inventory, Maintenance Windows, Patch Manager, Automation, and Parameter Store to remotely manage servers, define consistent configurations, gather inventory, schedule maintenance windows, automate patching, simplify deployments, and securely store parameters. Using these capabilities is expected to reduce the total cost of ownership for hybrid and cloud environments compared to traditional management tools.
The document discusses AWS Code services that can be used to automate the software release process. It describes CodeCommit for source control, CodeBuild for building and testing code, and CodeDeploy for deploying builds to EC2/on-premises servers. CodePipeline allows orchestrating builds and deployments across different environments through a visual workflow.
Configuration Management with AWS OpsWorks - November 2016 Webinar SeriesAmazon Web Services
To compete in today’s cloud-driven market, you need tools to help you automate bootstrapping, configuration, deployment, monitoring, scaling, recovery, and more. AWS OpsWorks helps you define, deploy, and operate your applications on AWS using Chef. AWS OpsWorks handles the heavy lifting for you, and it does it by using a single console. In this session, we will demonstrate how to bake automation and predictability into your application’s lifecycle with AWS OpsWorks, and showcase one of our latest collaborations with AWS CodePipeline that streamlines CI/CD processes.
Learning Objectives:
• Learn about AWS OpsWorks
• Learn about configuration management
• Learn about AWS CodePipeline
• Learn how to deploy applications using AWS CodePipeline and AWS OpsWorks
"Amazon Inspector is a new service from AWS that identifies security issues in your application deployments. Use Inspector with your applications to assess your security posture and identify areas that can be improved. Inspector works with your Amazon EC2 instances to monitor activity in your applications and system.
This session will cover getting started with Inspector, how to automate the process, how to manage and act on findings, and additional ways you can enhance your development and release lifecycle using Inspector."
Getting Started With Continuous Delivery on AWS - AWS April 2016 Webinar SeriesAmazon Web Services
Today’s cutting-edge companies have software release cycles measured in days instead of months. This agility is enabled by the DevOps practice of continuous delivery, which automates building, testing, and deploying code changes. This automation helps you catch bugs sooner and increases developer productivity.
In this webinar, we’ll share the processes that Amazon engineers use to practice DevOps and discuss how you can bring these processes to your company by using a new set of AWS tools (AWS CodePipeline and AWS CodeDeploy). These services were inspired by Amazon's own internal developer tools and DevOps culture.
Learning Objectives:
• Learn what is continuous delivery, its benefits, and how to implement it
• Learn how to increase the frequency and reliability of your application updates
• Learn to create an automated software release workflow on AWS
• Understand the basics of AWS CodePipeline and AWS CodeDeploy
Continuous Delivery with AWS Lambda - AWS April 2016 Webinar SeriesAmazon Web Services
Managing the deployment of code to multiple AWS Lambda functions and updating your API Gateway methods can be manual and time consuming.
In this webinar, we will show you how to build a deployment pipeline to AWS Lambda using AWS CodePipeline. We will discuss how to use versioning, allowing you to better manage the different variations of your Lambda function and API Gateway methods in your development workflow, such as development, staging, and production. We will walk through how to automate the entire release process of your application from development to staging and finally to production, performing automated integration tests at each stage.
Learning Objectives:
Understand the basics of AWS CodePipeline
Learn how to version AWS Lambda functions and API Gateway methods
Build a deployment pipeline to AWS Lambda
- AWS OpsWorks for Chef Automate provides a fully managed Chef Automate server on AWS to help with infrastructure configuration management.
- It allows users to easily create an AWS managed Chef server in about 10 minutes to define infrastructure using code.
- The service handles backups, security updates, and Chef software updates automatically so users can focus on writing cookbooks and recipes.
AWS offers a number of services that help you easily deploy and run applications in the cloud. Come to this session to learn how to choose among these options. Through interactive demonstrations, this session will show you how to get an application running using AWS OpsWorks and AWS Elastic Beanstalk application management services. You will also learn how to use AWS CloudFormation templates to document, version control, and share your application configuration. This session will cover topics like application updates, customization and working with resources such as load balancers and databases.
This session is recommended for people who understand AWS and want to know more about deployment options for their applications.
DevOps combines cultural philosophies, practices, and tools to increase collaboration between development and operations teams. It aims to improve reliability, speed, and scale through practices like microservices, continuous integration and delivery, infrastructure as code, and monitoring. AWS provides services like CodeCommit, CodePipeline, CodeDeploy, CloudFormation, OpsWorks, Config, CloudWatch, and CloudTrail to help customers implement DevOps practices on AWS.
In this session, we will discuss the difficulties of running Docker in production and how Amazon EC2 Container Service (ECS) can be used to reduce the operational burdens. We will give an overview of the core architectural principles underlying Amazon ECS, and we will walk through a number of patterns used by our customers to run their microservices platforms and batch applications.
(DVO313) Building Next-Generation Applications with Amazon ECSAmazon Web Services
Two trends are driving app development: The shift from the server-based web to rich applications that run on a diverse set of mobile devices and modern browsers, and the growth of microservices running in the cloud that serve these clients. The results are “connected clients” - apps with the processing power of the device that are statefully connected and scaled to the cloud. In this session, you will learn about the architecture for Meteor's JavaScript app platform, Galaxy, which uses Amazon ECS, Elastic Load Balancing, and AWS CloudFormation to provide highly available, scalable, isolated environments for stateful apps across browsers and devices. We will discuss the essential characteristics of the platform, how those are provided for, and why we decided to use Amazon ECS instead of alternatives, such as Kubernetes. We will also demonstrate the Galaxy system in production.
This document provides an overview of automating software deployments with AWS CodeDeploy. It discusses the value of automation in reducing human errors during deployments. AWS CodeDeploy is a managed service that allows for deploying code to EC2 instances or on-premises servers in an efficient, scalable, and reliable manner. It works by having an agent installed on instances that coordinates lifecycle events defined in an AppSpec file during deployments. Deployments can be configured for different deployment groups using options like one-at-a-time or all-at-once. AWS CodeDeploy integrates with services like AWS CodeCommit, CodePipeline, and Auto Scaling to enable fully automated deployments.
(DVO305) Turbocharge YContinuous Deployment Pipeline with ContainersAmazon Web Services
This document outlines best practices for using containers in a continuous delivery pipeline. It recommends using containers with tools like Docker, Docker Compose, Amazon ECS, Jenkins, and AWS CodePipeline to build, test, and deploy applications. The workflow involves developing code in a source code repository, building Docker images, running tests inside containers, and deploying containers to production using Amazon ECS and AWS services for automation and orchestration of the pipeline. Demo applications and architectures are presented to illustrate container-based continuous delivery.
Best Practices with IoT Security - February Online Tech TalksAmazon Web Services
AWS IoT is a managed cloud platform that lets connected devices easily and securely interact with cloud applications and other devices. This tech talk will introduce the best practices for IoT Security in the cloud and the access control mechanisms used by AWS IoT. These mechanisms can be used to not only securely build and provision devices, but also to integrate devices with other AWS services to create secure solutions.
Learning Objectives:
• Learn common Internet of Things security issues
• Learn about AWS IoT security and access control mechanisms
• Learn how to build secure interactions with the AWS Cloud
AWS re:Invent 2016: Configuration Management in the Cloud (DEV305)Amazon Web Services
To ensure that your application operates in a predictable manner in both your test and production environments, you must vigilantly maintain the configuration of your resources. By leveraging configuration management solutions, Dev and Ops engineers can define the state of their resources across their entire lifecycle. In this session, we will show you how to use AWS OpsWorks, AWS CodeDeploy, and AWS CodePipeline to build a reliable and consistent development pipeline that assures your production workloads behave in a predictable manner.
(DVO308) Docker & ECS in Production: How We Migrated Our Infrastructure from ...Amazon Web Services
This session will introduce you to Empire, a new self-hosted PaaS built on top of Amazon’s EC2 Container Service (ECS). Empire is a recently open-sourced project that provides a mostly Heroku-compatible API. It allows engineering teams to deploy and manage applications in a method similar to Heroku, but with the added flexibility and control of running your own ECS container instances. We'll talk about why Remind decided to move its infrastructure from Heroku to AWS, introduce you to ECS and the open source platform we built on top of it to make migration easier, and then we'll demo Empire to show you how you can try it today.
Just as serverless application development is rapidly becoming the most popular way to bring highly scalable applications to the cloud, .NET has undergone radical changes with .NET Core to become a premier development platform for the cloud. In this session, you will learn how to use the newly launched C# support for .NET Core with AWS Lambda to create highly scalable serverless applications that target platforms from the traditional desktop to mobile devices. We will demonstrate how to write, test, and deploy C# code to AWS Lambda and see how we can leverage our serverless back end from mobile applications.
AWS is an elastic, secure, flexible, and developer-centric ecosystem that serves as an ideal platform for Docker deployments. AWS offers the scalable infrastructure, APIs, and SDKs that integrate tightly into a development lifecycle and accentuate the benefits of the lightweight and portable containers that Docker offers to its users. This session familiarizes you with the benefits of containers, introduce Amazon EC2 Container Service, and demonstrates how to use Amazon ECS to run containerized applications at scale in production.
In this session we'll discuss and demonstrate key concepts and design patterns for continuous deployment and integration using technologies like AWS OpsWorks and Chef to enable better control of applications and infrastructures.
This document provides an overview of CloudFormation best practices:
- It discusses organizing infrastructure using CloudFormation stacks by layers, environments, and services to promote reuse and decoupling.
- It recommends starting with existing templates, validation tools, parameter types, and IAM roles to prevent errors.
- Debugging tips include viewing stack events, using wait conditions, and logging to CloudWatch.
- Safe stack updates involve change sets to review impacts and choosing update styles for minimal disruption.
Amazon EC2 Systems Manager for Hybrid Cloud Management at ScaleAmazon Web Services
Amazon EC2 Systems Manager provides capabilities that enable automated configuration and ongoing management of systems at scale across Windows and Linux workloads running in Amazon EC2 or on-premises at no additional charge. It offers components like Run Command, State Manager, Inventory, Maintenance Windows, Patch Manager, Automation, and Parameter Store to remotely manage servers, define consistent configurations, gather inventory, schedule maintenance windows, automate patching, simplify deployments, and securely store parameters. Using these capabilities is expected to reduce the total cost of ownership for hybrid and cloud environments compared to traditional management tools.
Migrate your Data Warehouse to Amazon Redshift - September Webinar SeriesAmazon Web Services
- TrueCar migrated their data warehouse from an on-premises Hadoop cluster to Amazon Redshift. They load clickstream, transactions, inventory, and lead data into Redshift for analytics and reporting.
- They use ETL tools like Talend and Hive to process data and load it into HDFS and S3, then load it into Redshift using a custom utility. The data is organized into schemas separating raw, user, and reporting data.
- Best practices for Redshift include designing tables for compression, sort keys, and distribution, managing cluster size and workloads over time, and vacuuming and analyzing tables regularly. TrueCar's migration to Redshift improved performance and reduced costs.
AWS Webcast - Archiving in the Cloud - Best Practices for Amazon GlacierAmazon Web Services
Join our webinar to learn more about how to build a cost effective archive application using Amazon Glacier, an extremely low cost, secure, highly durable, and easy to use storage service in the AWS cloud.
We will explain how Amazon Glacier works and walk through some best practices to get the most out of the service
We will also highlight how to choose between Amazon Glacier and Amazon S3’s Glacier storage option.
Learn more: http://aws.amazon.com/glacier/
Amazon Redshift is a fully managed data warehouse service that makes it fast, simple and cost effective to analyze data using SQL and existing business intelligence tools. The document provides an overview of Amazon Redshift and its benefits including speed, low cost, security, scalability and ease of use. It also provides examples of how various companies use Redshift for big data analytics including analyzing social media firehoses, mobile usage and real-time IoT streaming data.
An overview of the Amazon ElastiCache managed service, with examples of how it can be used to increase performance, lower costs and augment other database services and databases to make things faster, easier and less expensive.
The document provides an introduction to Amazon DynamoDB, a fully managed NoSQL database service. It discusses how DynamoDB provides fast and consistent performance at scale without the need to provision or manage infrastructure. It also demonstrates how to build a serverless web application using DynamoDB along with AWS Lambda and API Gateway.
(STG202) AWS Import/Export Snowball: Large-Scale Data Ingest into AWSAmazon Web Services
Moving terabyte and petabyte volumes of data into the cloud can be a challenge for many businesses. Come learn how you can use Snowball, a new AWS feature, to move large-scale (terabyte and petabyte) data to AWS storage services.
AWS re:Invent 2016: ElastiCache Deep Dive: Best Practices and Usage Patterns ...Amazon Web Services
In this session, we provide a peek behind the scenes to learn about Amazon ElastiCache's design and architecture. See common design patterns with our Redis and Memcached offerings and how customers have used them for in-memory operations to reduce latency and improve application throughput. During this session, we review ElastiCache best practices, design patterns, and anti-patterns.
Not just for archiving or compliance use cases, Amazon Glacier accommodates customers simply looking to replace their on-premises long term storage with a cost efficient, durable, cloud option, from which they can easily and quickly access their data when they need to. This session will introduce newly launched features for Amazon Glacier, review the current service feature set, and share the global data center shut down and storage strategy for Sony DADC New Media Solutions (NMS). NMS is Sony’s digital servicing division providing global digital distribution, linear playout and white label OTT/Commerce solutions for clients such as BBC Worldwide, NBCUniversal, Sony Playstation, and Funimation Entertainment.
Hear from Andy Shenkler, NMS’s Chief Technology and Solutions Officer as he talks about the key factors that drove the organization’s decision to move away from tape and go towards the cloud and out of the infrastructure business overall. Learn more about the impact and operational practices inside a world class digital supply chain as they were able to move over 20 petabytes of data, over 1M hours of video, to the cloud and never looked back.
AWS re:Invent 2016: Migrating Your Data Warehouse to Amazon Redshift (DAT202)Amazon Web Services
Amazon Redshift is a fast, simple, cost-effective data warehousing solution, and in this session, we look at the tools and techniques you can use to migrate your existing data warehouse to Amazon Redshift. We will then present a case study on Scholastic’s migration to Amazon Redshift. Scholastic, a large 100-year-old publishing company, was running their business with older, on-premise, data warehousing and analytics solutions, which could not keep up with business needs and were expensive. Scholastic also needed to include new capabilities like streaming data and real time analytics. Scholastic migrated to Amazon Redshift, and achieved agility and faster time to insight while dramatically reducing costs. In this session, Scholastic will discuss how they achieved this, including options considered, technical architecture implemented, results, and lessons learned.
Best Practices for Managing Security Operations in AWS - March 2017 AWS Onlin...Amazon Web Services
To help prevent unexpected access to your AWS resources, it is critical to maintain strong identity and access policies. It is equally important to track and alert on changes to your AWS resources. In this tech talk, you will learn how to use AWS Identity and Access Management (IAM) to control access to your AWS resources and integrate your existing authentication system with AWS IAM. We will cover how you can deploy and control your AWS infrastructure using code templates, including change management policies with AWS CloudFormation. In addition, we will explore different options for managing both your AWS access logs and your Amazon Elastic Compute Cloud (EC2) system logs using Amazon CloudWatch Logs. We also will cover how to use these logs to implement an audit and compliance validation process using services such as AWS Config, AWS CloudTrail, and Amazon Inspector.
Learning Objectives:
• Understand the AWS Shared Responsibility Model.
• Understand AWS account and identity management options and configuration.
• Learn the concept of infrastructure as code and change management using AWS CloudFormation.
• Learn how to audit and log your AWS service usage.
• Learn about AWS services to add automatic compliance checks to your AWS infrastructure.
Hands-on Labs: Getting Started with AWS - March 2017 AWS Online Tech TalksAmazon Web Services
The document provides information about a webinar on getting started with AWS, including deploying a static website. It outlines the agenda which includes: watching a 15 minute presentation on AWS; watching a 25 minute demo of deploying a static website; and having 45-60 minutes to complete the demo independently. It then details the various sections of the webinar which cover creating an AWS account, enabling security features, using S3 buckets to host the website, configuring permissions, associating a domain name, and using CloudFront for acceleration.
Amazon Elastic Block Store (Amazon EBS) provides persistent block level storage volumes for use with Amazon EC2 instances. In this technical session, we conduct a detailed analysis of the differences among the three types of Amazon EBS block storage: General Purpose (SSD), Provisioned IOPS (SSD), and Magnetic. We discuss how to maximize Amazon EBS performance, with a special eye towards low-latency, high-throughput applications like databases. We discuss Amazon EBS encryption and share best practices for Amazon EBS snapshot management. Throughout, we share tips for success.
AWS re:Invent 2016: Deep Dive on Amazon DynamoDB (DAT304)Amazon Web Services
Explore Amazon DynamoDB capabilities and benefits in detail and learn how to get the most out of your DynamoDB database. We go over best practices for schema design with DynamoDB across multiple use cases, including gaming, AdTech, IoT, and others. We explore designing efficient indexes, scanning, and querying, and go into detail on a number of recently released features, including JSON document support, DynamoDB Streams, and more. We also provide lessons learned from operating DynamoDB at scale, including provisioning DynamoDB for IoT.
(STG312) Amazon Glacier Deep Dive: Cold Data Storage in AWSAmazon Web Services
This session explores some of the key features of Amazon Glacier, including security, durability, and configuration for storing compliance and regulatory data. It covers best practices for managing your cold data, including ingest, retrieval, and security controls. Other topics include: how to optimize storage, upload, and retrieval costs; how to identify the most applicable workloads; and recommended optimizations based on a few sample use cases from a number of industry verticals.
AWS Data Transfer Services - AWS Gateway, AWS Snowball, AWS Snowball Edge, an...Amazon Web Services
by Everett Dolgner, Business Development Manager, AWS
AWS offers a suite of tools to help you surmount limitations associated to data migration from on premise to the cloud. Attend this session to learn about moving data by using networks, roads, and AWS technology partners. We will also discuss how to move data into and out of the Cloud in batches, increments, and streams.
This document provides an overview and agenda for a presentation on Amazon DynamoDB. It discusses key concepts like tables, data types, partitioning, indexing and scaling in DynamoDB. It also provides best practices and examples for modeling different data scenarios like event logging, product catalogs, messaging apps and multiplayer games.
Accelerating Application Performance with Amazon ElastiCache (DAT207) | AWS r...Amazon Web Services
Learn how you can use Amazon ElastiCache to easily deploy a Memcached or Redis compatible, in-memory caching system to speed up your application performance. We show you how to use Amazon ElastiCache to improve your application latency and reduce the load on your database servers. We'll also show you how to build a caching layer that is easy to manage and scale as your application grows. During this session, we go over various scenarios and use cases that can benefit by enabling caching, and discuss the features provided by Amazon ElastiCache.
Configuration Management with AWS OpsWorks for Chef AutomateAmazon Web Services
AWS OpsWorks for Chef Automate provides a fully managed Chef server and suite of automation tools that give you workflow automation for continuous deployment, automated testing for compliance and security, and a user interface that gives you visibility into your nodes and their status. The Chef server gives you full stack automation by handling operational tasks such as software and operating system configurations, package installations, database setups, and more. The Chef server centrally stores your configuration tasks and provides them to each node in your compute environment at any scale, from a few nodes to thousands of nodes. OpsWorks for Chef Automate is completely compatible with tooling and cookbooks from the Chef community and automatically registers new nodes with your Chef server.
Announcing AWS OpsWorks for Chef Automate - January 2017 AWS Online Tech TalksAmazon Web Services
AWS OpsWorks for Chef Automate provides a fully managed Chef server and suite of automation tools that give you workflow automation for continuous deployment, automated testing for compliance and security, and a user interface that gives you visibility into your nodes and their status.
Learning Objectives:
• Learn about the capabilities, features and benefits of AWS OpsWorks for Chef Automate
• Learn how you can automate configuration management using AWS OpsWorks for Chef Automate
• Learn how to get started using AWS OpsWorks for Chef Automate
Configuration Management in the Cloud - AWS Online Tech TalksAmazon Web Services
Learning Objectives:
- Learn how to use AWS OpsWorks, AWS CodeDeploy, and AWS CodePipeline to build a reliable and consistent development pipeline
- Understand about continous integration and delivery for Infrastructure as Code
- Learn how to get started with these services.
This document discusses Chef, an open source infrastructure automation tool. It provides concise summaries in 3 sentences or less:
Chef is a systems and cloud infrastructure automation framework that makes it easy to deploy servers and applications to any physical, virtual, or cloud location. It uses code and templates to abstractly define how infrastructure should be configured. Chef can be used to configure single machines or entire infrastructures for provisioning, configuration, and integration tasks.
Mohit Sethi gives a presentation on Chef, an automation and configuration management tool. He defines Chef as a systems integration framework that brings configuration management benefits to infrastructure. Chef allows users to define what state servers should be in and enforces that state. Key principles of Chef include idempotence, provisioning often, treating infrastructure as code, being data-driven, and having thick clients and a thin server.
Chef is a systems integration framework that allows you to define the state that your servers should be in and enforce that state. It provides architecture where Chef clients run on servers and talk to a central Chef server. Key principles of Chef include idempotence, provisioning often, treating infrastructure as code, being data-driven, and having thick clients and a thin server. Chef uses resources, providers, recipes, roles, cookbooks, attributes, and data bags to automate server configuration and management.
This document provides instructions for setting up a Chef environment, including installing a Chef server, configuring a workstation, and registering a node. It discusses the basics of Chef and its architecture involving workstations, nodes, and a server managed through Knife. Administrators can opt for a hosted or on-premises Chef server. The workstation is configured using Knife and keys are used to authenticate nodes which run Chef client.
Overview of Chef - Fundamentals Webinar Series Part 1Chef
This is an Overview of Chef. After viewing this webinar you will be able to:
- Describe how Chef thinks about Infrastructure Automation
- Define the following terms:
- Resource
- Recipe
- Node
- Run List
- Search
- Login to Hosted Chef
- Run `knife` commands from your workstation
Video of this webinar can be found at the following URL
https://www.youtube.com/watch?v=S5lHUpzoCYo&list=PL11cZfNdwNyPnZA9D1MbVqldGuOWqbumZ
Chef Fundamentals Training Series Module 1: Overview of ChefChef Software, Inc.
This document provides an overview of Chef fundamentals. It introduces Nathen Harvey as the presenter and outlines objectives to teach attendees how to automate infrastructure tasks with Chef. Key concepts discussed include Chef's architecture, tools, and how to apply its primitives to solve problems. The document explains that learning Chef is like learning a language and emphasizes using Chef to learn it. It provides an agenda covering topics like workstation setup, the node object, cookbooks, and using community cookbooks.
DOO-009_Powering High Velocity Development for your Infrastructuredecode2016
Chef is a leader in automation for DevOps that provides a platform to automate infrastructure and applications. The Chef platform allows teams to treat infrastructure as code by defining policies and configurations in code that can be versioned, tested, and deployed. This enables teams to collaborate safely at a high velocity. The Chef platform includes tools for test-driven infrastructure development, integration with cloud platforms like Azure, and ensuring compliance and security.
This document discusses simplifying, standardizing, and automating application deployment processes before moving to the cloud. It recommends using central configuration repositories and automation tools like Chef to deploy identical environments for development, staging, and production. This allows using the same processes and tools across environments. AWS services like OpsWorks can then be used to deploy production using the same Chef configurations. The key is treating the cloud as a tool to deploy standardized, automated applications at scale.
Chef is an automation platform that transforms infrastructure into code. It uses recipes written in Ruby and Erlang languages to configure, deploy, and manage applications across networks. Chef includes a server to store configuration data and recipes, workstations where developers write recipes, and nodes (physical or virtual machines) that are configured by recipes. Key components of Chef include cookbooks (which contain recipes, attributes, files, and templates), nodes, Ohai (which collects node data), and a workflow involving verifying, building, accepting, and delivering changes through shared pipelines.
Chef is a configuration management tool that turns infrastructure into code. It allows automating how systems are built, deployed, and managed. With Chef, infrastructure is versioned, tested, and repeatable like application code. The document provides an overview of key Chef concepts including the Chef server, nodes, organizations, environments, cookbooks, roles, and data bags. It also describes the basic Chef environment and components like the workstation, Chef client, and knife tool.
Opscode Webinar: Managing Your VMware Infrastructure with ChefChef Software, Inc.
This document provides an overview of how Chef can be used to manage VMware infrastructure. It discusses four main integration points: 1) VMware Fusion/Workstation and Vagrant to provision development VMs locally, 2) knife-esx to manage individual ESXi hosts, 3) knife-vsphere to manage vCenter and provision/configure VMs, and 4) knife-vcloud to manage vCloud Director and deploy vApps. The document emphasizes that Chef allows infrastructure to be defined as code through recipes and cookbooks rather than using VM templates, making infrastructure more flexible and standardized. It concludes with a demo of Vagrant/Fusion and knife-vsphere.
Introduction to Chef: Automate Your Infrastructure by Modeling It In CodeJosh Padnick
Presentation by Josh Padnick given at Desert Code Camp on April 5, 2014. Introduces OpsCode Chef with a special emphasis on learning the key Chef concepts. Also includes tips & tricks and references to best practices.
Chef Fundamentals Training Series Module 2: Workstation SetupChef Software, Inc.
This document provides instructions for setting up a workstation to manage infrastructure with Chef. It covers installing Chef, creating an account on the hosted Chef server, downloading the starter kit which contains files like cookbooks and roles, and configuring the knife command line tool to connect to the Chef server. The document also gives an overview of the components that make up a Chef-managed infrastructure including nodes, roles, environments and data bags.
This document discusses using Chef with AWS. It provides an overview of key points on Chef including its execution methods and use of Chef Server for bootstrapping nodes. It describes using Chef for continuous delivery processes on AWS and outlines architectures for high availability Chef implementations on AWS. The document concludes with links to additional Chef resources and a demo of dependency management with Chef Server on AWS.
Configuration Management in the Cloud | AWS Public Sector Summit 2017Amazon Web Services
In order for your application to operate in a predictable manner in both your test and production environments, you must vigilantly maintain the configuration of your resources. By leveraging configuration management solutions, Dev and Ops engineers can define the state of their resources across their entire lifecycle. In this session, you will learn how to use AWS OpsWorks, AWS CodeDeploy, and AWS CodePipeline to build a reliable and consistent development pipeline that assures your production workloads behave in a predictable manner. Learn More: https://aws.amazon.com/government-education/
Habitat-managed Chef with Policyfiles: Learn how to leverage the power of Habitat, chef-client and Policyfiles to produce an immutable application containing all of your chef cookbooks that can be locally tested and provides a consistent and guaranteed picture of desired configuration state across all target environments.
This document provides an introduction to using Chef for infrastructure automation and configuration management. It discusses what Chef is, why it is used, and its core components like recipes, resources, attributes, cookbooks, roles, environments, and more. It also covers how to set up a development environment for Chef, write recipes, and test Chef configurations using tools like Chefspec, Foodcritic, and Test Kitchen with Serverspec. The document aims to help readers understand Chef and be able to use it to define reusable infrastructure configurations.
Come costruire servizi di Forecasting sfruttando algoritmi di ML e deep learn...Amazon Web Services
Il Forecasting è un processo importante per tantissime aziende e viene utilizzato in vari ambiti per cercare di prevedere in modo accurato la crescita e distribuzione di un prodotto, l’utilizzo delle risorse necessarie nelle linee produttive, presentazioni finanziarie e tanto altro. Amazon utilizza delle tecniche avanzate di forecasting, in parte questi servizi sono stati messi a disposizione di tutti i clienti AWS.
In questa sessione illustreremo come pre-processare i dati che contengono una componente temporale e successivamente utilizzare un algoritmo che a partire dal tipo di dato analizzato produce un forecasting accurato.
Big Data per le Startup: come creare applicazioni Big Data in modalità Server...Amazon Web Services
La varietà e la quantità di dati che si crea ogni giorno accelera sempre più velocemente e rappresenta una opportunità irripetibile per innovare e creare nuove startup.
Tuttavia gestire grandi quantità di dati può apparire complesso: creare cluster Big Data su larga scala sembra essere un investimento accessibile solo ad aziende consolidate. Ma l’elasticità del Cloud e, in particolare, i servizi Serverless ci permettono di rompere questi limiti.
Vediamo quindi come è possibile sviluppare applicazioni Big Data rapidamente, senza preoccuparci dell’infrastruttura, ma dedicando tutte le risorse allo sviluppo delle nostre le nostre idee per creare prodotti innovativi.
Ora puoi utilizzare Amazon Elastic Kubernetes Service (EKS) per eseguire pod Kubernetes su AWS Fargate, il motore di elaborazione serverless creato per container su AWS. Questo rende più semplice che mai costruire ed eseguire le tue applicazioni Kubernetes nel cloud AWS.In questa sessione presenteremo le caratteristiche principali del servizio e come distribuire la tua applicazione in pochi passaggi
Vent'anni fa Amazon ha attraversato una trasformazione radicale con l'obiettivo di aumentare il ritmo dell'innovazione. In questo periodo abbiamo imparato come cambiare il nostro approccio allo sviluppo delle applicazioni ci ha permesso di aumentare notevolmente l'agilità, la velocità di rilascio e, in definitiva, ci ha consentito di creare applicazioni più affidabili e scalabili. In questa sessione illustreremo come definiamo le applicazioni moderne e come la creazione di app moderne influisce non solo sull'architettura dell'applicazione, ma sulla struttura organizzativa, sulle pipeline di rilascio dello sviluppo e persino sul modello operativo. Descriveremo anche approcci comuni alla modernizzazione, compreso l'approccio utilizzato dalla stessa Amazon.com.
Come spendere fino al 90% in meno con i container e le istanze spot Amazon Web Services
L’utilizzo dei container è in continua crescita.
Se correttamente disegnate, le applicazioni basate su Container sono molto spesso stateless e flessibili.
I servizi AWS ECS, EKS e Kubernetes su EC2 possono sfruttare le istanze Spot, portando ad un risparmio medio del 70% rispetto alle istanze On Demand. In questa sessione scopriremo insieme quali sono le caratteristiche delle istanze Spot e come possono essere utilizzate facilmente su AWS. Impareremo inoltre come Spreaker sfrutta le istanze spot per eseguire applicazioni di diverso tipo, in produzione, ad una frazione del costo on-demand!
In recent months, many customers have been asking us the question – how to monetise Open APIs, simplify Fintech integrations and accelerate adoption of various Open Banking business models. Therefore, AWS and FinConecta would like to invite you to Open Finance marketplace presentation on October 20th.
Event Agenda :
Open banking so far (short recap)
• PSD2, OB UK, OB Australia, OB LATAM, OB Israel
Intro to Open Finance marketplace
• Scope
• Features
• Tech overview and Demo
The role of the Cloud
The Future of APIs
• Complying with regulation
• Monetizing data / APIs
• Business models
• Time to market
One platform for all: a Strategic approach
Q&A
Rendi unica l’offerta della tua startup sul mercato con i servizi Machine Lea...Amazon Web Services
Per creare valore e costruire una propria offerta differenziante e riconoscibile, le startup di successo sanno come combinare tecnologie consolidate con componenti innovativi creati ad hoc.
AWS fornisce servizi pronti all'utilizzo e, allo stesso tempo, permette di personalizzare e creare gli elementi differenzianti della propria offerta.
Concentrandoci sulle tecnologie di Machine Learning, vedremo come selezionare i servizi di intelligenza artificiale offerti da AWS e, anche attraverso una demo, come costruire modelli di Machine Learning personalizzati utilizzando SageMaker Studio.
OpsWorks Configuration Management: automatizza la gestione e i deployment del...Amazon Web Services
Con l'approccio tradizionale al mondo IT per molti anni è stato difficile implementare tecniche di DevOps, che finora spesso hanno previsto attività manuali portando di tanto in tanto a dei downtime degli applicativi interrompendo l'operatività dell'utente. Con l'avvento del cloud, le tecniche di DevOps sono ormai a portata di tutti a basso costo per qualsiasi genere di workload, garantendo maggiore affidabilità del sistema e risultando in dei significativi miglioramenti della business continuity.
AWS mette a disposizione AWS OpsWork come strumento di Configuration Management che mira ad automatizzare e semplificare la gestione e i deployment delle istanze EC2 per mezzo di workload Chef e Puppet.
Scopri come sfruttare AWS OpsWork a garanzia e affidabilità del tuo applicativo installato su Instanze EC2.
Microsoft Active Directory su AWS per supportare i tuoi Windows WorkloadsAmazon Web Services
Vuoi conoscere le opzioni per eseguire Microsoft Active Directory su AWS? Quando si spostano carichi di lavoro Microsoft in AWS, è importante considerare come distribuire Microsoft Active Directory per supportare la gestione, l'autenticazione e l'autorizzazione dei criteri di gruppo. In questa sessione, discuteremo le opzioni per la distribuzione di Microsoft Active Directory su AWS, incluso AWS Directory Service per Microsoft Active Directory e la distribuzione di Active Directory su Windows su Amazon Elastic Compute Cloud (Amazon EC2). Trattiamo argomenti quali l'integrazione del tuo ambiente Microsoft Active Directory locale nel cloud e l'utilizzo di applicazioni SaaS, come Office 365, con AWS Single Sign-On.
Dal riconoscimento facciale al riconoscimento di frodi o difetti di fabbricazione, l'analisi di immagini e video che sfruttano tecniche di intelligenza artificiale, si stanno evolvendo e raffinando a ritmi elevati. In questo webinar esploreremo le possibilità messe a disposizione dai servizi AWS per applicare lo stato dell'arte delle tecniche di computer vision a scenari reali.
Amazon Web Services e VMware organizzano un evento virtuale gratuito il prossimo mercoledì 14 Ottobre dalle 12:00 alle 13:00 dedicato a VMware Cloud ™ on AWS, il servizio on demand che consente di eseguire applicazioni in ambienti cloud basati su VMware vSphere® e di accedere ad una vasta gamma di servizi AWS, sfruttando a pieno le potenzialità del cloud AWS e tutelando gli investimenti VMware esistenti.
Molte organizzazioni sfruttano i vantaggi del cloud migrando i propri carichi di lavoro Oracle e assicurandosi notevoli vantaggi in termini di agilità ed efficienza dei costi.
La migrazione di questi carichi di lavoro, può creare complessità durante la modernizzazione e il refactoring delle applicazioni e a questo si possono aggiungere rischi di prestazione che possono essere introdotti quando si spostano le applicazioni dai data center locali.
Crea la tua prima serverless ledger-based app con QLDB e NodeJSAmazon Web Services
Molte aziende oggi, costruiscono applicazioni con funzionalità di tipo ledger ad esempio per verificare lo storico di accrediti o addebiti nelle transazioni bancarie o ancora per tenere traccia del flusso supply chain dei propri prodotti.
Alla base di queste soluzioni ci sono i database ledger che permettono di avere un log delle transazioni trasparente, immutabile e crittograficamente verificabile, ma sono strumenti complessi e onerosi da gestire.
Amazon QLDB elimina la necessità di costruire sistemi personalizzati e complessi fornendo un database ledger serverless completamente gestito.
In questa sessione scopriremo come realizzare un'applicazione serverless completa che utilizzi le funzionalità di QLDB.
Con l’ascesa delle architetture di microservizi e delle ricche applicazioni mobili e Web, le API sono più importanti che mai per offrire agli utenti finali una user experience eccezionale. In questa sessione impareremo come affrontare le moderne sfide di progettazione delle API con GraphQL, un linguaggio di query API open source utilizzato da Facebook, Amazon e altro e come utilizzare AWS AppSync, un servizio GraphQL serverless gestito su AWS. Approfondiremo diversi scenari, comprendendo come AppSync può aiutare a risolvere questi casi d’uso creando API moderne con funzionalità di aggiornamento dati in tempo reale e offline.
Inoltre, impareremo come Sky Italia utilizza AWS AppSync per fornire aggiornamenti sportivi in tempo reale agli utenti del proprio portale web.
Database Oracle e VMware Cloud™ on AWS: i miti da sfatareAmazon Web Services
Molte organizzazioni sfruttano i vantaggi del cloud migrando i propri carichi di lavoro Oracle e assicurandosi notevoli vantaggi in termini di agilità ed efficienza dei costi.
La migrazione di questi carichi di lavoro, può creare complessità durante la modernizzazione e il refactoring delle applicazioni e a questo si possono aggiungere rischi di prestazione che possono essere introdotti quando si spostano le applicazioni dai data center locali.
In queste slide, gli esperti AWS e VMware presentano semplici e pratici accorgimenti per facilitare e semplificare la migrazione dei carichi di lavoro Oracle accelerando la trasformazione verso il cloud, approfondiranno l’architettura e dimostreranno come sfruttare a pieno le potenzialità di VMware Cloud ™ on AWS.
1) The document discusses building a minimum viable product (MVP) using Amazon Web Services (AWS).
2) It provides an example of an MVP for an omni-channel messenger platform that was built from 2017 to connect ecommerce stores to customers via web chat, Facebook Messenger, WhatsApp, and other channels.
3) The founder discusses how they started with an MVP in 2017 with 200 ecommerce stores in Hong Kong and Taiwan, and have since expanded to over 5000 clients across Southeast Asia using AWS for scaling.
This document discusses pitch decks and fundraising materials. It explains that venture capitalists will typically spend only 3 minutes and 44 seconds reviewing a pitch deck. Therefore, the deck needs to tell a compelling story to grab their attention. It also provides tips on tailoring different types of decks for different purposes, such as creating a concise 1-2 page teaser, a presentation deck for pitching in-person, and a more detailed read-only or fundraising deck. The document stresses the importance of including key information like the problem, solution, product, traction, market size, plans, team, and ask.
This document discusses building serverless web applications using AWS services like API Gateway, Lambda, DynamoDB, S3 and Amplify. It provides an overview of each service and how they can work together to create a scalable, secure and cost-effective serverless application stack without having to manage servers or infrastructure. Key services covered include API Gateway for hosting APIs, Lambda for backend logic, DynamoDB for database needs, S3 for static content, and Amplify for frontend hosting and continuous deployment.
This document provides tips for fundraising from startup founders Roland Yau and Sze Lok Chan. It discusses generating competition to create urgency for investors, fundraising in parallel rather than sequentially, having a clear fundraising narrative focused on what you do and why it's compelling, and prioritizing relationships with people over firms. It also notes how the pandemic has changed fundraising, with examples of deals done virtually during this time. The tips emphasize being fully prepared before fundraising and cultivating connections with investors in advance.
AWS_HK_StartupDay_Building Interactive websites while automating for efficien...Amazon Web Services
This document discusses Amazon's machine learning services for building conversational interfaces and extracting insights from unstructured text and audio. It describes Amazon Lex for creating chatbots, Amazon Comprehend for natural language processing tasks like entity extraction and sentiment analysis, and how they can be used together for applications like intelligent call centers and content analysis. Pre-trained APIs simplify adding machine learning to apps without requiring ML expertise.
Amazon Elastic Container Service (Amazon ECS) è un servizio di gestione dei container altamente scalabile, che semplifica la gestione dei contenitori Docker attraverso un layer di orchestrazione per il controllo del deployment e del relativo lifecycle. In questa sessione presenteremo le principali caratteristiche del servizio, le architetture di riferimento per i differenti carichi di lavoro e i semplici passi necessari per poter velocemente migrare uno o più dei tuo container.
This presentation by OECD, OECD Secretariat, was made during the discussion “The Intersection between Competition and Data Privacy” held at the 143rd meeting of the OECD Competition Committee on 13 June 2024. More papers and presentations on the topic can be found at oe.cd/ibcdp.
This presentation was uploaded with the author’s consent.
This presentation by OECD, OECD Secretariat, was made during the discussion “Competition and Regulation in Professions and Occupations” held at the 77th meeting of the OECD Working Party No. 2 on Competition and Regulation on 10 June 2024. More papers and presentations on the topic can be found at oe.cd/crps.
This presentation was uploaded with the author’s consent.
This presentation by Nathaniel Lane, Associate Professor in Economics at Oxford University, was made during the discussion “Pro-competitive Industrial Policy” held at the 143rd meeting of the OECD Competition Committee on 12 June 2024. More papers and presentations on the topic can be found at oe.cd/pcip.
This presentation was uploaded with the author’s consent.
Collapsing Narratives: Exploring Non-Linearity • a micro report by Rosie WellsRosie Wells
Insight: In a landscape where traditional narrative structures are giving way to fragmented and non-linear forms of storytelling, there lies immense potential for creativity and exploration.
'Collapsing Narratives: Exploring Non-Linearity' is a micro report from Rosie Wells.
Rosie Wells is an Arts & Cultural Strategist uniquely positioned at the intersection of grassroots and mainstream storytelling.
Their work is focused on developing meaningful and lasting connections that can drive social change.
Please download this presentation to enjoy the hyperlinks!
This presentation by OECD, OECD Secretariat, was made during the discussion “Pro-competitive Industrial Policy” held at the 143rd meeting of the OECD Competition Committee on 12 June 2024. More papers and presentations on the topic can be found at oe.cd/pcip.
This presentation was uploaded with the author’s consent.
This presentation by Yong Lim, Professor of Economic Law at Seoul National University School of Law, was made during the discussion “Artificial Intelligence, Data and Competition” held at the 143rd meeting of the OECD Competition Committee on 12 June 2024. More papers and presentations on the topic can be found at oe.cd/aicomp.
This presentation was uploaded with the author’s consent.
Carrer goals.pptx and their importance in real lifeartemacademy2
Career goals serve as a roadmap for individuals, guiding them toward achieving long-term professional aspirations and personal fulfillment. Establishing clear career goals enables professionals to focus their efforts on developing specific skills, gaining relevant experience, and making strategic decisions that align with their desired career trajectory. By setting both short-term and long-term objectives, individuals can systematically track their progress, make necessary adjustments, and stay motivated. Short-term goals often include acquiring new qualifications, mastering particular competencies, or securing a specific role, while long-term goals might encompass reaching executive positions, becoming industry experts, or launching entrepreneurial ventures.
Moreover, having well-defined career goals fosters a sense of purpose and direction, enhancing job satisfaction and overall productivity. It encourages continuous learning and adaptation, as professionals remain attuned to industry trends and evolving job market demands. Career goals also facilitate better time management and resource allocation, as individuals prioritize tasks and opportunities that advance their professional growth. In addition, articulating career goals can aid in networking and mentorship, as it allows individuals to communicate their aspirations clearly to potential mentors, colleagues, and employers, thereby opening doors to valuable guidance and support. Ultimately, career goals are integral to personal and professional development, driving individuals toward sustained success and fulfillment in their chosen fields.
Suzanne Lagerweij - Influence Without Power - Why Empathy is Your Best Friend...Suzanne Lagerweij
This is a workshop about communication and collaboration. We will experience how we can analyze the reasons for resistance to change (exercise 1) and practice how to improve our conversation style and be more in control and effective in the way we communicate (exercise 2).
This session will use Dave Gray’s Empathy Mapping, Argyris’ Ladder of Inference and The Four Rs from Agile Conversations (Squirrel and Fredrick).
Abstract:
Let’s talk about powerful conversations! We all know how to lead a constructive conversation, right? Then why is it so difficult to have those conversations with people at work, especially those in powerful positions that show resistance to change?
Learning to control and direct conversations takes understanding and practice.
We can combine our innate empathy with our analytical skills to gain a deeper understanding of complex situations at work. Join this session to learn how to prepare for difficult conversations and how to improve our agile conversations in order to be more influential without power. We will use Dave Gray’s Empathy Mapping, Argyris’ Ladder of Inference and The Four Rs from Agile Conversations (Squirrel and Fredrick).
In the session you will experience how preparing and reflecting on your conversation can help you be more influential at work. You will learn how to communicate more effectively with the people needed to achieve positive change. You will leave with a self-revised version of a difficult conversation and a practical model to use when you get back to work.
Come learn more on how to become a real influencer!
This presentation by Tim Capel, Director of the UK Information Commissioner’s Office Legal Service, was made during the discussion “The Intersection between Competition and Data Privacy” held at the 143rd meeting of the OECD Competition Committee on 13 June 2024. More papers and presentations on the topic can be found at oe.cd/ibcdp.
This presentation was uploaded with the author’s consent.
This presentation by Juraj Čorba, Chair of OECD Working Party on Artificial Intelligence Governance (AIGO), was made during the discussion “Artificial Intelligence, Data and Competition” held at the 143rd meeting of the OECD Competition Committee on 12 June 2024. More papers and presentations on the topic can be found at oe.cd/aicomp.
This presentation was uploaded with the author’s consent.
This presentation by Professor Alex Robson, Deputy Chair of Australia’s Productivity Commission, was made during the discussion “Competition and Regulation in Professions and Occupations” held at the 77th meeting of the OECD Working Party No. 2 on Competition and Regulation on 10 June 2024. More papers and presentations on the topic can be found at oe.cd/crps.
This presentation was uploaded with the author’s consent.
This presentation by Professor Giuseppe Colangelo, Jean Monnet Professor of European Innovation Policy, was made during the discussion “The Intersection between Competition and Data Privacy” held at the 143rd meeting of the OECD Competition Committee on 13 June 2024. More papers and presentations on the topic can be found at oe.cd/ibcdp.
This presentation was uploaded with the author’s consent.
2. What to expect from this session?
Understand how configuration management lets you refer
to your infrastructure as code
Understand how AWS can help you use configuration
management to save time
Discover the best practices of setting up your
infrastructure, host configuration, and application
3. Background
Moving to the cloud and AWS allows you to provision and
manage infrastructure in new ways:
•Scale can be achieved without complicated capacity
planning
•Infrastructure can be provisioned in minutes
•You are now a part of a fast moving environment that
requires constant attention
4. What is configuration management?
A practice in which code is used to define and maintain the
state of both new and existing resources throughout their
entire life cycle.
5. Why do I need configuration management?
• Store your configuration information in one place
• Spin up blank resources that work perfectly every time
• Make changes things in a single place and propagate them
• Create dev and test environments that mimic your production
6. Compute Resources
Operating System and
Host Configuration
Application Configuration
Amazon Elastic Compute
Cloud (EC2)
On-premises compute
resources (Servers)
…
Files
Directories
Networking
Symlinks
Mounts
Registry Key
Users
Groups
Packages
Filesystems
…
Application dependencies
Application configuration
Service registration
Credentials
…
7. Infrastructure needs ongoing management
• Package updates?
• New software?
• New configurations?
• New app deployments?
• Environment specific changes?
• Run commands across all hosts?
• Be on top of all running resources?
8. Ongoing management requires proper tooling
Some common challenges:
• Changing a vhost configuration on every web server across
multiple environments (dev, stage, prod)
• Installing a package on certain hosts to test out newer versions
• Changing LDAP config on every running Amazon EC2 Linux host
What tools can I use to tackle some of these challenges?
9.
10. What is Chef Automate?
• Refer to your infrastructure as code (cookbooks & recipes)
• Consistently install, configure, manage, deploy and scale
applications
• Align resources with specific policies
• Save time by automating manual tasks
11. How does it work?
• Simple client-server
architecture
• Connecting resources to a
Chef server
• Resources pull
configuration updates from
the Chef server Config A Config B
12. How can you set this up?
1. Setup the Chef server with cookbooks, recipes roles.
2. Install the Chef client on the instance (or server).
3. Register the instance with the Chef server as a Chef node.
4. Assign node with a role (e.g. web server, app server, db server).
5. The Chef client pulls the recipes from Chef server (based on role).
6. The Chef server determines the applicable recipes (by role).
7. The Chef client applies the recipes on the node by doing a “Chef run”.
8. The Chef client pulls the Chef server every 30 minutes.
13. What does it look like?
• The Chef client pulls
configuration updates from the
Chef server every 30 minutes.
• The Chef client will only make
configuration changes when
the node is out of spec.
• The Chef client can react to
changes using by using Chef
search.
14. Support for community tools
• ChefDK
• Knife
• Chef Client
• Community cookbooks and recipes
• TestKitchen
15. Chef recipe example – configure Apache
# Install Apache and start the service.
httpd_service ‘default' do
listen_ports ['81', '82']
threadlimit '4096'
action [:create, :start]
end
# Add the site configuration.
httpd_config ‘default' do
instance ‘default'
source ‘mysite.conf.erb'
notifies :restart, 'httpd_service[default]'
end
.....
16. Chef recipe example – configure Apache
# Create the document root directory.
directory '/var/www/default/public_html' do
recursive true
end
# Write the home page.
file '/var/www/default/public_html/index.html' do
content '<html>This is a placeholder</html>'
mode '0644'
owner 'web_admin'
group 'web_admin'
end
.....
17. Chef recipe example – configure PHP
# Install the mod_php5 Apache module.
httpd_module 'php5' do
instance ‘default'
end
# Install php5-mysql.
package 'php5-mysql' do
action :install
notifies :restart, 'httpd_service[default]'
end
18. Get visibility into the state of your nodes
Visibility – A view into convergence, compliance, cookbooks, recipes and more.
19. Not only a Configuration Management tool
Workflow – A continuous delivery pipeline of infrastructure and applications.
20. Not only a Configuration Management tool
Compliance - Discovery and analysis of compliance risks across environments
22. What is AWS OpsWorks for Chef Automate?
The place you go to for configuration management on AWS
Offers a fully managed Chef Automate server
OpsWorks
23. How can I create an AWS managed Chef server?
Easy to get started, get a Chef Automate server in 10 minutes.
24. What else can I set up?
Setup a weekly maintenance window
•Automatic security updates
•Automatic Chef version upgrades
25. What else can I set up?
Setup a daily/weekly backup schedule
26. What else is left for me to do?
Nothing, this is a fully managed configuration management
service:
•Automatic backups
•Automatic security updates
•Automatic Chef software updates
You can focus on writing cookbooks and recipes that meet
your needs.
27. What other benefits do I get from the service?
• Automatic instance to Chef server registration
• Secure and easy scaling using Auto Scaling Groups
• No separate license fees, only pay for what you use
• Supports both Amazon EC2 and on-prem resources
• Best practices, AWS support and guidance
28. Where does it come in the tool chain?
• Bootstrap instances with the right configuration
• Update the configuration of running instances
• Assure instances comply with a pre-defined policy
• A part of your Continuous Integration and Continuous
Delivery pipeline
30. How do I get started?
Ø Grab some community cookbooks
https://supermarket.chef.io/
Ø Learn more
https://www.chef.io/automate/
Get started
https://aws.amazon.com/opsworks/