This document summarizes an event promoting open data on AWS. It introduces open data initiatives like Landsat and NEXRAD imagery available on S3. The document highlights how AWS tools like S3, EC2, and EMR can be used to access, analyze and visualize open data at scale. It discusses how AWS aims to eliminate the "undifferentiated heavy lifting" of data preparation by pre-processing datasets and notifying subscribers of updates.
Making Earth observation data available by using Amazon S3 is accelerating scientific discovery and enabling the creation of new products. Attend and learn how the scale and performance of Amazon S3 lets earth scientists, researchers, startups, and GIS professionals gather and analyse planetary-scale data without worrying about limitations of bandwidth, storage, memory, or processing power. Co-presented with support of the Australian Geoscience Data Cube collaboration, DigitalGlobe’s Geospatial Big Data Platform and the developer of the popular ObservedEarth mobile app.
Speakers:
Craig Lawton, Public Sector Solutions Architect, Amazon Web Services
Lachlan Hurst, Observed Earth
Matt Paget, Senior Experimental Scientist, CSIRO
Dan Getman, Digital Globe
AWS re:Invent 2016: Visualizing Big Data Insights with Amazon QuickSight (BDM...Amazon Web Services
Amazon QuickSight is a fast BI service that makes it easy for you to build visualizations, perform ad-hoc analysis, and quickly get business insights from your data. QuickSight is built to harness the power and scalability of the cloud, so you can easily run analysis on large datasets, and support hundreds of thousands of users. In this session, we’ll demonstrate how you can easily get started with Amazon QuickSight, uploading files, connecting to S3 and Redshift and creating analyses from visualizations that are optimized based on the underlying data. Once we’ve built our analysis and dashboard, we’ll show you easy it is to share it with colleagues and stakeholders in just a few seconds. And with SPICE – QuckSight’s in-memory calculation engine – you can go from data to insights, faster than ever.
Working with big volumes of data is a complicated task, but it's even harder if you have to do everything in real time and try to figure it all out yourself. Over the past decades many open-source projects helped solve problems within the data analytics lifecycle around ingestion, storage, processing and visualisation of data. This session will use practical examples to discuss architectural best practices and lessons learned when solving real-time analytics and data visualisation decision-making problems with open-source at scale with the power of Amazon Web Services. It furthermore dives into a demo, using source code from the AWS Labs to visualise live data streams at scale.
Olivier Klein, Solutions Architect, Amazon Web Services, Greater China
Innovation is based on many components – a great idea, creativity, persistence, the right data, and technology tools. Amazon Web Services has an engine of innovation for the start-up community, and it’s now being used to power innovative solutions for big societal problems. As government data becomes more widely available, more people can use AWS computing and big data analytics tools to tackle problems that were, until recently, exclusively the domain of government projects. Scientists, developers, and curious citizens are more equipped than ever to find forward-thinking and entirely new solutions for some of the world’s biggest challenges. These opportunities for innovation are improving citizen services and creating opportunities for a new class of civic tech entrepreneur. This session will highlight real examples of open data enabling transformative innovation on the national and local levels. You will hear about GIS Open Data, NASA’s Citizen Science program and how the new Landsat AWS Public Data Set has supported the development of new applications for citizens and government, and other examples of how open data has helped drive citizen engagement.
Steve Sofian, Solution Architect, Amazon Web Services, WWPS, ASEAN
Understanding AWS Managed Database and Analytics Services | AWS Public Sector...Amazon Web Services
The world is creating more data in more ways than ever before. The average internet user in 2017 generates 1.5GB of data per day, with the rate doubling every 18 months. A single autonomous vehicle can generate 4TB per day. Each smart manufacturing plant generates 1PB per day. Storing, managing, and analyzing this data requires integrated database and analytic services that provide reliability and security at scale. AWS offers a range of managed data services that let customers focus on making data useful, including Amazon Aurora, RDS, DynamoDB, Redshift, Spectrum, ElastiCache, Kinesis, EMR, Elasticsearch Service, and Glue. In this session, we discuss these services, share our vision for innovation, and show how our customers use these services today. Learn More: https://aws.amazon.com/government-education/
Making Earth observation data available by using Amazon S3 is accelerating scientific discovery and enabling the creation of new products. Attend and learn how the scale and performance of Amazon S3 lets earth scientists, researchers, startups, and GIS professionals gather and analyse planetary-scale data without worrying about limitations of bandwidth, storage, memory, or processing power. Co-presented with support of the Australian Geoscience Data Cube collaboration, DigitalGlobe’s Geospatial Big Data Platform and the developer of the popular ObservedEarth mobile app.
Speakers:
Craig Lawton, Public Sector Solutions Architect, Amazon Web Services
Lachlan Hurst, Observed Earth
Matt Paget, Senior Experimental Scientist, CSIRO
Dan Getman, Digital Globe
AWS re:Invent 2016: Visualizing Big Data Insights with Amazon QuickSight (BDM...Amazon Web Services
Amazon QuickSight is a fast BI service that makes it easy for you to build visualizations, perform ad-hoc analysis, and quickly get business insights from your data. QuickSight is built to harness the power and scalability of the cloud, so you can easily run analysis on large datasets, and support hundreds of thousands of users. In this session, we’ll demonstrate how you can easily get started with Amazon QuickSight, uploading files, connecting to S3 and Redshift and creating analyses from visualizations that are optimized based on the underlying data. Once we’ve built our analysis and dashboard, we’ll show you easy it is to share it with colleagues and stakeholders in just a few seconds. And with SPICE – QuckSight’s in-memory calculation engine – you can go from data to insights, faster than ever.
Working with big volumes of data is a complicated task, but it's even harder if you have to do everything in real time and try to figure it all out yourself. Over the past decades many open-source projects helped solve problems within the data analytics lifecycle around ingestion, storage, processing and visualisation of data. This session will use practical examples to discuss architectural best practices and lessons learned when solving real-time analytics and data visualisation decision-making problems with open-source at scale with the power of Amazon Web Services. It furthermore dives into a demo, using source code from the AWS Labs to visualise live data streams at scale.
Olivier Klein, Solutions Architect, Amazon Web Services, Greater China
Innovation is based on many components – a great idea, creativity, persistence, the right data, and technology tools. Amazon Web Services has an engine of innovation for the start-up community, and it’s now being used to power innovative solutions for big societal problems. As government data becomes more widely available, more people can use AWS computing and big data analytics tools to tackle problems that were, until recently, exclusively the domain of government projects. Scientists, developers, and curious citizens are more equipped than ever to find forward-thinking and entirely new solutions for some of the world’s biggest challenges. These opportunities for innovation are improving citizen services and creating opportunities for a new class of civic tech entrepreneur. This session will highlight real examples of open data enabling transformative innovation on the national and local levels. You will hear about GIS Open Data, NASA’s Citizen Science program and how the new Landsat AWS Public Data Set has supported the development of new applications for citizens and government, and other examples of how open data has helped drive citizen engagement.
Steve Sofian, Solution Architect, Amazon Web Services, WWPS, ASEAN
Understanding AWS Managed Database and Analytics Services | AWS Public Sector...Amazon Web Services
The world is creating more data in more ways than ever before. The average internet user in 2017 generates 1.5GB of data per day, with the rate doubling every 18 months. A single autonomous vehicle can generate 4TB per day. Each smart manufacturing plant generates 1PB per day. Storing, managing, and analyzing this data requires integrated database and analytic services that provide reliability and security at scale. AWS offers a range of managed data services that let customers focus on making data useful, including Amazon Aurora, RDS, DynamoDB, Redshift, Spectrum, ElastiCache, Kinesis, EMR, Elasticsearch Service, and Glue. In this session, we discuss these services, share our vision for innovation, and show how our customers use these services today. Learn More: https://aws.amazon.com/government-education/
#EarthOnAWS: How the Cloud Is Transforming Earth Observation | AWS Public Sec...Amazon Web Services
Making earth observation data available in the cloud is accelerating scientific discovery and enabling the creation of new products. Attend and learn how the cloud lets earth scientists, researchers, startups, and GIS professionals gather and analyze earth observation data without worrying about limitations of bandwidth, storage, memory, or processing power. Join us and learn how earth science data projects are becoming more scalable, agile, and efficient with AWS on-demand IT infrastructure.
Antoine Genereux takes us on a detailed overview of the Database solutions available on the AWS Cloud, addressing the needs and requirements of customers at all levels. He also discusses Business Intelligence and Analytics solutions.
Organizations around the world are facing a "data tsunami" as next-generation sensors produce enormous volumes of Earth observation data. Come learn how NASA is leveraging AWS to efficiently work with data and computing resources at massive scales. NASA is transforming its Earth Sciences EOSDIS (Earth Observing System Data Information System) program by moving data processing and archiving to the cloud. NASA anticipates that their Data Archives will grow from 16PB today to over 400PB by 2023 and 1 Exabyte by 2030, and they are moving to the cloud in order to scale their operations for this new paradigm. Learn More: https://aws.amazon.com/government-education/
Join us for a series of introductory and technical sessions on AWS Big Data solutions. Gain a thorough understanding of what Amazon Web Services offers across the big data lifecycle and learn architectural best practices for applying those solutions to your projects.
We will kick off this technical seminar in the morning with an introduction to the AWS Big Data platform, including a discussion of popular use cases and reference architectures. In the afternoon, we will deep dive into Machine Learning and Streaming Analytics. We will then walk everyone through building your first Big Data application with AWS.
Analytics on AWS with Amazon Redshift, Amazon QuickSight, and Amazon Machine ...Amazon Web Services
AWS has a large and growing portfolio of big data management and analytics services, designed to be integrated into solution architectures that meet the needs of your business. In this session, we look at analytics through the eyes of a business intelligence analyst, a data scientist, and an application developer, and we explore how to quickly leverage Amazon Redshift, Amazon QuickSight, RStudio, and Amazon Machine Learning to create powerful, yet straightforward, business solutions.
Speaker:
Paul Armstrong, Solutions Architect, Amazon Web Services
Real-time Analytics using Data from IoT Devices - AWS Online Tech TalksAmazon Web Services
Learning Objectives:
- Learn the different options available to stream data from IoT sensors to AWS
- Understand how to architect an analytics solution using AWS services to ingest and process IoT data
- Take away best practices for building IoT applications with scalability, cost-effectiveness, and security
AWS Compute Overview: Servers, Containers, Serverless, and Batch | AWS Public...Amazon Web Services
The AWS Compute platform has expanded EC2 instance types including FPGA and new GPU instances. There are also other ways to run workloads in AWS including Lambda (serverless), ECS (managed Docker), and AWS Batch (batch computing). This session will cover the newest instance types in EC2 and review AWS Lambda, ECS, and Batch. Learn More: https://aws.amazon.com/government-education/
AWS re:Invent 2016: How to Build a Big Data Analytics Data Lake (LFS303)Amazon Web Services
For discovery-phase research, life sciences companies have to support infrastructure that processes millions to billions of transactions. The advent of a data lake to accomplish such a task is showing itself to be a stable and productive data platform pattern to meet the goal. We discuss how to build a data lake on AWS, using services and techniques such as AWS CloudFormation, Amazon EC2, Amazon S3, IAM, and AWS Lambda. We also review a reference architecture from Amgen that uses a data lake to aid in their Life Science Research.
Businesses are generating more data than ever before.
Doing real time data analytics requires IT infrastructure that often needs to be scaled up quickly and running an on-premise environment in this setting has its limitations.
Organisations often require a massive amount of IT resources to analyse their data and the upfront capital cost can deter them from embarking on these projects.
What’s needed is scalable, agile and secure cloud-based infrastructure at the lowest possible cost so they can spin up servers that support their data analysis projects exactly when they are required. This infrastructure must enable them to create proof-of-concepts quickly and cheaply – to fail fast and move on.
AWS re:Invent 2016: How Fulfillment by Amazon (FBA) and Scopely Improved Resu...Amazon Web Services
We’ll share an overview of leveraging serverless architectures to support high performance data intensive applications. Fulfillment by Amazon (FBA) built the Seller Inventory Authority Platform (IAP) using Amazon DynamoDB Streams, AWS Lambda functions, Amazon Elasticsearch Service, and Amazon Redshift to improve results and reduce costs. Scopely will share how they used a flexible logging system built on Kinesis, Lambda, and Amazon Elasticsearch to provide high-fidelity reporting on hotkeys in Memcached and DynamoDB, and drastically reduce the incidence of hotkeys. Both of these customers are using managed services and serverless architecture to build scalable systems that can meet the projected business growth without a corresponding increase in operational costs.
AWS re:Invent 2016: Earth on AWS—Next-Generation Open Data Platforms (STG203)Amazon Web Services
Making earth observation data available by using Amazon S3 is accelerating scientific discovery and enabling the creation of new products. Attend and learn how the scale and performance of Amazon S3 lets earth scientists, researchers, startups, and GIS professionals gather and analyze planetary-scale data without worrying about limitations of bandwidth, storage, memory, or processing power. Learn how AWS is being used to combine satellite imagery, social data, and telemetry data to produce new products and services. Learn also how Amazon S3 provides much more than storage, and how an open geospatial data lake on Amazon S3 can be used as the basis for planetary-scale applications built with Amazon EMR, Amazon API Gateway, and AWS Lambda. As part of this talk, AWS customer Digital Globe demonstrates how they use open data stored in S3 to distribute high-resolution satellite imagery to their customers around the world.
AWS provides a broad platform of managed services to help you build, secure, and seamlessly scale end-to-end Big Data applications quickly and with ease. Want to get ramped up on how to use Amazon's big data web services? Learn when to use which service? Want to write your first big data application on AWS? Join us in this session as we discuss reference architecture, design patterns, and best practices for pulling together various AWS services to meet your big data challenges.
AWS Summit 2013 | India - Big Data Analytics, Abhishek SinhaAmazon Web Services
The volume, velocity and variety of data has changed drastically in the last decade. Everything generates data today, from your customers on social networks, to the instances running your web applications. The tools to support collecting, storing, organizing, analyzing and sharing of data are all available in a couple of clicks, with Amazon Web Services. Attend this session to learn how Big Data in the cloud can help you easily unlock business opportunities hidden in your data today.
AWS-powered services for analytics can handle the scale, agility, and flexibility required to combine different types of data and analytics approaches that will allow you to transform your data into a valuable corporate asset. In this session, AWS will provide an overview of the different AWS services available for your data analytics needs. You can combine these blocks to build data flows that will extend your organization’s agility, ability to derive more insights and value from its data, and capability to adopt more sophisticated analytics tools and processes as your needs evolve. In the second part of the session, Paddy Power Betfair’s Data team will discuss the adoption and large scale operation of a broad range of AWS services that make up PPB’s scalable, mixed workload, multi-brand data platform. The data capabilities developed by PPB and powered by AWS were implemented to enable low-latency, high-volume and near real-time advanced analytics use cases, in the highly regulated and fast-paced betting industry. This was only possible through a focus on automation, innovation and continuous improvement.
Stream Data Analytics with Amazon Kinesis Firehose & Redshift - AWS August We...Amazon Web Services
Evolving your analytics from batch processing to real-time processing can have a major business impact, but ingesting streaming data into your data warehouse requires building complex streaming data pipelines. Amazon Kinesis Firehose solves this problem by making it easy to ingest streaming data into Amazon Redshift so that you can use existing analytics and business intelligence tools to extract information in near real-time and respond promptly. In this webinar, we will dive deep using Amazon Kinesis Firehose to load streaming data into Amazon Redshift reliably, scalably, and cost-effectively. Join us to: - Understand the basics of ingesting streaming data from sources such as mobile devices, servers, and websites with Amazon Kinesis Firehose - Get a closer look at how to automate delivery of streaming data to Amazon Redshift reliably using Amazon Kinesis Firehose - Learn techniques to detect, troubleshoot, and avoid data loading problems Who should attend: Developers, data analysts, data engineers, architects
NEW LAUNCH! Introducing AWS Batch: Easy and efficient batch computingAmazon Web Services
AWS Batch is a fully-managed service that enables developers, scientists, and engineers to easily and efficiently run batch computing workloads of any scale on AWS. AWS Batch automatically provisions compute resources and optimizes the workload distribution based on the quantity and scale of the workloads. With AWS Batch, there is no need to install or manage batch computing software, allowing you to focus on analyzing results and solving problems. AWS Batch plans, schedules, and executes your batch computing workloads across the full range of AWS compute services and features, such as Amazon EC2, Spot Instances, and AWS Lambda. AWS Batch reduces operational complexities, saving time and reducing costs. In this session, Principal Product Managers Jamie Kinney and Dougal Ballantyne describe the core concepts behind AWS Batch and details of how the service functions. The presentation concludes with relevant use cases and sample code.
AWS Summit 2013 | Singapore - Big Data Analytics, Presented by AWS, Intel and...Amazon Web Services
Learn more about the tools, techniques and technologies for working productively with data at any scale. This session will introduce the family of data analytics tools on AWS which you can use to collect, compute and collaborate around data, from gigabytes to petabytes. We'll discuss Amazon Elastic MapReduce, Hadoop, structured and unstructured data, and the EC2 instance types which enable high performance analytics.
#EarthOnAWS: How the Cloud Is Transforming Earth Observation | AWS Public Sec...Amazon Web Services
Making earth observation data available in the cloud is accelerating scientific discovery and enabling the creation of new products. Attend and learn how the cloud lets earth scientists, researchers, startups, and GIS professionals gather and analyze earth observation data without worrying about limitations of bandwidth, storage, memory, or processing power. Join us and learn how earth science data projects are becoming more scalable, agile, and efficient with AWS on-demand IT infrastructure.
Antoine Genereux takes us on a detailed overview of the Database solutions available on the AWS Cloud, addressing the needs and requirements of customers at all levels. He also discusses Business Intelligence and Analytics solutions.
Organizations around the world are facing a "data tsunami" as next-generation sensors produce enormous volumes of Earth observation data. Come learn how NASA is leveraging AWS to efficiently work with data and computing resources at massive scales. NASA is transforming its Earth Sciences EOSDIS (Earth Observing System Data Information System) program by moving data processing and archiving to the cloud. NASA anticipates that their Data Archives will grow from 16PB today to over 400PB by 2023 and 1 Exabyte by 2030, and they are moving to the cloud in order to scale their operations for this new paradigm. Learn More: https://aws.amazon.com/government-education/
Join us for a series of introductory and technical sessions on AWS Big Data solutions. Gain a thorough understanding of what Amazon Web Services offers across the big data lifecycle and learn architectural best practices for applying those solutions to your projects.
We will kick off this technical seminar in the morning with an introduction to the AWS Big Data platform, including a discussion of popular use cases and reference architectures. In the afternoon, we will deep dive into Machine Learning and Streaming Analytics. We will then walk everyone through building your first Big Data application with AWS.
Analytics on AWS with Amazon Redshift, Amazon QuickSight, and Amazon Machine ...Amazon Web Services
AWS has a large and growing portfolio of big data management and analytics services, designed to be integrated into solution architectures that meet the needs of your business. In this session, we look at analytics through the eyes of a business intelligence analyst, a data scientist, and an application developer, and we explore how to quickly leverage Amazon Redshift, Amazon QuickSight, RStudio, and Amazon Machine Learning to create powerful, yet straightforward, business solutions.
Speaker:
Paul Armstrong, Solutions Architect, Amazon Web Services
Real-time Analytics using Data from IoT Devices - AWS Online Tech TalksAmazon Web Services
Learning Objectives:
- Learn the different options available to stream data from IoT sensors to AWS
- Understand how to architect an analytics solution using AWS services to ingest and process IoT data
- Take away best practices for building IoT applications with scalability, cost-effectiveness, and security
AWS Compute Overview: Servers, Containers, Serverless, and Batch | AWS Public...Amazon Web Services
The AWS Compute platform has expanded EC2 instance types including FPGA and new GPU instances. There are also other ways to run workloads in AWS including Lambda (serverless), ECS (managed Docker), and AWS Batch (batch computing). This session will cover the newest instance types in EC2 and review AWS Lambda, ECS, and Batch. Learn More: https://aws.amazon.com/government-education/
AWS re:Invent 2016: How to Build a Big Data Analytics Data Lake (LFS303)Amazon Web Services
For discovery-phase research, life sciences companies have to support infrastructure that processes millions to billions of transactions. The advent of a data lake to accomplish such a task is showing itself to be a stable and productive data platform pattern to meet the goal. We discuss how to build a data lake on AWS, using services and techniques such as AWS CloudFormation, Amazon EC2, Amazon S3, IAM, and AWS Lambda. We also review a reference architecture from Amgen that uses a data lake to aid in their Life Science Research.
Businesses are generating more data than ever before.
Doing real time data analytics requires IT infrastructure that often needs to be scaled up quickly and running an on-premise environment in this setting has its limitations.
Organisations often require a massive amount of IT resources to analyse their data and the upfront capital cost can deter them from embarking on these projects.
What’s needed is scalable, agile and secure cloud-based infrastructure at the lowest possible cost so they can spin up servers that support their data analysis projects exactly when they are required. This infrastructure must enable them to create proof-of-concepts quickly and cheaply – to fail fast and move on.
AWS re:Invent 2016: How Fulfillment by Amazon (FBA) and Scopely Improved Resu...Amazon Web Services
We’ll share an overview of leveraging serverless architectures to support high performance data intensive applications. Fulfillment by Amazon (FBA) built the Seller Inventory Authority Platform (IAP) using Amazon DynamoDB Streams, AWS Lambda functions, Amazon Elasticsearch Service, and Amazon Redshift to improve results and reduce costs. Scopely will share how they used a flexible logging system built on Kinesis, Lambda, and Amazon Elasticsearch to provide high-fidelity reporting on hotkeys in Memcached and DynamoDB, and drastically reduce the incidence of hotkeys. Both of these customers are using managed services and serverless architecture to build scalable systems that can meet the projected business growth without a corresponding increase in operational costs.
AWS re:Invent 2016: Earth on AWS—Next-Generation Open Data Platforms (STG203)Amazon Web Services
Making earth observation data available by using Amazon S3 is accelerating scientific discovery and enabling the creation of new products. Attend and learn how the scale and performance of Amazon S3 lets earth scientists, researchers, startups, and GIS professionals gather and analyze planetary-scale data without worrying about limitations of bandwidth, storage, memory, or processing power. Learn how AWS is being used to combine satellite imagery, social data, and telemetry data to produce new products and services. Learn also how Amazon S3 provides much more than storage, and how an open geospatial data lake on Amazon S3 can be used as the basis for planetary-scale applications built with Amazon EMR, Amazon API Gateway, and AWS Lambda. As part of this talk, AWS customer Digital Globe demonstrates how they use open data stored in S3 to distribute high-resolution satellite imagery to their customers around the world.
AWS provides a broad platform of managed services to help you build, secure, and seamlessly scale end-to-end Big Data applications quickly and with ease. Want to get ramped up on how to use Amazon's big data web services? Learn when to use which service? Want to write your first big data application on AWS? Join us in this session as we discuss reference architecture, design patterns, and best practices for pulling together various AWS services to meet your big data challenges.
AWS Summit 2013 | India - Big Data Analytics, Abhishek SinhaAmazon Web Services
The volume, velocity and variety of data has changed drastically in the last decade. Everything generates data today, from your customers on social networks, to the instances running your web applications. The tools to support collecting, storing, organizing, analyzing and sharing of data are all available in a couple of clicks, with Amazon Web Services. Attend this session to learn how Big Data in the cloud can help you easily unlock business opportunities hidden in your data today.
AWS-powered services for analytics can handle the scale, agility, and flexibility required to combine different types of data and analytics approaches that will allow you to transform your data into a valuable corporate asset. In this session, AWS will provide an overview of the different AWS services available for your data analytics needs. You can combine these blocks to build data flows that will extend your organization’s agility, ability to derive more insights and value from its data, and capability to adopt more sophisticated analytics tools and processes as your needs evolve. In the second part of the session, Paddy Power Betfair’s Data team will discuss the adoption and large scale operation of a broad range of AWS services that make up PPB’s scalable, mixed workload, multi-brand data platform. The data capabilities developed by PPB and powered by AWS were implemented to enable low-latency, high-volume and near real-time advanced analytics use cases, in the highly regulated and fast-paced betting industry. This was only possible through a focus on automation, innovation and continuous improvement.
Stream Data Analytics with Amazon Kinesis Firehose & Redshift - AWS August We...Amazon Web Services
Evolving your analytics from batch processing to real-time processing can have a major business impact, but ingesting streaming data into your data warehouse requires building complex streaming data pipelines. Amazon Kinesis Firehose solves this problem by making it easy to ingest streaming data into Amazon Redshift so that you can use existing analytics and business intelligence tools to extract information in near real-time and respond promptly. In this webinar, we will dive deep using Amazon Kinesis Firehose to load streaming data into Amazon Redshift reliably, scalably, and cost-effectively. Join us to: - Understand the basics of ingesting streaming data from sources such as mobile devices, servers, and websites with Amazon Kinesis Firehose - Get a closer look at how to automate delivery of streaming data to Amazon Redshift reliably using Amazon Kinesis Firehose - Learn techniques to detect, troubleshoot, and avoid data loading problems Who should attend: Developers, data analysts, data engineers, architects
NEW LAUNCH! Introducing AWS Batch: Easy and efficient batch computingAmazon Web Services
AWS Batch is a fully-managed service that enables developers, scientists, and engineers to easily and efficiently run batch computing workloads of any scale on AWS. AWS Batch automatically provisions compute resources and optimizes the workload distribution based on the quantity and scale of the workloads. With AWS Batch, there is no need to install or manage batch computing software, allowing you to focus on analyzing results and solving problems. AWS Batch plans, schedules, and executes your batch computing workloads across the full range of AWS compute services and features, such as Amazon EC2, Spot Instances, and AWS Lambda. AWS Batch reduces operational complexities, saving time and reducing costs. In this session, Principal Product Managers Jamie Kinney and Dougal Ballantyne describe the core concepts behind AWS Batch and details of how the service functions. The presentation concludes with relevant use cases and sample code.
AWS Summit 2013 | Singapore - Big Data Analytics, Presented by AWS, Intel and...Amazon Web Services
Learn more about the tools, techniques and technologies for working productively with data at any scale. This session will introduce the family of data analytics tools on AWS which you can use to collect, compute and collaborate around data, from gigabytes to petabytes. We'll discuss Amazon Elastic MapReduce, Hadoop, structured and unstructured data, and the EC2 instance types which enable high performance analytics.
Your data has value for multiple business functions in your organization. Shorten your time to analytics and take faster, better decisions based on data.
In this session you will learn how you can access your data from a myriad of tools such as multiple EMR clusters, Athena & Redshift.
AWS re:Invent 2016: Big Data Mini Con State of the Union (BDM205)Amazon Web Services
Join us for this general session where AWS big data experts present an in-depth look at the current state of big data. Learn about the latest big data trends and industry use cases. Hear how other organizations are using the AWS big data platform to innovate and remain competitive. Take a look at some of the most recent AWS big data announcements, as we kick off the Big Data re:Source Mini Con.
(BDT201) Big Data and HPC State of the Union | AWS re:Invent 2014Amazon Web Services
Leveraging big data and high performance computing (HPC) solutions enables your organization to make smarter and faster decisions that influence strategy, increase productivity, and ultimately grow your business. We kick off the Big Data and HPC track with the latest advancements in data analytics, databases, storage, and HPC at AWS. Hear customer success stories and discover how to put data to work in your own organization.
Understanding AWS Managed Database and Analytics Services | AWS Public Sector...Amazon Web Services
The world is creating more data in more ways than ever before. The average internet user in 2017 generates 1.5GB of data per day, with the rate doubling every 18 months. A single autonomous vehicle can generate 4TB per day. Each smart manufacturing plant generates 1PB per day. Storing, managing, and analyzing this data requires integrated database and analytic services that provide reliability and security at scale. AWS offers a range of managed data services that let customers focus on making data useful, including Amazon Aurora, RDS, DynamoDB, Redshift, Spectrum, ElastiCache, Kinesis, EMR, Elasticsearch Service, and Glue. In this session, we discuss these services, share our vision for innovation, and show how our customers use these services today. Learn More: https://aws.amazon.com/government-education/
This overview presentation discusses big data challenges and provides an overview of the AWS Big Data Platform by covering:
- How AWS customers leverage the platform to manage massive volumes of data from a variety of sources while containing costs.
- Reference architectures for popular use cases, including, connected devices (IoT), log streaming, real-time intelligence, and analytics.
- The AWS big data portfolio of services, including, Amazon S3, Kinesis, DynamoDB, Elastic MapReduce (EMR), and Redshift.
- The latest relational database engine, Amazon Aurora— a MySQL-compatible, highly-available relational database engine, which provides up to five times better performance than MySQL at one-tenth the cost of a commercial database.
Created by: Rahul Pathak,
Sr. Manager of Software Development
Hybrid as a Stepping Stone: It’s Not All or Nothing for Your Cloud Transforma...Amazon Web Services
The implementation of highly scalable, easy-to-deploy technology is transforming the public sector, but it's not a one-size-fits-all approach. Organizations begin their cloud adoption journeys in many ways. Some start with pilot projects and others jump into mission critical programs, but they are all starting with an existing infrastructure. Adopting cloud doesn't mean scrapping it all and starting over. This session explores how organizations can extend their existing IT platforms into the cloud to enable hybrid capabilities capable of supporting every phase of their transformation. Learn More: https://aws.amazon.com/government-education/
The Presentation Talks about how Cloud Computing is Big Data's Best Friend and How AWS Cloud Components Fit in to complete your Big Data Life Cycle.
Agenda:
- How Big is Big Data Actually growing?
- How Cloud has the potential to become Big Data's Best Friend
- A tour on The Big Data Life Cycle
- How AWS Cloud Components Fit in to this Life Cycle
- A Case Study of Our Log Analytics Tool Cloudlytics, using Big Data Implementation
on AWS Cloud.
Companies, from startups to enterprises across the globe, are looking to migrate data warehousing to the cloud to increase performance and lower costs. Data engineers, data analysts, and developers also need to access and consume this important data. The landscape is constantly evolving and there are many solutions available for enterprises of all sizes. In this workshop, we dive deep into architectural patterns, use cases, and best practices when designing an enterprise data warehouse in the cloud. We also address key issues such as data governance and democratization. At the end of this workshop, you’ll be equipped to design and implement a cloud enterprise data warehouse platform that provides the most benefit for your enterprise, data consumers, and customers.
How to use big data to improve a EC platform is a hot topic. In this session, we will discuss some big data case studies in retail and EC, and introduce how to create a recommendation service with Amazon Machine Learning.
Slides: Proven Strategies for Hybrid Cloud Computing with Mainframes — From A...DATAVERSITY
Mainframes continue to perform mission-critical transaction processing and contain massive amounts of core business data. But digital transformation initiatives and cloud computing have created both opportunities and challenges for unlocking and utilizing this data. Qlik and AWS will share some of the proven strategies from successful customer deployments across a range of different mainframe to cloud use cases, including legacy application modernization, data analytics, and data migrations.
In this presentation, you will learn how to:
• Replicate very large volumes of mainframe data in real-time to the cloud
• Automate the creation of analytics-ready data lakes and data warehouses
• Achieve a 30% reduction in cost of compute
What is Innovation? How can cloud computing help you innovate? How can you make your applications smarter? Predictive? How can you interpret data and anticipate trends? With AWS Artificial Intelligence Solutions: Machine Learning, Rekognition, Polly; with serverless - Lambda, Step Functions.
Data Con LA 2020
Description
In this session, I introduce the Amazon Redshift lake house architecture which enables you to query data across your data warehouse, data lake, and operational databases to gain faster and deeper insights. With a lake house architecture, you can store data in open file formats in your Amazon S3 data lake.
Speaker
Antje Barth, Amazon Web Services, Sr. Developer Advocate, AI and Machine Learning
Understanding AWS Managed Databases and Analytic Services - AWS Innovate Otta...Amazon Web Services
• Overview of database services to elevate your applications, analytic services to engage your data, and migration services to help you reach database freedom.
• Survey of how Canadian and other organizations are using the cloud to make data scalable, reliable, and secure.
Come costruire servizi di Forecasting sfruttando algoritmi di ML e deep learn...Amazon Web Services
Il Forecasting è un processo importante per tantissime aziende e viene utilizzato in vari ambiti per cercare di prevedere in modo accurato la crescita e distribuzione di un prodotto, l’utilizzo delle risorse necessarie nelle linee produttive, presentazioni finanziarie e tanto altro. Amazon utilizza delle tecniche avanzate di forecasting, in parte questi servizi sono stati messi a disposizione di tutti i clienti AWS.
In questa sessione illustreremo come pre-processare i dati che contengono una componente temporale e successivamente utilizzare un algoritmo che a partire dal tipo di dato analizzato produce un forecasting accurato.
Big Data per le Startup: come creare applicazioni Big Data in modalità Server...Amazon Web Services
La varietà e la quantità di dati che si crea ogni giorno accelera sempre più velocemente e rappresenta una opportunità irripetibile per innovare e creare nuove startup.
Tuttavia gestire grandi quantità di dati può apparire complesso: creare cluster Big Data su larga scala sembra essere un investimento accessibile solo ad aziende consolidate. Ma l’elasticità del Cloud e, in particolare, i servizi Serverless ci permettono di rompere questi limiti.
Vediamo quindi come è possibile sviluppare applicazioni Big Data rapidamente, senza preoccuparci dell’infrastruttura, ma dedicando tutte le risorse allo sviluppo delle nostre le nostre idee per creare prodotti innovativi.
Ora puoi utilizzare Amazon Elastic Kubernetes Service (EKS) per eseguire pod Kubernetes su AWS Fargate, il motore di elaborazione serverless creato per container su AWS. Questo rende più semplice che mai costruire ed eseguire le tue applicazioni Kubernetes nel cloud AWS.In questa sessione presenteremo le caratteristiche principali del servizio e come distribuire la tua applicazione in pochi passaggi
Vent'anni fa Amazon ha attraversato una trasformazione radicale con l'obiettivo di aumentare il ritmo dell'innovazione. In questo periodo abbiamo imparato come cambiare il nostro approccio allo sviluppo delle applicazioni ci ha permesso di aumentare notevolmente l'agilità, la velocità di rilascio e, in definitiva, ci ha consentito di creare applicazioni più affidabili e scalabili. In questa sessione illustreremo come definiamo le applicazioni moderne e come la creazione di app moderne influisce non solo sull'architettura dell'applicazione, ma sulla struttura organizzativa, sulle pipeline di rilascio dello sviluppo e persino sul modello operativo. Descriveremo anche approcci comuni alla modernizzazione, compreso l'approccio utilizzato dalla stessa Amazon.com.
Come spendere fino al 90% in meno con i container e le istanze spot Amazon Web Services
L’utilizzo dei container è in continua crescita.
Se correttamente disegnate, le applicazioni basate su Container sono molto spesso stateless e flessibili.
I servizi AWS ECS, EKS e Kubernetes su EC2 possono sfruttare le istanze Spot, portando ad un risparmio medio del 70% rispetto alle istanze On Demand. In questa sessione scopriremo insieme quali sono le caratteristiche delle istanze Spot e come possono essere utilizzate facilmente su AWS. Impareremo inoltre come Spreaker sfrutta le istanze spot per eseguire applicazioni di diverso tipo, in produzione, ad una frazione del costo on-demand!
In recent months, many customers have been asking us the question – how to monetise Open APIs, simplify Fintech integrations and accelerate adoption of various Open Banking business models. Therefore, AWS and FinConecta would like to invite you to Open Finance marketplace presentation on October 20th.
Event Agenda :
Open banking so far (short recap)
• PSD2, OB UK, OB Australia, OB LATAM, OB Israel
Intro to Open Finance marketplace
• Scope
• Features
• Tech overview and Demo
The role of the Cloud
The Future of APIs
• Complying with regulation
• Monetizing data / APIs
• Business models
• Time to market
One platform for all: a Strategic approach
Q&A
Rendi unica l’offerta della tua startup sul mercato con i servizi Machine Lea...Amazon Web Services
Per creare valore e costruire una propria offerta differenziante e riconoscibile, le startup di successo sanno come combinare tecnologie consolidate con componenti innovativi creati ad hoc.
AWS fornisce servizi pronti all'utilizzo e, allo stesso tempo, permette di personalizzare e creare gli elementi differenzianti della propria offerta.
Concentrandoci sulle tecnologie di Machine Learning, vedremo come selezionare i servizi di intelligenza artificiale offerti da AWS e, anche attraverso una demo, come costruire modelli di Machine Learning personalizzati utilizzando SageMaker Studio.
OpsWorks Configuration Management: automatizza la gestione e i deployment del...Amazon Web Services
Con l'approccio tradizionale al mondo IT per molti anni è stato difficile implementare tecniche di DevOps, che finora spesso hanno previsto attività manuali portando di tanto in tanto a dei downtime degli applicativi interrompendo l'operatività dell'utente. Con l'avvento del cloud, le tecniche di DevOps sono ormai a portata di tutti a basso costo per qualsiasi genere di workload, garantendo maggiore affidabilità del sistema e risultando in dei significativi miglioramenti della business continuity.
AWS mette a disposizione AWS OpsWork come strumento di Configuration Management che mira ad automatizzare e semplificare la gestione e i deployment delle istanze EC2 per mezzo di workload Chef e Puppet.
Scopri come sfruttare AWS OpsWork a garanzia e affidabilità del tuo applicativo installato su Instanze EC2.
Microsoft Active Directory su AWS per supportare i tuoi Windows WorkloadsAmazon Web Services
Vuoi conoscere le opzioni per eseguire Microsoft Active Directory su AWS? Quando si spostano carichi di lavoro Microsoft in AWS, è importante considerare come distribuire Microsoft Active Directory per supportare la gestione, l'autenticazione e l'autorizzazione dei criteri di gruppo. In questa sessione, discuteremo le opzioni per la distribuzione di Microsoft Active Directory su AWS, incluso AWS Directory Service per Microsoft Active Directory e la distribuzione di Active Directory su Windows su Amazon Elastic Compute Cloud (Amazon EC2). Trattiamo argomenti quali l'integrazione del tuo ambiente Microsoft Active Directory locale nel cloud e l'utilizzo di applicazioni SaaS, come Office 365, con AWS Single Sign-On.
Dal riconoscimento facciale al riconoscimento di frodi o difetti di fabbricazione, l'analisi di immagini e video che sfruttano tecniche di intelligenza artificiale, si stanno evolvendo e raffinando a ritmi elevati. In questo webinar esploreremo le possibilità messe a disposizione dai servizi AWS per applicare lo stato dell'arte delle tecniche di computer vision a scenari reali.
Amazon Web Services e VMware organizzano un evento virtuale gratuito il prossimo mercoledì 14 Ottobre dalle 12:00 alle 13:00 dedicato a VMware Cloud ™ on AWS, il servizio on demand che consente di eseguire applicazioni in ambienti cloud basati su VMware vSphere® e di accedere ad una vasta gamma di servizi AWS, sfruttando a pieno le potenzialità del cloud AWS e tutelando gli investimenti VMware esistenti.
Molte organizzazioni sfruttano i vantaggi del cloud migrando i propri carichi di lavoro Oracle e assicurandosi notevoli vantaggi in termini di agilità ed efficienza dei costi.
La migrazione di questi carichi di lavoro, può creare complessità durante la modernizzazione e il refactoring delle applicazioni e a questo si possono aggiungere rischi di prestazione che possono essere introdotti quando si spostano le applicazioni dai data center locali.
Crea la tua prima serverless ledger-based app con QLDB e NodeJSAmazon Web Services
Molte aziende oggi, costruiscono applicazioni con funzionalità di tipo ledger ad esempio per verificare lo storico di accrediti o addebiti nelle transazioni bancarie o ancora per tenere traccia del flusso supply chain dei propri prodotti.
Alla base di queste soluzioni ci sono i database ledger che permettono di avere un log delle transazioni trasparente, immutabile e crittograficamente verificabile, ma sono strumenti complessi e onerosi da gestire.
Amazon QLDB elimina la necessità di costruire sistemi personalizzati e complessi fornendo un database ledger serverless completamente gestito.
In questa sessione scopriremo come realizzare un'applicazione serverless completa che utilizzi le funzionalità di QLDB.
Con l’ascesa delle architetture di microservizi e delle ricche applicazioni mobili e Web, le API sono più importanti che mai per offrire agli utenti finali una user experience eccezionale. In questa sessione impareremo come affrontare le moderne sfide di progettazione delle API con GraphQL, un linguaggio di query API open source utilizzato da Facebook, Amazon e altro e come utilizzare AWS AppSync, un servizio GraphQL serverless gestito su AWS. Approfondiremo diversi scenari, comprendendo come AppSync può aiutare a risolvere questi casi d’uso creando API moderne con funzionalità di aggiornamento dati in tempo reale e offline.
Inoltre, impareremo come Sky Italia utilizza AWS AppSync per fornire aggiornamenti sportivi in tempo reale agli utenti del proprio portale web.
Database Oracle e VMware Cloud™ on AWS: i miti da sfatareAmazon Web Services
Molte organizzazioni sfruttano i vantaggi del cloud migrando i propri carichi di lavoro Oracle e assicurandosi notevoli vantaggi in termini di agilità ed efficienza dei costi.
La migrazione di questi carichi di lavoro, può creare complessità durante la modernizzazione e il refactoring delle applicazioni e a questo si possono aggiungere rischi di prestazione che possono essere introdotti quando si spostano le applicazioni dai data center locali.
In queste slide, gli esperti AWS e VMware presentano semplici e pratici accorgimenti per facilitare e semplificare la migrazione dei carichi di lavoro Oracle accelerando la trasformazione verso il cloud, approfondiranno l’architettura e dimostreranno come sfruttare a pieno le potenzialità di VMware Cloud ™ on AWS.
Amazon Elastic Container Service (Amazon ECS) è un servizio di gestione dei container altamente scalabile, che semplifica la gestione dei contenitori Docker attraverso un layer di orchestrazione per il controllo del deployment e del relativo lifecycle. In questa sessione presenteremo le principali caratteristiche del servizio, le architetture di riferimento per i differenti carichi di lavoro e i semplici passi necessari per poter velocemente migrare uno o più dei tuo container.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Let's dive deeper into the world of ODC! Ricardo Alves (OutSystems) will join us to tell all about the new Data Fabric. After that, Sezen de Bruijn (OutSystems) will get into the details on how to best design a sturdy architecture within ODC.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Search and Society: Reimagining Information Access for Radical FuturesBhaskar Mitra
The field of Information retrieval (IR) is currently undergoing a transformative shift, at least partly due to the emerging applications of generative AI to information access. In this talk, we will deliberate on the sociotechnical implications of generative AI for information access. We will argue that there is both a critical necessity and an exciting opportunity for the IR community to re-center our research agendas on societal needs while dismantling the artificial separation between the work on fairness, accountability, transparency, and ethics in IR and the rest of IR research. Instead of adopting a reactionary strategy of trying to mitigate potential social harms from emerging technologies, the community should aim to proactively set the research agenda for the kinds of systems we should build inspired by diverse explicitly stated sociotechnical imaginaries. The sociotechnical imaginaries that underpin the design and development of information access technologies needs to be explicitly articulated, and we need to develop theories of change in context of these diverse perspectives. Our guiding future imaginaries must be informed by other academic fields, such as democratic theory and critical theory, and should be co-developed with social science scholars, legal scholars, civil rights and social justice activists, and artists, among others.
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Tobias Schneck
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
7. Our goals for this event
• Show off amazing work being done by our customers
• Provide opportunities for you to network
• Highlight the diversity of work made possible by Earth
observation data
• Learn about your priorities and needs
8. New whitepaper
This Amazon Web Services, Inc. (AWS) package is provided for informational purposes only. The services discussed
in this package are standard commercial services. This package may include a set of suggested solutions for this
opportunity that are based on our limited information, and should not be construed as a binding offer from AWS. For
current prices for AWS services, please refer to the AWS website at www.aws.amazon.com.
This package includes Amazon Web Services, Inc. commercial, financial, or trade secret data that includes
confidential, and/or trade secret information.
AWS Whitepaper: Minimizing Variable
Costs for Shared Data
November 2015
Amazon Web Services, Inc.
410 Terry Avenue North
Seattle, WA 98109-5210
Cage Code: 66EB1
DUNS Number: 965048981
NAICS: 518210
We have just published a new AWS
Whitepaper on Minimizing Variable
Costs for Shared Data.
Download it at:
http://bit.ly/s3-requester-pays-open-data
10. Why does AWS care about open data?
Open data is data that can be used by anyone for any purpose for free.
Many of our customers rely on quality open data as much as they rely on
our computing, storage, and other web services.
10
11. Data on AWS
Amazon Web Services provides a comprehensive toolkit for gathering,
storing, analyzing, and working with data at any scale.
Amazon Elastic MapReduce
(Amazon EMR) provides the
Apache Hadoop analytics
framework as an easy-to-use
managed service.
Amazon S3 lets you store
and retrieve any amount of
data, at any time, from
anywhere on the web.
Amazon DynamoDB is a
fully-managed NoSQL
database service that makes
it cost-effective to store and
retrieve any amount of data.
11
12. 1-click deployment to launch, on
multiple regions around the world
Pay-as-you-go pricing
Advanced AnalyticsData Integration Analysis & Visualization
http://bit.ly/awsAnalytics
12
13. The power of open data in the cloud
Making data open on AWS enables more innovation by making data
available for rapid access to our flexible and low-cost computing
resources.
Amazon S3
Bucket
Amazon
EMR
Amazon
EC2
AWS
Lambda
Amazon
Redshift
Amazon
DynamoDB
13
14. Making data open on AWS enables more innovation by making data
available for rapid access to our flexible and low-cost computing
resources.
Amazon S3
Bucket
Amazon
EMR
Amazon
EC2
AWS
Lambda
Amazon
Redshift
Amazon
DynamoDB
The power of open data in the cloud
14
16. History of Innovation
AWS has been continually expanding its services to support virtually any cloud
workload, now offering more than 40 services.
Amazon S3
Amazon SQS
Amazon EC2
Amazon SimpleDB
Amazon EBS
Amazon CloudFront
Elastic Load
Balancing
Auto Scaling
Amazon VPC
Amazon RDS
Amazon SNS
AWS Identity and Access
Management
Amazon Route 53
Amazon SES
AWS Elastic Beanstalk
AWS CloudFormation
Amazon ElastiCache
AWS Direct Connect
AWS GovCloud
AWS Storage
Gateway
Amazon DynamoDB
Amazon CloudSearch
Amazon SWF
Amazon Glacier
Amazon Redshift
AWS Data Pipeline
Amazon Elastic
Transcoder
AWS OpsWorks
AWS CloudHSM
Amazon AppStream
AWS CloudTrail
Amazon WorkSpaces
Amazon Kinesis
Amazon ECS
Amazon Lambda
AWS Config
AWS CodeDeploy
Amazon RDS for Aurora
AWS KMS
Amazon Cognito
Amazon WorkDocs
AWS Directory Service
Amazon Mobile Analytics
2006 2007 2008 2009 2010 2011 2012 2013 2014 2015
Amazon EFS
Amazon WorkMail
Amazon Machine
Learning
16
17. AWS has announced price reductions 49* times since
our inception in 2006. Recent price drops included…
Amazon
ElastiCache
reduces
prices for
cache nodes
by an average
of 34%
March 26, 2014
34%
Amazon S3
reduces prices
for Standard
and Reduced
Redundancy
Storage, by an
average of 51%
March 26, 2014
51%
Amazon Route 53
lowers prices for
both standard
queries and
latency-based
routing queries
by 20%
July 31, 2014
20%
17
* As of June 2015
18. Open data as a platform
18
Data Enrichment
Sensemaking
Data Creation
Data at Rest
(Object storage)
Basic APIs
Complex APIs
Consumer
applications
Algorithmic
policy
Data-driven
journalism
Data Catalogs
Focused data
dashboards
Predictive
modeling
Visualizations
Lower cost of
knowledge
20. An Amazonian approach to open data
Two ideas that inform how we approach public data sets:
• Work backwards from the customer
• Eliminate undifferentiated heavy lifting
20
21. Working Backwards
• Think of data sets as products
• Seek out valuable data by listening to customer needs
• Consider real-world use cases for the data
• Consider the size of the user community or market
opportunity
21
22. Undifferentiated heavy lifting
“…data must be organized, well-documented, consistently
formatted, and error free. Cleaning the data is often the
most taxing part of data science, and is frequently 80% of
the work.”
— Data Driven by DJ Patil and Hilary Mason
22
23. Undifferentiated heavy lifting
“…data must be organized, well-documented, consistently
formatted, and error free. Cleaning the data is often the
most taxing part of data science, and is frequently 80% of
the work.”
— Data Driven by DJ Patil and Hilary Mason
We ask: How can we get rid of that 80%?
23
24. Public datasets on AWS
To enable more innovation, AWS hosts a selection of datasets that anyone
can access for free. Data in our public datasets is available for rapid
access to our flexible and low-cost computing resources.
Earth Science
Landsat on AWS
Life Sciences
1000 Genomes Project
Internet Science
Common Crawl Corpus
24
26. NEXRAD on AWS
The Next Generation Weather Radar (NEXRAD) is
a network of 160 high-resolution Doppler radar
sites that detects precipitation and atmospheric
movement and disseminates data in 5 minute
intervals from each site.
It has traditionally been time consuming and
expensive to acquire, store, and analyze NEXRAD
data. Accessing the full historical archive has been
impossible.
26
27. NEXRAD on AWS
NEXRAD
Sites
Public
Amazon S3 Bucket
Amazon
EC2
Public
Amazon S3 Bucket
Real-time
data chunks
Volume scan
file assembly
Continuously
updated archive
With NEXRAD on AWS, we provide an archive of individual volume scan files and
real-time chunks as objects in Amazon S3.
This allows the data to be accessed programmatically via a RESTful interface and
quickly deployed to any of our products for analysis and processing.
27
28. NEXRAD on AWS
Our collaborators, including Unidata, The Weather Company, NOAA, Climate
Corporation, and CartoDB, have provided early use cases and tutorials on how to
use this data in the cloud.
A wide range of users are interested in using NEXRAD on AWS for longitudinal
analysis, to study and visualize specific weather events, and develop new
products.
More info at http://aws.amazon.com/public-data-sets/nexrad
28
30. Landsat on AWS
We have committed to make up to 1
petabyte of Landsat imagery readily
available as objects on Amazon S3.
All Landsat 8 scenes from 2015 are
available, along with a selection of
cloud-free scenes from 2013 and
2014. All new Landsat 8 scenes are
made available each day (~700 per
day), often within hours of
production.
30
31. The Traditional Approach
Data is most commonly accessed via a
web interface and downloaded on
premises before being loaded into a
web server.
All bands are downloaded in a .tar
archive, even if you only need a few
bands.
Data acquisition is time consuming and
inherently redundant. Analysis is limited
by user’s access to bandwidth, storage,
memory, and processing power.
31
32. Landsat on AWS
Landsat on AWS makes each band of
each scene readily available as objects
on Amazon S3. Data can be accessed
programmatically via HTTP and quickly
deployed to any of our products for
analysis and processing.
Users do not need to worry about local
storage and have access to virtually
unlimited computing power on demand.
32
Amazon
EC2
s3://landsat-pds
.tarUSGS
.tiff
33. Undifferentiated heavy lifting
We use GDAL to add “internal tiling”
on each Landsat on AWS tiff, which
allows developers to use HTTP range
gets to access specific portions of
each scene.
This allows people to only access the
data they need when they need it. Standard tiff
object
Internal tiled tiff
object
1 2 3 4 5 6
7 8 9 10 11 12
13 14 15 16 17 18
19 20 21 22 23 24
25 26 27 28 29 30
31 32 33 34 35 36
1 2 3
4 5 6
7 8 9
10 11 12
13 14 15
16 17 18
19 20 21
22 23 24
25 26 27
28 29 30
31 32 33
34 35 36
33
36. Landsat on AWS
In the first 150 days (19 Mar – 16 Aug 2015)
• Over 200,000 scenes available
• Over 500 million hits globally
Image shows frequency of scene requests by
path/row.
White: ~100 requests
Orange: >300k requests
36
Visualization by Drew Bollinger
Development Seed
38. New SNS topic for Landsat on AWS
38
arn:aws:sns:us-west-2:274514004127:OpenObjectAddL8
You can now subscribe to a publicly available Amazon Simple Notification
Service (Amazon SNS) topic to be notified whenever a new batch of
Landsat scenes are available at s3://landsat-pds.
HELLO! **WELCOME!** I’m Jed Sundwall, and I lead AWS’s open data program. Each of you has received an email from me about this event, so it’s nice to meet you in person. I can’t express how excited we are to have you all here. This is the first big event the AWS Open Data team has put on and it’s very cool to see the response.
I want to take this time to tell you about why we’re here today, explain our open data program, and give you a few updates on our NEXRAD and Landsat Public Data Sets. Let’s go.
I love emoji.
6. First off, let’s thank our sponsors for making today possible. There are plenty of folks from Esri, DigitalGlobe, and Development Seed here, so please seek them out and thank them if you can. And I want to give a special thanks to Chief for sharing their venue with us. I’ve been a big fan of the team at Chief for years. They are fully committed to helping government take full advantage of technology, which is why they were the first people I reached out to when planning this event.
So why are we here? First off, we wanted to show off the state of the art of Earth data analysis, which is being done by our customers who you’ll hear from today. We also wanted to provide a venue for leaders at the Department of Interior and NOAA to connect with those of you who use their data. We also want this to be a networking event for all of you. You’ll notice a lot of breaks throughout the day. Those are there to give you ample opportunities to chat between sessions. We’ve also told all of the speakers that they get 30 minutes to present *and* take questions. We want this to be a very conversational event. One of the neat things about AWS is that we have over a million customers who use our services to do a wide variety of work. Hopefully you’ll get to learn about how someone outside of your field is working with Earth observation today. And last, but not least, we want to learn from you, particularly about what series AWS could provide to make your lives easier. Our product roadmap is largely based upon customer feedback, so please give us feedback! You should have a survey, so please fill that out, and feel free to cast your vote for your most wanted data on the window throughout the day.
8. Hot off the presses! In addition to this being our first event, we’ve also recently published our first Whitepaper, which is about understanding and mitigating the costs of using Amazon S3 to share data. Just email me for a copy or you can download it at this link.
OK. Let’s talk about why AWS has an open data program.
We use a very basic definition of open data. The beauty of open data – if it’s good – is that it gives people something to do with our computing and storage resources.
We have very significant customers who rely on open data for their work. Their products wouldn’t exist or wouldn’t be very useful without the availability of high quality, license-free, government information.
Open data matters to our customers, so it matters to us, and our goal, as a program, is to continually make more data more available to more people.
AWS provides a comprehensive toolkit for working with data at any scale. I’m not going to get into describing our products today, but I simply want to make the point that our platform is designed to store and and work with virtually any amount of data using our on-demand computing resources.
Through the AWS Marketplace we also provide a catalog of data analytics, integration, and visualization applications that you can quickly deploy in the cloud with pay-as-you-go pricing.
The value of open data on AWS is very basic: once the data is on S3, it becomes easier to work with using any of our computing and database resources.
This reduces the cost – in terms of dollars and time – of product development, of analysis, and of scientific discovery.
This is tremendously empowering for researchers, product developers, and analysts because it allows them to take their algorithms to the data rather than spending time and money to find, download, and store data in their own computing environments. When data is made available in the cloud, people can pay only for the computing resources they need to analyze the data and don’t have to worry about downloading and paying to store the data themselves.
The value of open data on AWS is very basic: once the data is on S3, it becomes easier to work with using any of our computing and database resources.
This reduces the cost – in terms of dollars and time – of product development, of analysis, and of scientific discovery.
We have over a million customers. This includes many many partners focused on the public sector. These are just part of a large and growing ecosystem of users who are advancing the state of the art of what can be done with data in the cloud. Of course, you’re going to hear from them today. They are the focus of this event.
Benefits are also realized as we continue to develop new services and features of our platform. We now offer more than 40 services to support virtually any kind of application in the cloud. As we add services, new applications of data become possible.
And finally, new possibilities are created as we lower prices. As of June this year, we’ve dropped our prices 49 times since AWS got started in 2006. The on-demand and pay-as-you-go nature of the cloud removes barriers for people to get started, and we continually pass whatever savings we can on to our customers to make it easier for them to access the IT resources they need.
Talk about temperature data.
Talk about lowering the cost of knowledge.
Talk about the great grain robbery. 10MM tons of grain sold to Russia in 1972.
Naturally, we have products that can help. LOLOLOLOL. Seriously, we do have products that help with creation of data, strong it, sharing it, analyzing it, and making it available through applications. Where the open data team focuses is in the bottom left there, which is the point from which data is shared. What can we do to ensure that shared data can realize its value? There are two very Amazonian ideas that guide us.
These concepts inform the way we think about a lot of our products and they can be applied to open data as well.
Working backwards from the customer is how we determine what to focus our energies on. A common practice throughout Amazon is to write a hypothetical press release as a first step when considering a new product. It challenges us to write in few words, and in plain language, why what we’re working on would matter and who it would matter to. If we can’t write a plausible press release, our efforts may be better spent elsewhere, or we need to try harder. The same thing can apply when prioritizing data sets. There is a lot of data out there, and it’s hard to tell which is most valuable. The working backwards approach can help with that.
Undifferentiated heavy lifting is all of the chores that keep you from focusing on what you’re good at and what your mission is. You’re not better at it than anyone else and it distracts from what you’ve actually set out to do. It’s one of the core benefits of the cloud, in that you can access as much storage and computing power as you need without needing to worry about ordering gear, setting it up, securing it, cooling it and worrying about running out of capacity. You just get what you need when you need it. The same thing applies to data. I love this quote from Data Driven (which is free on the Kindle, btw) by DJ Patil and Hilary Mason.
What we ask ourselves is what we can do to eliminate all of the work that gets in the way of actually analyzing data. That might mean making it analysis ready, but we’ve learned that we can make a huge difference by simply eliminating the need for people to download and store copies of it.
Our Public Data Sets Program is our lab where we can test out these theories and identify best practices for sharing data in the cloud. We select data sets by listening to our customers and when we foresee opportunities to make an impact on emerging sectors, as we did when we started hosting the 1,000 Genomes data set several years ago. Many of you are here today because of our Landsat and NEXRAD Public Data Sets.
We launched NEXRAD on AWS last month.
My DC-based counterpart, Ariel Gold, has managed the NEXRAD project throughout the year and she deserves a round of applause for this groundbreaking project – I don’t think it’s possible to overstate how hard she worked on this.
The Next Generation Weather Radar (NEXRAD) is a network of 160 high-resolution Doppler radar sites that detects precipitation and atmospheric movement and disseminates data in approximately 5 minute intervals from each site. NEXRAD enables severe storm prediction and is used by researchers and commercial enterprises to study and address the impact of weather across multiple sectors.
Traditionally, there has been no full historical archive available on demand to NOAA or external users and you have to wait 12 to 24 hours for new volume scan files.
A real-time feed of NEXRAD data and a full historical archive of original resolution (Level II) NEXRAD data, from June 1991 to present, is now freely available on Amazon S3 for anyone to use. This is the first time the full NEXRAD Level II archive has been accessible to the public on demand. Now anyone can use the data on-demand in the cloud without worrying about storage costs and download time. We are making NEXRAD data available as part of our research agreement with the US National Oceanic and Atmospheric Administration (NOAA) to enable new product development and analysis. Jeff De La Beujardier is going to talk about NOAA’s work to make its data more widely available later today.
NEXRAD on AWS is new, but we already know of one customer who was able to complete a project *two weeks* in advance because of the easy access to the data in the cloud. If you go to this URL, you’ll find documentation on how to access the data along with tutorials from Climate Corporation and CartoDB.
Landsat!
Just under a year ago, we announced that we would make up to 1PB of Landsat data available on AWS over a two year period as a contribution to the White House’s Climate Data Initiative. We launched Landsat on AWS on March 19th this year.
We had heard from multiple customers that we should look harder at Landsat, so we did. We spoke with a lot of customers who use Landsat data and learned that there was a lot we could do to help people work with it. One of the first things we learned was that acquiring data could be incredibly time consuming, particularly when trying to work with a lot of data. One big reason for this is that scenes are made available at 1GB .tar archives. To work with a scene, a researcher needs to download all 12 bands of that scene’s data even if they only need three or four scenes. That means that 60-70% of bandwidth (and therefore time spent downloading) is wasted
We subscribe to a feed of Landsat scenes from USGS to acquire all new scenes as they’re created. We then unpack the tarballs and use GDAL to apply some lossless compression and data optimization for each GeoTIFF of each band, which is then made available as a stand-alone object on S3. This means that researchers can access only the data they need when they need it.
Internal tiling!
By lowering the cost of accessing Landsat data, we’ve enabled more people to work with it in unexpected ways. I made the visualizations you see here on snapsat.org, which is a web app created by a group of students just a few weeks after we launched Landsat on AWS. It’s a remarkable web app that provides a very novel interface to browse the planet and create Landsat composites in seconds. I love it.
The Snapsat students were able to build it because they were able to access Landsat data without needing to download or copy it, which they wouldn’t have had the time or resources to do. If you have data to share, share it via URLs!
Landsat on AWS gets used a lot. Within the first 150 days, data was requested from the bucket over 500MM times. We don’t know how good that is because we’ve never released a comparable data set like this, but we know that a lot of people are accessing a lot of the data on purpose. As you’ll learn today, it’s being used for real work. It exceeded all of our expectations and it’s been amazing to learn about what people are doing with it.
Remember this thing? This is what it looks like in action. One copy of the data, made available near computing resources, and optimized for analysis. In very little time we’ve seen a lot of activity come out of it, and we couldn’t be happier about it. We’re here today to discover what’s next.
There’s one more thing. We now have a public SNS topic that you can use to subscribe to notifications whenever a new batch of scenes are available. You can use this to automatically trigger analysis of data from within AWS. I don’t expect you to copy it down, so just email me to get it.