Data lakes are transforming the way enterprises store, analyze, and learn insights from their data. While data lakes are a relatively new concept, many enterprises have already generated significant business value from the insights gleaned. In this session, AWS experts and technology leaders from Sysco, a Fortune 50 company and leader in food distribution and marketing, explain why Sysco decided to evolve its data management capabilities to include data lakes and how they customized them to support diverse querying capabilities and data science use cases. They also discuss how to architect different aspects of a data lake—ingestion from disparate sources, data consumption, and usability layers—and how to track data ingestion and consumption, monitor associated costs, enforce wanted levels of user access, manage data file formats, synchronize production and non-production environments, and maintain data integrity. Services to be discussed include Amazon S3 and S3 Select, Amazon Athena, Amazon EMR, Amazon EC2, and Amazon Redshift Spectrum.
This document outlines an agenda for a 90-minute workshop on Snowflake. The agenda includes introductions, an overview of Snowflake and data warehousing, demonstrations of how users utilize Snowflake, hands-on exercises loading sample data and running queries, and discussions of Snowflake architecture and capabilities. Real-world customer examples are also presented, such as a pharmacy building new applications on Snowflake and an education company using it to unify their data sources and achieve a 16x performance improvement.
How to build a data lake with aws glue data catalog (ABD213-R) re:Invent 2017Amazon Web Services
As data volumes grow and customers store more data on AWS, they often have valuable data that is not easily discoverable and available for analytics. The AWS Glue Data Catalog provides a central view of your data lake, making data readily available for analytics. We introduce key features of the AWS Glue Data Catalog and its use cases. Learn how crawlers can automatically discover your data, extract relevant metadata, and add it as table definitions to the AWS Glue Data Catalog. We will also explore the integration between AWS Glue Data Catalog and Amazon Athena, Amazon EMR, and Amazon Redshift Spectrum.
Apache Kafka and the Data Mesh | Ben Stopford and Michael Noll, ConfluentHostedbyConfluent
The document discusses the principles of a data mesh architecture using Apache Kafka for event streaming. It describes a data mesh as having four key principles: 1) domain-driven decentralization where each domain owns the data it creates, 2) treating data as a first-class product, 3) providing a self-serve data platform for easy access to real-time and historical data, and 4) establishing federated governance with global standards. Event streaming is presented as a good fit for data meshing due to its scalability, ability to handle real-time and historical data, and immutability. The document provides examples and recommendations for implementing each principle in a data mesh.
This document provides an overview and summary of the author's background and expertise. It states that the author has over 30 years of experience in IT working on many BI and data warehouse projects. It also lists that the author has experience as a developer, DBA, architect, and consultant. It provides certifications held and publications authored as well as noting previous recognition as an SQL Server MVP.
How will my on-premises data migrate to the cloud? How can I make it transparent to my users? Afterwards, how will on-premises and cloud data interact? In this session you will learn about the AWS Database Migration Service (DMS) and the AWS Schema Migration Tool (SCT). You can use these tools to convert your commercial database and database warehouse to open-source engines or AWS-native services, such as Amazon Aurora and Redshift.
Speakers:
Saurabh Saxena - Principal Technical Account Manager, AWS
Chris England - Sr. Technical Account Manager, AWS
Enterprise Architecture vs. Data ArchitectureDATAVERSITY
Enterprise Architecture (EA) provides a visual blueprint of the organization, and shows key interrelationships between data, process, applications, and more. By abstracting these assets in a graphical view, it’s possible to see key interrelationships, particularly as they relate to data and its business impact across the organization. Join us for a discussion on how data architecture is a key component of an overall enterprise architecture for enhanced business value and success.
Snowflake's Kent Graziano talks about what makes a data warehouse as a service and some of the key features of Snowflake's data warehouse as a service.
Snowflake: Your Data. No Limits (Session sponsored by Snowflake) - AWS Summit...Amazon Web Services
Snowflake is a cloud-based data warehouse that is built for the cloud. It was founded in 2012 and has raised $1 billion in funding. Snowflake's architecture separates storage, compute, and metadata services, allowing it to offer unlimited scalability, multiple clusters that can access shared data with no downtime, and full transactional consistency across the system. Snowflake has over 2000 customers including large enterprises that use it for analytics, data science, and sharing large volumes of data securely.
This document outlines an agenda for a 90-minute workshop on Snowflake. The agenda includes introductions, an overview of Snowflake and data warehousing, demonstrations of how users utilize Snowflake, hands-on exercises loading sample data and running queries, and discussions of Snowflake architecture and capabilities. Real-world customer examples are also presented, such as a pharmacy building new applications on Snowflake and an education company using it to unify their data sources and achieve a 16x performance improvement.
How to build a data lake with aws glue data catalog (ABD213-R) re:Invent 2017Amazon Web Services
As data volumes grow and customers store more data on AWS, they often have valuable data that is not easily discoverable and available for analytics. The AWS Glue Data Catalog provides a central view of your data lake, making data readily available for analytics. We introduce key features of the AWS Glue Data Catalog and its use cases. Learn how crawlers can automatically discover your data, extract relevant metadata, and add it as table definitions to the AWS Glue Data Catalog. We will also explore the integration between AWS Glue Data Catalog and Amazon Athena, Amazon EMR, and Amazon Redshift Spectrum.
Apache Kafka and the Data Mesh | Ben Stopford and Michael Noll, ConfluentHostedbyConfluent
The document discusses the principles of a data mesh architecture using Apache Kafka for event streaming. It describes a data mesh as having four key principles: 1) domain-driven decentralization where each domain owns the data it creates, 2) treating data as a first-class product, 3) providing a self-serve data platform for easy access to real-time and historical data, and 4) establishing federated governance with global standards. Event streaming is presented as a good fit for data meshing due to its scalability, ability to handle real-time and historical data, and immutability. The document provides examples and recommendations for implementing each principle in a data mesh.
This document provides an overview and summary of the author's background and expertise. It states that the author has over 30 years of experience in IT working on many BI and data warehouse projects. It also lists that the author has experience as a developer, DBA, architect, and consultant. It provides certifications held and publications authored as well as noting previous recognition as an SQL Server MVP.
How will my on-premises data migrate to the cloud? How can I make it transparent to my users? Afterwards, how will on-premises and cloud data interact? In this session you will learn about the AWS Database Migration Service (DMS) and the AWS Schema Migration Tool (SCT). You can use these tools to convert your commercial database and database warehouse to open-source engines or AWS-native services, such as Amazon Aurora and Redshift.
Speakers:
Saurabh Saxena - Principal Technical Account Manager, AWS
Chris England - Sr. Technical Account Manager, AWS
Enterprise Architecture vs. Data ArchitectureDATAVERSITY
Enterprise Architecture (EA) provides a visual blueprint of the organization, and shows key interrelationships between data, process, applications, and more. By abstracting these assets in a graphical view, it’s possible to see key interrelationships, particularly as they relate to data and its business impact across the organization. Join us for a discussion on how data architecture is a key component of an overall enterprise architecture for enhanced business value and success.
Snowflake's Kent Graziano talks about what makes a data warehouse as a service and some of the key features of Snowflake's data warehouse as a service.
Snowflake: Your Data. No Limits (Session sponsored by Snowflake) - AWS Summit...Amazon Web Services
Snowflake is a cloud-based data warehouse that is built for the cloud. It was founded in 2012 and has raised $1 billion in funding. Snowflake's architecture separates storage, compute, and metadata services, allowing it to offer unlimited scalability, multiple clusters that can access shared data with no downtime, and full transactional consistency across the system. Snowflake has over 2000 customers including large enterprises that use it for analytics, data science, and sharing large volumes of data securely.
The Zen of DataOps – AWS Lake Formation and the Data Supply Chain PipelineAmazon Web Services
Many organizations have adopted or are in the process of adopting DevOps methodologies in their quest to accelerate the delivery of software capabilities, features, and functionalities to support their organizational objectives. By applying the same practices, DataOps aims to provide the same level of agility in delivering data and information to the organization. AWS Lake Formation, in coordination with other AWS Services, enables DevOps methodologies to be realized through the Data Supply Chain Pipeline.
Recolectar y analizar grandes cantidades de datos se ha convertido en algo esencial para muchas organizaciones. El uso de Data Lakes se ha convertido en una popular estrategia para almacenar todo tipo de datos estructurados y no-estructurados, y centralizarlos en una única fuente. Únase a este webinar para descubrir cómo puede crear y administrar facilmente un data lake seguro usando servicios de AWS.
In this webinar you will learn how the AWS Database Migration Service (DMS) and AWS Schema Conversion Tool (SCT) can help migrate your databases to AWS for homogeneous and heterogeneous migrations. We will also discuss new sources and targets, together with new features that make DMS and SCT a powerful combination for both your database migration and data replication requirements.
This is a Level 100 webinar.
Speaker: Blair Layton, APAC Business Development, Database,
The document discusses building a data lake on AWS. It describes various AWS services that can be used to ingest, store, transform, analyze and visualize data in the data lake. These services include Amazon S3 for storage, AWS Glue for ETL/data cataloging, AWS Lake Formation for governance, Amazon Athena/EMR for analytics and Amazon QuickSight for visualization. The document also covers data movement options from on-premises to the data lake and real-time streaming of data using services like Kinesis. Machine learning workloads can leverage Amazon SageMaker for training and deployment.
Delta Lake brings reliability, performance, and security to data lakes. It provides ACID transactions, schema enforcement, and unified handling of batch and streaming data to make data lakes more reliable. Delta Lake also features lightning fast query performance through its optimized Delta Engine. It enables security and compliance at scale through access controls and versioning of data. Delta Lake further offers an open approach and avoids vendor lock-in by using open formats like Parquet that can integrate with various ecosystems.
This document discusses AWS Lake Formation and provides an overview of its key capabilities. It describes how Lake Formation can help users build clean and secure data lakes in days by automating tasks like data loading, cleaning, and governance that traditionally take months. It also outlines recent innovations to Lake Formation, including new features that enable row-level security, ACID transactions on data lakes, and acceleration of analytics on data stored in Amazon S3.
This session provides IT pros and application owners an overview of AWS options for building hybrid storage architectures or even entirely migrating datacenter storage to the AWS cloud. The AWS Storage Gateway connects existing on-premises block, file or tape storage systems to AWS cloud storage over the WAN in a hybrid model. The AWS Snow family of physical devices can capture, pre-process and migrate data into and out of AWS without any network connection at all. Join us to learn how you can close down datacenters, reduce storage footprints, and build solutions for tiering, data lakes, backup, disaster recovery, and migration.
This document discusses data governance challenges in the era of big data and proposes solutions. It begins by outlining the rise of data-driven businesses and the challenges they face with data quality, access, and trust issues. This has led to the rise of the Chief Data Officer role. The document then discusses how data governance approaches need to shift from hierarchical systems of record to more networked systems of engagement to manage expanding data volumes and types from sources like IoT and big data analytics. Key challenges discussed include digitalizing trust in data and addressing risks from opaque big data models. The document proposes taking a hybrid governance approach and implementing a system of record for data assets to provide findability, understandability and trust for all organizational data. Example use
Master the Multi-Clustered Data Warehouse - SnowflakeMatillion
Snowflake is one of the most powerful, efficient data warehouses on the market today—and we joined forces with the Snowflake team to show you how it works!
In this webinar:
- Learn how to optimize Snowflake
- Hear insider tips and tricks on how to improve performance
- Get expert insights from Craig Collier, Technical Architect from Snowflake, and Kalyan Arangam, Solution Architect from Matillion
- Find out how leading brands like Converse, Duo Security, and Pets at Home use Snowflake and Matillion ETL to make data-driven decisions
- Discover how Matillion ETL and Snowflake work together to modernize your data world
- Learn how to utilize the impressive scalability of Snowflake and Matillion
The document provides an overview of the Databricks platform, which offers a unified environment for data engineering, analytics, and AI. It describes how Databricks addresses the complexity of managing data across siloed systems by providing a single "data lakehouse" platform where all data and analytics workloads can be run. Key features highlighted include Delta Lake for ACID transactions on data lakes, auto loader for streaming data ingestion, notebooks for interactive coding, and governance tools to securely share and catalog data and models.
The document discusses the challenges of modern data, analytics, and AI workloads. Most enterprises struggle with siloed data systems that make integration and productivity difficult. The future of data lies with a data lakehouse platform that can unify data engineering, analytics, data warehousing, and machine learning workloads on a single open platform. The Databricks Lakehouse platform aims to address these challenges with its open data lake approach and capabilities for data engineering, SQL analytics, governance, and machine learning.
Snowflake: The Good, the Bad, and the UglyTyler Wishnoff
Learn how to solve the top 3 challenges Snowflake customers face, and what you can do to ensure high-performance, intelligent analytics at any scale. Ideal for those currently using Snowflake and those considering it. Learn more at: https://kyligence.io/
Build a simple data lake on AWS using a combination of services, including AWS Glue Data Catalog, AWS Glue Crawlers, AWS Glue Jobs, AWS Glue Studio, Amazon Athena, Amazon Relational Database Service (Amazon RDS), and Amazon S3.
Link to the blog post and video: https://garystafford.medium.com/building-a-simple-data-lake-on-aws-df21ca092e32
This document discusses how companies are increasingly data-centric and how data has become a strategic asset. It introduces several AWS database and data storage services like Amazon Aurora, DynamoDB, DocumentDB, ElastiCache, Neptune, Timestream, and QLDB. These services provide different data models and use cases like relational, key-value, document, in-memory, graph, time-series, and ledger data. The document highlights features of each service like performance, scalability, availability, security, and ease of use. It also discusses how the AWS Database Migration Service can help migrate databases to AWS.
How to Take Advantage of an Enterprise Data Warehouse in the CloudDenodo
Watch full webinar here: [https://buff.ly/2CIOtys]
As organizations collect increasing amounts of diverse data, integrating that data for analytics becomes more difficult. Technology that scales poorly and fails to support semi-structured data fails to meet the ever-increasing demands of today’s enterprise. In short, companies everywhere can’t consolidate their data into a single location for analytics.
In this Denodo DataFest 2018 session we’ll cover:
Bypassing the mandate of a single enterprise data warehouse
Modern data sharing to easily connect different data types located in multiple repositories for deeper analytics
How cloud data warehouses can scale both storage and compute, independently and elastically, to meet variable workloads
Presentation by Harsha Kapre, Snowflake
AIOps is becoming imperative to the management of today’s complex IT systems and their ability to support changing business conditions. This slide explains the role that AIOps can and will play in the enterprise of the future, how the scope of AIOps platforms will expand, and what new functionality may be deployed.
Watch the webinar here. https://www.moogsoft.com/resources/aiops/webinar/aiops-the-next-five-years
The document discusses building data lakes with AWS. It recommends using Amazon S3 as the storage layer for the data lake due to its scalability, durability and integration with other AWS analytics services. It also recommends using AWS Glue to catalog and ingest data into the data lake through automated crawlers. This allows for easy discovery, querying and analysis of data in the lake.
Databricks CEO Ali Ghodsi introduces Databricks Delta, a new data management system that combines the scale and cost-efficiency of a data lake, the performance and reliability of a data warehouse, and the low latency of streaming.
만들자! 데이터 기반의 스마트 팩토리 - 문태양 AWS 솔루션즈 아키텍트 / 배권 팀장, OCI 정보통신 :: AWS Summit Seou...Amazon Web Services Korea
제조 산업의 데이터는 내부 장치 및 장비에 담겨있기 때문에 활용되지 못하는 경우가 많습니다. AWS IoT로 산업 현장의 원격 감시 제어 데이터 (SCADA)를 수집하고 전사적 자원관리 (ERP), 제조 실행 시스템 (MES)의 데이터와 산업 현장의 데이터를 통합하여 대시보드에서 거의 실시간에 가까운 운영 메트릭을 모니터링하여 비즈니스 인사이트를 얻은 사례를 살펴봅니다.
The first step towards knowing your customer is to collect and extract insights and actionable information from your data. Learn how AWS enables you to cost efficiently store any amount of data and build an agile approach to data mining and visualization - helping you to make efficient business decisions and targeted offerings.
The first step towards knowing your customer is to collect and extract insights and actionable information from your data. Learn how AWS enables you to cost efficiently store any amount of data and build an agile approach to data mining and visualization - helping you to make efficient business decisions and targeted offerings.
The Zen of DataOps – AWS Lake Formation and the Data Supply Chain PipelineAmazon Web Services
Many organizations have adopted or are in the process of adopting DevOps methodologies in their quest to accelerate the delivery of software capabilities, features, and functionalities to support their organizational objectives. By applying the same practices, DataOps aims to provide the same level of agility in delivering data and information to the organization. AWS Lake Formation, in coordination with other AWS Services, enables DevOps methodologies to be realized through the Data Supply Chain Pipeline.
Recolectar y analizar grandes cantidades de datos se ha convertido en algo esencial para muchas organizaciones. El uso de Data Lakes se ha convertido en una popular estrategia para almacenar todo tipo de datos estructurados y no-estructurados, y centralizarlos en una única fuente. Únase a este webinar para descubrir cómo puede crear y administrar facilmente un data lake seguro usando servicios de AWS.
In this webinar you will learn how the AWS Database Migration Service (DMS) and AWS Schema Conversion Tool (SCT) can help migrate your databases to AWS for homogeneous and heterogeneous migrations. We will also discuss new sources and targets, together with new features that make DMS and SCT a powerful combination for both your database migration and data replication requirements.
This is a Level 100 webinar.
Speaker: Blair Layton, APAC Business Development, Database,
The document discusses building a data lake on AWS. It describes various AWS services that can be used to ingest, store, transform, analyze and visualize data in the data lake. These services include Amazon S3 for storage, AWS Glue for ETL/data cataloging, AWS Lake Formation for governance, Amazon Athena/EMR for analytics and Amazon QuickSight for visualization. The document also covers data movement options from on-premises to the data lake and real-time streaming of data using services like Kinesis. Machine learning workloads can leverage Amazon SageMaker for training and deployment.
Delta Lake brings reliability, performance, and security to data lakes. It provides ACID transactions, schema enforcement, and unified handling of batch and streaming data to make data lakes more reliable. Delta Lake also features lightning fast query performance through its optimized Delta Engine. It enables security and compliance at scale through access controls and versioning of data. Delta Lake further offers an open approach and avoids vendor lock-in by using open formats like Parquet that can integrate with various ecosystems.
This document discusses AWS Lake Formation and provides an overview of its key capabilities. It describes how Lake Formation can help users build clean and secure data lakes in days by automating tasks like data loading, cleaning, and governance that traditionally take months. It also outlines recent innovations to Lake Formation, including new features that enable row-level security, ACID transactions on data lakes, and acceleration of analytics on data stored in Amazon S3.
This session provides IT pros and application owners an overview of AWS options for building hybrid storage architectures or even entirely migrating datacenter storage to the AWS cloud. The AWS Storage Gateway connects existing on-premises block, file or tape storage systems to AWS cloud storage over the WAN in a hybrid model. The AWS Snow family of physical devices can capture, pre-process and migrate data into and out of AWS without any network connection at all. Join us to learn how you can close down datacenters, reduce storage footprints, and build solutions for tiering, data lakes, backup, disaster recovery, and migration.
This document discusses data governance challenges in the era of big data and proposes solutions. It begins by outlining the rise of data-driven businesses and the challenges they face with data quality, access, and trust issues. This has led to the rise of the Chief Data Officer role. The document then discusses how data governance approaches need to shift from hierarchical systems of record to more networked systems of engagement to manage expanding data volumes and types from sources like IoT and big data analytics. Key challenges discussed include digitalizing trust in data and addressing risks from opaque big data models. The document proposes taking a hybrid governance approach and implementing a system of record for data assets to provide findability, understandability and trust for all organizational data. Example use
Master the Multi-Clustered Data Warehouse - SnowflakeMatillion
Snowflake is one of the most powerful, efficient data warehouses on the market today—and we joined forces with the Snowflake team to show you how it works!
In this webinar:
- Learn how to optimize Snowflake
- Hear insider tips and tricks on how to improve performance
- Get expert insights from Craig Collier, Technical Architect from Snowflake, and Kalyan Arangam, Solution Architect from Matillion
- Find out how leading brands like Converse, Duo Security, and Pets at Home use Snowflake and Matillion ETL to make data-driven decisions
- Discover how Matillion ETL and Snowflake work together to modernize your data world
- Learn how to utilize the impressive scalability of Snowflake and Matillion
The document provides an overview of the Databricks platform, which offers a unified environment for data engineering, analytics, and AI. It describes how Databricks addresses the complexity of managing data across siloed systems by providing a single "data lakehouse" platform where all data and analytics workloads can be run. Key features highlighted include Delta Lake for ACID transactions on data lakes, auto loader for streaming data ingestion, notebooks for interactive coding, and governance tools to securely share and catalog data and models.
The document discusses the challenges of modern data, analytics, and AI workloads. Most enterprises struggle with siloed data systems that make integration and productivity difficult. The future of data lies with a data lakehouse platform that can unify data engineering, analytics, data warehousing, and machine learning workloads on a single open platform. The Databricks Lakehouse platform aims to address these challenges with its open data lake approach and capabilities for data engineering, SQL analytics, governance, and machine learning.
Snowflake: The Good, the Bad, and the UglyTyler Wishnoff
Learn how to solve the top 3 challenges Snowflake customers face, and what you can do to ensure high-performance, intelligent analytics at any scale. Ideal for those currently using Snowflake and those considering it. Learn more at: https://kyligence.io/
Build a simple data lake on AWS using a combination of services, including AWS Glue Data Catalog, AWS Glue Crawlers, AWS Glue Jobs, AWS Glue Studio, Amazon Athena, Amazon Relational Database Service (Amazon RDS), and Amazon S3.
Link to the blog post and video: https://garystafford.medium.com/building-a-simple-data-lake-on-aws-df21ca092e32
This document discusses how companies are increasingly data-centric and how data has become a strategic asset. It introduces several AWS database and data storage services like Amazon Aurora, DynamoDB, DocumentDB, ElastiCache, Neptune, Timestream, and QLDB. These services provide different data models and use cases like relational, key-value, document, in-memory, graph, time-series, and ledger data. The document highlights features of each service like performance, scalability, availability, security, and ease of use. It also discusses how the AWS Database Migration Service can help migrate databases to AWS.
How to Take Advantage of an Enterprise Data Warehouse in the CloudDenodo
Watch full webinar here: [https://buff.ly/2CIOtys]
As organizations collect increasing amounts of diverse data, integrating that data for analytics becomes more difficult. Technology that scales poorly and fails to support semi-structured data fails to meet the ever-increasing demands of today’s enterprise. In short, companies everywhere can’t consolidate their data into a single location for analytics.
In this Denodo DataFest 2018 session we’ll cover:
Bypassing the mandate of a single enterprise data warehouse
Modern data sharing to easily connect different data types located in multiple repositories for deeper analytics
How cloud data warehouses can scale both storage and compute, independently and elastically, to meet variable workloads
Presentation by Harsha Kapre, Snowflake
AIOps is becoming imperative to the management of today’s complex IT systems and their ability to support changing business conditions. This slide explains the role that AIOps can and will play in the enterprise of the future, how the scope of AIOps platforms will expand, and what new functionality may be deployed.
Watch the webinar here. https://www.moogsoft.com/resources/aiops/webinar/aiops-the-next-five-years
The document discusses building data lakes with AWS. It recommends using Amazon S3 as the storage layer for the data lake due to its scalability, durability and integration with other AWS analytics services. It also recommends using AWS Glue to catalog and ingest data into the data lake through automated crawlers. This allows for easy discovery, querying and analysis of data in the lake.
Databricks CEO Ali Ghodsi introduces Databricks Delta, a new data management system that combines the scale and cost-efficiency of a data lake, the performance and reliability of a data warehouse, and the low latency of streaming.
만들자! 데이터 기반의 스마트 팩토리 - 문태양 AWS 솔루션즈 아키텍트 / 배권 팀장, OCI 정보통신 :: AWS Summit Seou...Amazon Web Services Korea
제조 산업의 데이터는 내부 장치 및 장비에 담겨있기 때문에 활용되지 못하는 경우가 많습니다. AWS IoT로 산업 현장의 원격 감시 제어 데이터 (SCADA)를 수집하고 전사적 자원관리 (ERP), 제조 실행 시스템 (MES)의 데이터와 산업 현장의 데이터를 통합하여 대시보드에서 거의 실시간에 가까운 운영 메트릭을 모니터링하여 비즈니스 인사이트를 얻은 사례를 살펴봅니다.
The first step towards knowing your customer is to collect and extract insights and actionable information from your data. Learn how AWS enables you to cost efficiently store any amount of data and build an agile approach to data mining and visualization - helping you to make efficient business decisions and targeted offerings.
The first step towards knowing your customer is to collect and extract insights and actionable information from your data. Learn how AWS enables you to cost efficiently store any amount of data and build an agile approach to data mining and visualization - helping you to make efficient business decisions and targeted offerings.
A data lake is an architectural approach that allows you to store massive amounts of data into a central location, so it's readily available to be categorized, processed, analyzed and consumed by diverse groups within an organization.In this session, we will introduce the Data Lake concept and its implementation on AWS.We will explain the different roles our services play and how they fit into the Data Lake picture.
AWS Floor 28 - Building Data lake on AWSAdir Sharabi
AWS makes it easy to build and operate a highly scalable and flexible data platforms to collect, process, and analyze data so you can get timely insights and react quickly to new information. In this session we will talk about how to improve over time using your data. How do you take your everyday data and build relevant business insights, to help and continuously improve your business processes, and keep your innovation going based on your data.
Building a Data Lake in Amazon S3 & Amazon Glacier (STG401-R1) - AWS re:Inven...Amazon Web Services
The document discusses building a data lake using Amazon S3 and Amazon Glacier for storage. It covers topics like what is big data, what is a data lake, achievable business outcomes from a data lake, securing the data lake, and examples of what can be done with analytics services on AWS. The presentation provides examples of using services like Amazon Comprehend, Amazon Transcribe, Kinesis, Athena and QuickSight for natural language processing, audio analysis, real-time streaming and visualization.
Customizing Data Lakes to Work for Your Enterprise with Sysco (STG340) - AWS ...Amazon Web Services
Data lakes are helping enterprises of all sizes and industries make the most of their data. However, building a data lake requires consideration of your goals and an understanding of data lakes, including data ingestion, data consumption, and usability layers. In this chalk talk, AWS experts and representatives from Sysco, a Fortune 50 company and leader in food distribution and marketing, discuss parts of a data lake, design considerations, and the pros and cons of different architectural designs. They share guidance around data tracking, costs, user access, synchronization, and data integrity so that your data lake complies with governance requirements and works towards your data goals. Sysco representatives share their data lake experiences, best practices, and lessons learned. We highlight Amazon S3 and S3 Select, Amazon Athena, Amazon EMR, Amazon EC2, and Amazon Redshift Spectrum.
Connecting the dots - How Amazon Neptune and Graph Databases can transform yo...Amazon Web Services
This document discusses Amazon Neptune, a fully managed graph database service. It provides an overview of graph databases and their advantages over traditional databases for modeling connected data. It then describes Amazon Neptune's key features, like automatic scaling, high availability across Availability Zones, integration with open standards like Gremlin and SPARQL, and ease of use on AWS. Examples are given showing how to model and query graph data using Gremlin and SPARQL. Finally, it discusses Amazon Neptune's architecture and roadmap for general availability later in 2018.
Get to Know Your Customers - Build and Innovate with a Modern Data ArchitectureAmazon Web Services
Your customers probably want a better experience with your brand. Your different business teams want and need better insights in their decision making. Almost certainly, your finance and operations teams require this to happen at a fraction of the cost of traditional on-premises options. Modern data architectures on AWS help many of our best customers realise all of those goals. Your business data contains critical information about customer behaviours, operational decisions, and many factors that have financial impact on your organisation. Increasingly, this data sits beyond your transactional systems, and is too big, too fast, and too complex for existing systems to handle. AWS Data and Analytics services are designed from our customers' requirements to ingest, store, analyse, and consume information at record-breaking scale. In this session you will learn how these services work together to deliver business automation, enhance customer engagement and intelligence.
Modern Cloud Data Warehousing ft. Equinox Fitness Clubs: Optimize Analytics P...Amazon Web Services
Most companies are overrun with data, yet they lack critical insights to make timely and accurate business decisions. They are missing the opportunity to combine large amounts of new, unstructured big data that resides outside their data warehouse with trusted, structured data inside their data warehouse. In this session, we discuss the most common use cases with Amazon Redshift, and we take an in-depth look at how modern data warehousing blends and analyzes all your data to give you deeper insights to run your business. Equinox Fitness Clubs joins us to share their journey from static reports, redundant data, and inefficient data intergration to a modern and flexible data lake and data warehouse architecture that delivers dynamic reports based on trusted data.
Driving Machine Learning and Analytics Use Cases with AWS Storage (STG302) - ...Amazon Web Services
You’ve designed and built a well-architected data lake and ingested extreme amounts of structured and unstructured data. Now what? In this session, we explore real-world use cases where data scientists, developers, and researchers have discovered new and valuable ways to extract business insights using advanced analytics and machine learning. We review Amazon S3, Amazon Glacier, and Amazon EFS, the foundation for the analytics clusters and data engines. We also explore analytics tools and databases, including Amazon Redshift, Amazon Athena, Amazon EMR, Amazon QuickSight, Amazon Kinesis, Amazon RDS, and Amazon Aurora; and we review the AWS machine learning portfolio and AI services such as Amazon SageMaker, AWS Deep Learning AMIs, Amazon Rekognition, and Amazon Lex. We discuss how all of these pieces fit together to build intelligent applications.
In this session, we will focus on the data architecture of the application. Enabling different personas in the organization, (e.g., senior management, data scientists, data engineer, etc.), the ability to access relevant data points and produce valuable insights. We will understand key concepts and architectural components of a data lake architecture as well as how to build speed layer and batch layer data processing flows.
Supercell – Scaling Mobile Games (GAM301) - AWS re:Invent 2018Amazon Web Services
This document discusses Supercell's approach to scaling their mobile games and analytics infrastructure. Supercell has 5 games with hundreds of millions of active users. They use a microservices architecture and sharding to scale their games across thousands of EC2 instances. Their analytics pipeline collects terabytes of data daily, storing it in S3 and processing it with EMR. They have transitioned to separating compute and storage to better scale their analytics capabilities.
Using Big Data Retail to Build a Single View of Your Customer.pdfAmazon Web Services
A key challenge faced by retailers is how to form an integrated single view of their customers across multiple retail channels to better understand purchasing behaviours and patterns. In this session, we’ll present a solution that merges web analytics data with customer purchase history based on AWS API Gateway, Lambda, and S3. Learn how to track customer purchase behaviours across channels to better predict future needs and make relevant, intelligent recommendations.
Emerging Trends in Big Data, Analytics, Machine Learning, and Internet-of-Thi...Michaela Bromfield
This presentation was delivered on March 7, 2018 at Gartner's Data and Analytics Summit in Grapevine, TX. Rahul Pathak, GM at AWS discusses Next Gen Architecture on AWS.
AWS Analytics Services - When to use what? | AWS Summit Tel Aviv 2019AWS Summits
AWS provides a wide range of data analytics tools with the power to analyze vast volumes of customer, business, and transactional data quickly and at low cost.
In this session, we provide an overview of AWS analytics services and discuss how customers are using these services today. We will also discuss the new database and analytics services and features we launched in the last year.
AWS Analytics Services - When to use what? | AWS Summit Tel Aviv 2019Amazon Web Services
AWS provides a wide range of data analytics tools with the power to analyze vast volumes of customer, business, and transactional data quickly and at low cost.
In this session, we provide an overview of AWS analytics services and discuss how customers are using these services today. We will also discuss the new database and analytics services and features we launched in the last year.
SRV309 AWS Purpose-Built Database Strategy: The Right Tool for the Right JobAmazon Web Services
In this session, Shawn Bice, VP of NoSQL and QuickSight, covers the AWS purpose-built strategy for databases and explains why your application should drive the requirements of a database, not the other way around. We introduce AWS databases that are purpose-built for your application use cases. Learn why you should select different data services to solve different aspects of an application, and watch a demonstration on which application use cases lend themselves well to which data services. If you’re a developer building modern applications that require flexibility and consistent millisecond performance, and you’re trying to determine what relational and non-relational data services to use, this session is for you.
This document discusses big data and machine learning. It begins by defining big data using the 5 V's: volume, velocity, variety, veracity, and value. It then discusses challenges organizations face with big data, including which tools to use and determining what data they have. The remainder discusses how to gain business value from data through architectures like data lakes, analytics, and machine learning services on AWS. It provides an example of how Netflix evolved its data pipeline and emphasizes agility. Finally, it discusses how machine learning relies on big data and new tools are needed for data scientists.
Similar to Building a Data Lake for Your Enterprise, ft. Sysco (STG309) - AWS re:Invent 2018 (20)
Come costruire servizi di Forecasting sfruttando algoritmi di ML e deep learn...Amazon Web Services
Il Forecasting è un processo importante per tantissime aziende e viene utilizzato in vari ambiti per cercare di prevedere in modo accurato la crescita e distribuzione di un prodotto, l’utilizzo delle risorse necessarie nelle linee produttive, presentazioni finanziarie e tanto altro. Amazon utilizza delle tecniche avanzate di forecasting, in parte questi servizi sono stati messi a disposizione di tutti i clienti AWS.
In questa sessione illustreremo come pre-processare i dati che contengono una componente temporale e successivamente utilizzare un algoritmo che a partire dal tipo di dato analizzato produce un forecasting accurato.
Big Data per le Startup: come creare applicazioni Big Data in modalità Server...Amazon Web Services
La varietà e la quantità di dati che si crea ogni giorno accelera sempre più velocemente e rappresenta una opportunità irripetibile per innovare e creare nuove startup.
Tuttavia gestire grandi quantità di dati può apparire complesso: creare cluster Big Data su larga scala sembra essere un investimento accessibile solo ad aziende consolidate. Ma l’elasticità del Cloud e, in particolare, i servizi Serverless ci permettono di rompere questi limiti.
Vediamo quindi come è possibile sviluppare applicazioni Big Data rapidamente, senza preoccuparci dell’infrastruttura, ma dedicando tutte le risorse allo sviluppo delle nostre le nostre idee per creare prodotti innovativi.
Ora puoi utilizzare Amazon Elastic Kubernetes Service (EKS) per eseguire pod Kubernetes su AWS Fargate, il motore di elaborazione serverless creato per container su AWS. Questo rende più semplice che mai costruire ed eseguire le tue applicazioni Kubernetes nel cloud AWS.In questa sessione presenteremo le caratteristiche principali del servizio e come distribuire la tua applicazione in pochi passaggi
Vent'anni fa Amazon ha attraversato una trasformazione radicale con l'obiettivo di aumentare il ritmo dell'innovazione. In questo periodo abbiamo imparato come cambiare il nostro approccio allo sviluppo delle applicazioni ci ha permesso di aumentare notevolmente l'agilità, la velocità di rilascio e, in definitiva, ci ha consentito di creare applicazioni più affidabili e scalabili. In questa sessione illustreremo come definiamo le applicazioni moderne e come la creazione di app moderne influisce non solo sull'architettura dell'applicazione, ma sulla struttura organizzativa, sulle pipeline di rilascio dello sviluppo e persino sul modello operativo. Descriveremo anche approcci comuni alla modernizzazione, compreso l'approccio utilizzato dalla stessa Amazon.com.
Come spendere fino al 90% in meno con i container e le istanze spot Amazon Web Services
L’utilizzo dei container è in continua crescita.
Se correttamente disegnate, le applicazioni basate su Container sono molto spesso stateless e flessibili.
I servizi AWS ECS, EKS e Kubernetes su EC2 possono sfruttare le istanze Spot, portando ad un risparmio medio del 70% rispetto alle istanze On Demand. In questa sessione scopriremo insieme quali sono le caratteristiche delle istanze Spot e come possono essere utilizzate facilmente su AWS. Impareremo inoltre come Spreaker sfrutta le istanze spot per eseguire applicazioni di diverso tipo, in produzione, ad una frazione del costo on-demand!
In recent months, many customers have been asking us the question – how to monetise Open APIs, simplify Fintech integrations and accelerate adoption of various Open Banking business models. Therefore, AWS and FinConecta would like to invite you to Open Finance marketplace presentation on October 20th.
Event Agenda :
Open banking so far (short recap)
• PSD2, OB UK, OB Australia, OB LATAM, OB Israel
Intro to Open Finance marketplace
• Scope
• Features
• Tech overview and Demo
The role of the Cloud
The Future of APIs
• Complying with regulation
• Monetizing data / APIs
• Business models
• Time to market
One platform for all: a Strategic approach
Q&A
Rendi unica l’offerta della tua startup sul mercato con i servizi Machine Lea...Amazon Web Services
Per creare valore e costruire una propria offerta differenziante e riconoscibile, le startup di successo sanno come combinare tecnologie consolidate con componenti innovativi creati ad hoc.
AWS fornisce servizi pronti all'utilizzo e, allo stesso tempo, permette di personalizzare e creare gli elementi differenzianti della propria offerta.
Concentrandoci sulle tecnologie di Machine Learning, vedremo come selezionare i servizi di intelligenza artificiale offerti da AWS e, anche attraverso una demo, come costruire modelli di Machine Learning personalizzati utilizzando SageMaker Studio.
OpsWorks Configuration Management: automatizza la gestione e i deployment del...Amazon Web Services
Con l'approccio tradizionale al mondo IT per molti anni è stato difficile implementare tecniche di DevOps, che finora spesso hanno previsto attività manuali portando di tanto in tanto a dei downtime degli applicativi interrompendo l'operatività dell'utente. Con l'avvento del cloud, le tecniche di DevOps sono ormai a portata di tutti a basso costo per qualsiasi genere di workload, garantendo maggiore affidabilità del sistema e risultando in dei significativi miglioramenti della business continuity.
AWS mette a disposizione AWS OpsWork come strumento di Configuration Management che mira ad automatizzare e semplificare la gestione e i deployment delle istanze EC2 per mezzo di workload Chef e Puppet.
Scopri come sfruttare AWS OpsWork a garanzia e affidabilità del tuo applicativo installato su Instanze EC2.
Microsoft Active Directory su AWS per supportare i tuoi Windows WorkloadsAmazon Web Services
Vuoi conoscere le opzioni per eseguire Microsoft Active Directory su AWS? Quando si spostano carichi di lavoro Microsoft in AWS, è importante considerare come distribuire Microsoft Active Directory per supportare la gestione, l'autenticazione e l'autorizzazione dei criteri di gruppo. In questa sessione, discuteremo le opzioni per la distribuzione di Microsoft Active Directory su AWS, incluso AWS Directory Service per Microsoft Active Directory e la distribuzione di Active Directory su Windows su Amazon Elastic Compute Cloud (Amazon EC2). Trattiamo argomenti quali l'integrazione del tuo ambiente Microsoft Active Directory locale nel cloud e l'utilizzo di applicazioni SaaS, come Office 365, con AWS Single Sign-On.
Dal riconoscimento facciale al riconoscimento di frodi o difetti di fabbricazione, l'analisi di immagini e video che sfruttano tecniche di intelligenza artificiale, si stanno evolvendo e raffinando a ritmi elevati. In questo webinar esploreremo le possibilità messe a disposizione dai servizi AWS per applicare lo stato dell'arte delle tecniche di computer vision a scenari reali.
Amazon Web Services e VMware organizzano un evento virtuale gratuito il prossimo mercoledì 14 Ottobre dalle 12:00 alle 13:00 dedicato a VMware Cloud ™ on AWS, il servizio on demand che consente di eseguire applicazioni in ambienti cloud basati su VMware vSphere® e di accedere ad una vasta gamma di servizi AWS, sfruttando a pieno le potenzialità del cloud AWS e tutelando gli investimenti VMware esistenti.
Molte organizzazioni sfruttano i vantaggi del cloud migrando i propri carichi di lavoro Oracle e assicurandosi notevoli vantaggi in termini di agilità ed efficienza dei costi.
La migrazione di questi carichi di lavoro, può creare complessità durante la modernizzazione e il refactoring delle applicazioni e a questo si possono aggiungere rischi di prestazione che possono essere introdotti quando si spostano le applicazioni dai data center locali.
Crea la tua prima serverless ledger-based app con QLDB e NodeJSAmazon Web Services
Molte aziende oggi, costruiscono applicazioni con funzionalità di tipo ledger ad esempio per verificare lo storico di accrediti o addebiti nelle transazioni bancarie o ancora per tenere traccia del flusso supply chain dei propri prodotti.
Alla base di queste soluzioni ci sono i database ledger che permettono di avere un log delle transazioni trasparente, immutabile e crittograficamente verificabile, ma sono strumenti complessi e onerosi da gestire.
Amazon QLDB elimina la necessità di costruire sistemi personalizzati e complessi fornendo un database ledger serverless completamente gestito.
In questa sessione scopriremo come realizzare un'applicazione serverless completa che utilizzi le funzionalità di QLDB.
Con l’ascesa delle architetture di microservizi e delle ricche applicazioni mobili e Web, le API sono più importanti che mai per offrire agli utenti finali una user experience eccezionale. In questa sessione impareremo come affrontare le moderne sfide di progettazione delle API con GraphQL, un linguaggio di query API open source utilizzato da Facebook, Amazon e altro e come utilizzare AWS AppSync, un servizio GraphQL serverless gestito su AWS. Approfondiremo diversi scenari, comprendendo come AppSync può aiutare a risolvere questi casi d’uso creando API moderne con funzionalità di aggiornamento dati in tempo reale e offline.
Inoltre, impareremo come Sky Italia utilizza AWS AppSync per fornire aggiornamenti sportivi in tempo reale agli utenti del proprio portale web.
Database Oracle e VMware Cloud™ on AWS: i miti da sfatareAmazon Web Services
Molte organizzazioni sfruttano i vantaggi del cloud migrando i propri carichi di lavoro Oracle e assicurandosi notevoli vantaggi in termini di agilità ed efficienza dei costi.
La migrazione di questi carichi di lavoro, può creare complessità durante la modernizzazione e il refactoring delle applicazioni e a questo si possono aggiungere rischi di prestazione che possono essere introdotti quando si spostano le applicazioni dai data center locali.
In queste slide, gli esperti AWS e VMware presentano semplici e pratici accorgimenti per facilitare e semplificare la migrazione dei carichi di lavoro Oracle accelerando la trasformazione verso il cloud, approfondiranno l’architettura e dimostreranno come sfruttare a pieno le potenzialità di VMware Cloud ™ on AWS.
1) The document discusses building a minimum viable product (MVP) using Amazon Web Services (AWS).
2) It provides an example of an MVP for an omni-channel messenger platform that was built from 2017 to connect ecommerce stores to customers via web chat, Facebook Messenger, WhatsApp, and other channels.
3) The founder discusses how they started with an MVP in 2017 with 200 ecommerce stores in Hong Kong and Taiwan, and have since expanded to over 5000 clients across Southeast Asia using AWS for scaling.
This document discusses pitch decks and fundraising materials. It explains that venture capitalists will typically spend only 3 minutes and 44 seconds reviewing a pitch deck. Therefore, the deck needs to tell a compelling story to grab their attention. It also provides tips on tailoring different types of decks for different purposes, such as creating a concise 1-2 page teaser, a presentation deck for pitching in-person, and a more detailed read-only or fundraising deck. The document stresses the importance of including key information like the problem, solution, product, traction, market size, plans, team, and ask.
This document discusses building serverless web applications using AWS services like API Gateway, Lambda, DynamoDB, S3 and Amplify. It provides an overview of each service and how they can work together to create a scalable, secure and cost-effective serverless application stack without having to manage servers or infrastructure. Key services covered include API Gateway for hosting APIs, Lambda for backend logic, DynamoDB for database needs, S3 for static content, and Amplify for frontend hosting and continuous deployment.
This document provides tips for fundraising from startup founders Roland Yau and Sze Lok Chan. It discusses generating competition to create urgency for investors, fundraising in parallel rather than sequentially, having a clear fundraising narrative focused on what you do and why it's compelling, and prioritizing relationships with people over firms. It also notes how the pandemic has changed fundraising, with examples of deals done virtually during this time. The tips emphasize being fully prepared before fundraising and cultivating connections with investors in advance.
AWS_HK_StartupDay_Building Interactive websites while automating for efficien...Amazon Web Services
This document discusses Amazon's machine learning services for building conversational interfaces and extracting insights from unstructured text and audio. It describes Amazon Lex for creating chatbots, Amazon Comprehend for natural language processing tasks like entity extraction and sentiment analysis, and how they can be used together for applications like intelligent call centers and content analysis. Pre-trained APIs simplify adding machine learning to apps without requiring ML expertise.
Amazon Elastic Container Service (Amazon ECS) è un servizio di gestione dei container altamente scalabile, che semplifica la gestione dei contenitori Docker attraverso un layer di orchestrazione per il controllo del deployment e del relativo lifecycle. In questa sessione presenteremo le principali caratteristiche del servizio, le architetture di riferimento per i differenti carichi di lavoro e i semplici passi necessari per poter velocemente migrare uno o più dei tuo container.