SlideShare a Scribd company logo
1 of 59
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Databases onAWS: HowToChoose
The Right Database
Randall Hunt
Software Engineer at AWS
@jrhunt
randhunt@amazon
Markus Ostertag
CEO at Team Internet AG
@Osterjour
Dr. Sebastian Brandt
Senior Key Expert - Knowledge
Graph at Siemens CT
S U M M I T © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
1970 1980 1990 2000
Oracle DB2
SQL Server
MySQL
PostgreSQL
DynamoDB
Redis
MongoDB
Elasticsearch
Neptune
CassandraAccess
Aurora
2010
Timestream
QLDB
DocumentDB
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Modern apps havemodern requirements
Users: 1 million+
Data volume: TB–PB–EB
Locality: Global
Performance: Milliseconds–microseconds
Request rate: Millions per second
Access: Web, Mobile, IoT, Devices
Scale: Up-down, Out-in
Economics: Pay for what you use
Developer access: No assembly requiredSocial mediaRide hailing Media streaming Dating
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Common datacategories and usecases
Relational
Referential
integrity, ACID
transactions,
schema-
on-write
Lift and shift, ERP,
CRM, finance
Key-value
High
throughput, low-
latency reads
and writes,
endless scale
Real-time bidding,
shopping cart,
social, product
catalog, customer
preferences
Document
Store
documents and
quickly query
on any attribute
Content
management,
personalization,
mobile
In-memory
Query by key
with
microsecond
latency
Leaderboards,
real-time analytics,
caching
Graph
Quickly and
easily create
and navigate
relationships
between
data
Fraud detection,
social networking,
recommendation
engine
Time-series
Collect, store,
and process
data sequenced
by time
IoT applications,
event tracking
Ledger
Complete,
immutable, and
verifiable history
of all changes to
application data
Systems
of record, supply
chain, health care,
registrations,
financial
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
AWS: Purpose-built databases
Relational Key-value Document In-memory Graph Search
Amazon
DynamoDB
Amazon
Neptune
Amazon
RDS
Aurora CommercialCommunity
Amazon
ElastiCache
Amazon
Elasticsearch
Service
Amazon
DocumentDB
Time-series Ledger
Amazon
Timestream
Amazon
Quantum
Ledger
DatabaseMemcachedRedi
s
S U M M I T © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Amazon RelationalDatabaseService(RDS)
Managed relational database service with a choice of six popular database engines
Easy to administer Available and durable Highly scalable Fast and secure
No need for infrastructure
provisioning, installing, and
maintaining DB software
Automatic Multi-AZ data
replication; automated backup,
snapshots, failover
Scale database compute
and storage with a few
clicks with no app
downtime
SSD storage and guaranteed
provisioned I/O; data
encryption at rest and in
transit
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
TraditionalSQL
• TCP based wire protocol
• Well Known, lots of uses
• Common drivers (JDBC)
• Frequently used with ORMs
• Scale UP individual instances
• Scale OUT with read replicas
• Sharding at application level
• Lots of flavors but very similar
language
• Joins
INSERT INTO table1
(id, first_name, last_name)
VALUES (1, Randall’, Hunt’);
SELECT col1, col2, col3
FROM table1
WHERE col4 = 1 AND col5 = 2
GROUP BY col1
HAVING count(*) > 1
ORDER BY col2
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
AmazonAurora
MySQL and PostgreSQL-compatible relational database built for the cloud
Performance and availability of commercial-grade databases at 1/10th the cost
Performance
and scalability
Availability
and durability
Highly secure Fully managed
5x throughput of standard MySQL
and 3x of standard PostgreSQL;
scale-out up to
15 read replicas
Fault-tolerant, self-healing
storage; six copies of data
across three Availability Zones;
continuous backup to Amazon S3
Network isolation,
encryption at rest/transit
Managed by RDS:
No hardware provisioning,
software patching, setup,
configuration, or backups
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
SQLvsNoSQL
Optimized for storage Optimized for compute
Normalized/relational Denormalized/hierarchical
Ad hoc queries Instantiated views
Scale vertically Scale horizontally
Good for OLAP Built for OLTP at scale
SQL NoSQL
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Amazon DynamoDB
Fast and flexible key value database service for any scale
Comprehensive
security
Encrypts all data by default
and fully integrates with
AWS Identity and Access
Management for robust
security
Performance at scale
Consistent, single-digit
millisecond response times at any
scale; build applications with
virtually unlimited throughput
Global database for
global users and apps
Build global applications with fast
access to local data by easily
replicating tables across multiple
AWS Regions
Serverless
No server provisioning,
software patching, or upgrades;
scales up or down
automatically; continuously
backs up your data
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Amazon DynamoDB DataStructure
Table
Items
Attributes
Partition
Key
Sort
Key
Mandatory
Key-value access pattern
Determines data distribution
Optional
Model 1:N relationships
Enables rich query capabilities
All items for key
==, <, >, >=, <=
“begins with”
“between”
“contains”
“in”
sorted results
counts
top/bottom N values
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
DynamoDBSchemaandQueries
• Connects over HTTP
• Global Secondary Indexes and Local
Secondary Indexes
• Speed up queries with DAX
• Global tables (mutli-region-multi-master)
• Transactions across multiple tables
• Change Streams
• Rich query language with expressions
• Provision read and write capacity units
separately with or without autoscaling
• Also supports pay per request model
import boto3
votes_table =
boto3.resource('dynamodb').Table('votes')
resp = votes_table.update_item(
Key={'name': editor},
UpdateExpression="ADD votes :incr",
ExpressionAttributeValues={":incr": 1},
ReturnValues="ALL_NEW"
)
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
DynamoDB
Advancements over the last 22 months
VPC
endpoints
April 2017
Auto
scaling
June 2017
DynamoDB
Accelerator (DAX)
April 2017
Time To
Live (TTL)
February 2017
Global tables
On-demand
backup
Encryption at rest
November 2017 November 2017 November 2017
Point-in-time
recovery
March 2018
SLA
June 2018
99.999% SLA
August 2018
Adaptive
capacity ACID
November 2018
Transactions
November 2018
On-demand
Amazon DocumentDB: Modern cloud-native
architecture
Decouple
compute and
storage
Distribute data in
smaller partitions
Increase the
replication of
data (6x)
What would you do to improve scalability and availability?
1 2 3
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
AmazonDocumentDB
Fast,scalable,andfullymanagedMongoDB-compatibledatabaseservice
Fast Scalable Fully managed MongoDB
compatible
Millions of requests per second
with millisecond latency; twice
the throughput of MongoDB
Separation of compute and
storage enables both layers
to scale independently; scale
out to 15 read replicas in
minutes
Managed by AWS:
no hardware provisioning;
auto patching, quick setup,
secure, and automatic
backups
Compatible with MongoDB 3.6;
use the same SDKs, tools, and
applications with Amazon
DocumentDB
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Document databases
• Data is stored in JSON-like
documents
• Documents map naturally to
how humans model data
• Flexible schema and
indexing
• Expressive query language
built for documents (ad hoc
queries and aggregations)
JSON documents are first-class objects
of the database
{
id: 1,
name: "sue",
age: 26,
email: "sue@example.com",
promotions: ["new user", "5%", "dog lover"],
memberDate: 2018-2-22,
shoppingCart: [
{product:"abc", quantity:2, cost:19.99},
{product:"edf", quantity:3, cost: 2.99}
]
}
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Amazon ElastiCache
Redis and Memcached compatible, in-memory data store and cache
Secure and
reliable
Network isolation,
encryption at rest/transit,
HIPAA, PCI, FedRAMP,
multi AZ, and automatic
failover
Redis & Memcached
compatible
Fully compatible with open source
Redis and Memcached
Easily scalable
Scale writes and reads with
sharding and replicas
Extreme
performance
In-memory data store and
cache for microsecond
response times
S U M M I T © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Redis
• Redis Serialization Protocol over TCP (RESP)
• Supports Strings, Hashes, Lists, Sets, and Sorted Sets
• Simple commands for manipulating in memory data structures
• Pub/Sub features
• Supports clustering via partitions
S U M M I T © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Memcached
• Text and Binary protocols over TCP
• Small number of commands: set, add, replace, append, prepend, cas, get, gets, delete, incr,decr
• Client/Application based partitioning
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Relationshipsenable newapplications
Retail fraud detectionRestaurant recommendationsSocial networks
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Usecasesfor highlyconnected data
Social networking
Life Sciences Network & IT operationsFraud detection
Recommendations Knowledge graphs
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Different approaches for highlyconnected data
Purpose-built for a business process
Purpose-built to answer questions about
relationships
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Amazon Neptune
Fully managed graph database
Easy
Build powerful queries
easily with Gremlin and
SPARQL
Fast
Query billions of relationships
with millisecond latency
Open
Supports Apache TinkerPop &
W3C RDF graph models
Reliable
Six replicas of your data across
three AZs with full backup and
restore
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Open Source Apache TinkerPop™
Gremlin Traversal Language
W3C Standard
SPARQL Query Language
R E S O U R C E D E S C R I P T I O N
F R A M E W O R K ( R D F )
P R O P E R T Y G R A P H
LEADING GRAPH MODELS AND
FRAMEWORKS
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
AmazonNeptunehighlevelarchitecture
Bulk load
from S3
Database
Mgmt.
S U M M I T © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Gremlin
g.addV('person').property(id, 1).property('name', 'randall')
g.V('1').property(single, 'age', 27)
g.addV('person').property(id, 2).property('name', 'markus')
g.addE('knows').from(g.V('1')).to(g.V('2')).property('weight', 1.0)
g.V().hasLabel('person')
g.V().has('name', 'randall').out('knows').valueMap()
http://tinkerpop.apache.org/docs/current/reference/#graph-traversal-steps
S U M M I T © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
SPARQLand RDF
https://www.w3.org/TR/sparql11-query/
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
THE BENEFIT OF URIs: LINKED DATA
Linking across datasets by referencing globally unique URIs
GeoNames
Wikidata
PermID
Example: PermID (re)uses <http://sws.geonames.org/6252001/>
as a global Identifier for the USA, which is an identifier rooted in GeoNames.
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
AmazonTimestream(signup for thepreview)
Fast, scalable, fully managed time-series database
1,000x faster and 1/10th the
cost of relational databases
Collect data at the rate of
millions of inserts per
second (10M/second)
Trillions of
daily events
Adaptive query processing
engine maintains steady,
predictable performance
Time-series
analytics
Built-in functions for
interpolation, smoothing, and
approximation
Serverless
Automated setup, configuration,
server provisioning, software
patching
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
AmazonQuantumLedger Database(QLDB)(Preview)
Fully managed ledger database
Track and verify history of all changes made to your application’s data
Immutable
Maintains a sequenced record of
all changes to your data, which
cannot be deleted or modified;
you have the ability to query and
analyze the full history
Cryptographically
verifiable
Uses cryptography to
generate a secure output
file of your data’s history
Easy to use
Easy to use, letting you
use familiar database
capabilities like SQL APIs
for querying the data
Highly scalable
Executes 2–3X as many
transactions than ledgers
in common blockchain
frameworks
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
AWS DatabaseMigrationService(AWS DMS)
M I G R A T I N G
D A T A B A S E S
T O A W S
Migrate between on-premises and AWS
Migrate between databases
Automated schema conversion
Data replication for
zero-downtime migration
100,000+
databases migrated
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Customer
Premises,
EC2, RDS
Application Users
Aurora
PostgreSQLVPN/Network
AWS DMS—Logical replication
Start a replication instance
Connect to source and target databases
Select tables, schemas, or databases
Let the AWS Database Migration
Service create tables and load data
Uses change data capture to keep
them in sync
Switch applications over to the target
at your convenience
S U M M I T © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Customers aremoving toAWSDatabases
Verizon is migrating over 1,000 business-critical applications and database backend systems to AWS,
several of which also include the migration of production databases to Amazon Aurora.
Wappa migrated from their Oracle database to Amazon Aurora and improved their reporting
time per user by 75 percent.
Trimble migrated their Oracle databases to Amazon RDS and project they will pay about 1/4th of
what they paid when managing their private infrastructure.
Intuit migrated from Microsoft SQL Server to Amazon Redshift to reduce data-processing timelines
and get insights to decision makers faster and more frequently.
Equinox Fitness migrated its Teradata on-premises data warehouse to Amazon Redshift. They went from
static reports to a modern data lake that delivers dynamic reports.
Eventbrite moved from Cloudera to Amazon EMR and were able to cut costs dramatically, spinning
clusters up/down on-demand and using Spot (saving > 80%) and Reserved Instances.
By December 2018, Amazon.com will have migrated 88% of their Oracle DBs (and 97% of critical
system DBs) moved to Amazon Aurora and Amazon DynamoDB. They also migrated their 50 PB
Oracle Data Warehouse to AWS (Amazon S3, Amazon Redshift, and Amazon EMR).
Samsung Electronics migrated their Cassandra clusters to Amazon DynamoDB for their Samsung
Cloud workload with 70% cost savings.
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Mostenterprisedatabase& analyticscloud customers
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Moststartupdatabase&analyticscloud customers
S U M M I T © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Problem: Matching IPs toRanges inReal-Time Bidding
Real-Time Bidding Requirements:
• High throughput (>25000 req/s)
• Response time is critical
• Scalability
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Possiblechoices
Relational
Well known,
Flexible Indices,
Query
language,
Data scheme
obvious
Latency?
Query volume?
Future scale?
Key-value
High throughput,
low-latency
reads,
endless scale
Data scheme?
Indices?
Sharding?
In-memory
Query by key
with
microsecond
latency
Data scheme?
Sharding?
Future scale?
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Possiblechoices
Relational
Well known,
Flexible Indices,
Query
language,
Data scheme
obvious
Latency?
Query volume?
Future scale?
Key-value
High throughput,
low-latency
reads,
endless scale
Data scheme?
Indices?
Sharding?
In-memory
Query by key
with
microsecond
latency
Data scheme?
Sharding?
Future scale?
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
DynamoDB?
But how?
• Sharding and Index downsides addressed by transforming data during
ingest
• Accept multiple entries per range -> sharding & query possible/optimized
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
DynamoDB itis!
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
DynamoDB itis!
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
DynamoDB itis!
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Learnings
• Focus on “core requirements“
• Be creative about data schemes
• Accept minor downsides – very often there is no perfect solution
• Use upsides of chosen database to optimize (i.e. adding DAX for
caching)
Knowledge Graphs
@ Siemens
Sebastian Brandt
Steffen Lamparter
Semantics and Reasoning Group
Siemens Corporate Technology
CT RDA BAM SMR-DE
February 2019
Unrestricted © Siemens AG 2019siemens.com
Dr. Sebastian Brandt, Dr. Steffen Lamparter, Siemens Corporate Technology Siemens AG 2019
Knowledge graphs become especially powerful for
managing complex queries and heterogeneous data
Why Knowledge Graphs?
• Graphs are a natural way to represent
entities and their relationships
• Graphs can capture a broad spectrum
of data (structured / unstructured)
• Graphs can be managed efficiently
Intuitive domain modelling
Flexibility & performance
Low up-front investment
Robust data quality assurance
Game-changing data integration
Dr. Sebastian Brandt, Dr. Steffen Lamparter, Siemens Corporate Technology Siemens AG 2019
Semantic descriptions
The semantic description indicates the
meaning of an object or relation, e.g. Joe
Kaeser is a person
What are Graphs? Knowledge representation formalism
semantic descriptions of entities and their relationships
Relationships
Logical connection between two objects
e.g. Joe Kaeser is born in Arnbruck
Rules make it possible to add further expert knowledge, e.g.
"Siemens has to be a company, as a person is working there"
Objects
Real-world objects (things, places,
people) and abstract concepts (genres,
religions, professions)
Is a
Is born on
Was founded by
Is headquartered in
Munich, Berlin
Was promoted
June 23, 1957
Is a
Is working at
Roesemarie
Kaeser
Is married to
Werner v. Siemens and
Friedrich Halske
August 1, 2013
Person
Company
Has pictures
Is born in
Arnbruck
Joe Kaeser
Dr. Sebastian Brandt, Dr. Steffen Lamparter, Siemens Corporate Technology Siemens AG 2019
Knowledge graphs become a powerful addition to traditional data
warehouses for managing heterogeneous data with complex relations
Dimensions
of data relations
High number & complex
relations, e.g. social networks
Limited relations across data,
e.g. time series data
Variety of
questions to be
answered
"We do not yet know all the
different questions, users will
ask"  Chatbot
"We have a clear scope of
user questions"  fixed set of
dashboard needed
Data heterogeneity
Analysis needs to combine he-
terogeneous data, e.g. sensor
data, text data, product data
Only few different data types
and data sources e.g.
invoicing database in ERP
Quality of data
Need to handle imperfect,
incomplete or inconsistent
data
Ensured consistency of stored
data
low high
low high
low high
low high
Traditional data warehouses
Knowledge graphs
Dr. Sebastian Brandt, Dr. Steffen Lamparter, Siemens Corporate Technology Siemens AG 2019
Use cases for knowledge graphs can be clustered into
five categories – overview and use case examples
Data quality Digital companion
Improving data availability
and quality by combining
and comparing data from
various sources to fill in
missing data sets or
identify potentially wrong
data and data duplicates
Enhancing features of
existing products or
services with digital
companions that are able
to understand and
process user questions
and providing the needed
data insights
Data access &
dashboarding
Maintaining up-to-date
meta-data, creating
transparency on all
available data and
making them accessible
to users via queries
Recommender
system
Providing users high
quality recommendations
by identifying similarities
in historical data
Constraints &
planning
Enabling autonomous
systems
to understand data and its
dependencies and take
own decisions, such as
autonomous planning of
production proces-ses
Degree of complexity
Dr. Sebastian Brandt, Dr. Steffen Lamparter, Siemens Corporate Technology Siemens AG 2019
Example: gas-turbine maintenance planning
Maintenance Repair
Time-series
Power turbine Configuration
Fleet browsing and search
DB Docs CSV, XML,
Excel, …
Siemens Knowledge Graph
Analytics for R&D• Cross-life-cycle data integration at PS DO:
gas-turbines, including maintenance, repair,
monitoring, and configuration data
• Holistic engineering-centric domain model
• Intuitive graph queries independent of
source data schemata
• Virtual integration of time-series data
DB
Maintenance planning
Dr. Sebastian Brandt, Dr. Steffen Lamparter, Siemens Corporate Technology Siemens AG 2019
A semantic data model enables flexible linking and an integrated,
intuitive API for applications
Semantic Data Model
Planning
tools
Simulation
tool
Building
Operations
Security aaS Building
Performance
Service Portal Location Ba-
sed Services
Digital Lifecycle Platform
Drive
applications
BT Knowledge
Graph
Customer Data
Weather Data
Public Energy Data
MindSphereBuilding Structure
BIM
Product Data
www
Data APIProduct APIBuilding API
Dr. Sebastian Brandt, Dr. Steffen Lamparter, Siemens Corporate Technology Siemens AG 2019
Creating perfect places based on Services –
a user-centric holistic approach to the modern workplace …
Customer Interest Relevant KPI’s
Cost per space unit
Employee satisfaction
CO2 emissions
Employee productivity
Optimizing CAPEX and OPEX
Energy and
asset efficiency
Space
efficiency
Individual efficiency
and comfort
Workplace Utilization
Revenue per space unit
Vacancy Rate
Asset Performance/Useful Life
Dr. Sebastian Brandt, Dr. Steffen Lamparter, Siemens Corporate Technology Siemens AG 2019
Industrializing Knowledge Graphs
R&D AreasRelevant technologies
Decision Making
• Explanation of AI decisions
• Data access: Semantic Search
• Machine Learning on Graphs for
recommendations, quality, etc.
Decision Making
• Reasoning and Constraint Solving
• Machine/Deep Learning
• Question Answering
Storage and Integration
• Reusable Semantic Modelling
and Knowledge Graphs
• Data integration and cleaning (e.g.
entity reconciliation)
Storage and Integration
• Graph/NoSQL databases
• Constraints and Rules
• Probabilistic programming
• Ontologies
Generation
• Extraction from unstructured data
(inclusive text, audio, image)
• Automatic semantic annotation
of structured data
• Learning of domain-specific
rules/patterns
Generation
• NLP/Text understanding
• Machine/Deep Learning
• Computer vision
• Sound recognition
• Virtual data Integration
• Information retrieval
• …
Industrial Knowledge Graph
Decision
Making
Storage and
Integration
Generation
Knowledge
Automation
HumansMachines
ML for
Graphs
Ontology
Library
ML for
automated
annotation
Simple
RDF +
PGM
Dr. Sebastian Brandt, Dr. Steffen Lamparter, Siemens Corporate Technology Siemens AG 2019
Get in touch!
Dr. Sebastian Brandt
Senior Key Expert – Knowledge Graph and Data Management
CT RDA BAM SMR-DE
Dr. Steffen Lamparter
Head of Research Group Semantics & Reasoning
CT RDA BAM SMR-DE
Siemens AG
Corporate Technology
Otto-Hahn-Ring 6
81739 München Germany
E-mail
steffen.lamparter@siemens.com
Intranet
intranet.siemens.com/ct
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T
Thank you!
S U M M I T © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
© 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I TS U M M I T © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.

More Related Content

What's hot

Data & Analytics ReInvent Recap [AWS Basel Meetup - Jan 2023].pdf
Data & Analytics ReInvent Recap [AWS Basel Meetup - Jan 2023].pdfData & Analytics ReInvent Recap [AWS Basel Meetup - Jan 2023].pdf
Data & Analytics ReInvent Recap [AWS Basel Meetup - Jan 2023].pdfChris Bingham
 
Neptune, the Graph Database | AWS Floor28
Neptune, the Graph Database | AWS Floor28Neptune, the Graph Database | AWS Floor28
Neptune, the Graph Database | AWS Floor28Amazon Web Services
 
The Zen of DataOps – AWS Lake Formation and the Data Supply Chain Pipeline
The Zen of DataOps – AWS Lake Formation and the Data Supply Chain PipelineThe Zen of DataOps – AWS Lake Formation and the Data Supply Chain Pipeline
The Zen of DataOps – AWS Lake Formation and the Data Supply Chain PipelineAmazon Web Services
 
Building End-to-End Delta Pipelines on GCP
Building End-to-End Delta Pipelines on GCPBuilding End-to-End Delta Pipelines on GCP
Building End-to-End Delta Pipelines on GCPDatabricks
 
Building the Data Lake with Azure Data Factory and Data Lake Analytics
Building the Data Lake with Azure Data Factory and Data Lake AnalyticsBuilding the Data Lake with Azure Data Factory and Data Lake Analytics
Building the Data Lake with Azure Data Factory and Data Lake AnalyticsKhalid Salama
 
Big Data Analytics Architectural Patterns and Best Practices (ANT201-R1) - AW...
Big Data Analytics Architectural Patterns and Best Practices (ANT201-R1) - AW...Big Data Analytics Architectural Patterns and Best Practices (ANT201-R1) - AW...
Big Data Analytics Architectural Patterns and Best Practices (ANT201-R1) - AW...Amazon Web Services
 
Data Platform Architecture Principles and Evaluation Criteria
Data Platform Architecture Principles and Evaluation CriteriaData Platform Architecture Principles and Evaluation Criteria
Data Platform Architecture Principles and Evaluation CriteriaScyllaDB
 
Emerging Trends in Data Architecture – What’s the Next Big Thing?
Emerging Trends in Data Architecture – What’s the Next Big Thing?Emerging Trends in Data Architecture – What’s the Next Big Thing?
Emerging Trends in Data Architecture – What’s the Next Big Thing?DATAVERSITY
 
An Overview of Best Practices for Large Scale Migrations - AWS Transformation...
An Overview of Best Practices for Large Scale Migrations - AWS Transformation...An Overview of Best Practices for Large Scale Migrations - AWS Transformation...
An Overview of Best Practices for Large Scale Migrations - AWS Transformation...Amazon Web Services
 
Work Backwards to Your Graph Data Model & Queries with Amazon Neptune (DAT330...
Work Backwards to Your Graph Data Model & Queries with Amazon Neptune (DAT330...Work Backwards to Your Graph Data Model & Queries with Amazon Neptune (DAT330...
Work Backwards to Your Graph Data Model & Queries with Amazon Neptune (DAT330...Amazon Web Services
 
Balance agility and governance with #TrueDataOps and The Data Cloud
Balance agility and governance with #TrueDataOps and The Data CloudBalance agility and governance with #TrueDataOps and The Data Cloud
Balance agility and governance with #TrueDataOps and The Data CloudKent Graziano
 
Using MLOps to Bring ML to Production/The Promise of MLOps
Using MLOps to Bring ML to Production/The Promise of MLOpsUsing MLOps to Bring ML to Production/The Promise of MLOps
Using MLOps to Bring ML to Production/The Promise of MLOpsWeaveworks
 
Best Practices in DataOps: How to Create Agile, Automated Data Pipelines
Best Practices in DataOps: How to Create Agile, Automated Data PipelinesBest Practices in DataOps: How to Create Agile, Automated Data Pipelines
Best Practices in DataOps: How to Create Agile, Automated Data PipelinesEric Kavanagh
 
Getting Started with AWS Database Migration Service
Getting Started with AWS Database Migration ServiceGetting Started with AWS Database Migration Service
Getting Started with AWS Database Migration ServiceAmazon Web Services
 
Building a Modern Data Platform in the Cloud
Building a Modern Data Platform in the CloudBuilding a Modern Data Platform in the Cloud
Building a Modern Data Platform in the CloudAmazon Web Services
 
Free Training: How to Build a Lakehouse
Free Training: How to Build a LakehouseFree Training: How to Build a Lakehouse
Free Training: How to Build a LakehouseDatabricks
 
Architect’s Open-Source Guide for a Data Mesh Architecture
Architect’s Open-Source Guide for a Data Mesh ArchitectureArchitect’s Open-Source Guide for a Data Mesh Architecture
Architect’s Open-Source Guide for a Data Mesh ArchitectureDatabricks
 

What's hot (20)

Data & Analytics ReInvent Recap [AWS Basel Meetup - Jan 2023].pdf
Data & Analytics ReInvent Recap [AWS Basel Meetup - Jan 2023].pdfData & Analytics ReInvent Recap [AWS Basel Meetup - Jan 2023].pdf
Data & Analytics ReInvent Recap [AWS Basel Meetup - Jan 2023].pdf
 
Neptune, the Graph Database | AWS Floor28
Neptune, the Graph Database | AWS Floor28Neptune, the Graph Database | AWS Floor28
Neptune, the Graph Database | AWS Floor28
 
The Zen of DataOps – AWS Lake Formation and the Data Supply Chain Pipeline
The Zen of DataOps – AWS Lake Formation and the Data Supply Chain PipelineThe Zen of DataOps – AWS Lake Formation and the Data Supply Chain Pipeline
The Zen of DataOps – AWS Lake Formation and the Data Supply Chain Pipeline
 
Building End-to-End Delta Pipelines on GCP
Building End-to-End Delta Pipelines on GCPBuilding End-to-End Delta Pipelines on GCP
Building End-to-End Delta Pipelines on GCP
 
Building the Data Lake with Azure Data Factory and Data Lake Analytics
Building the Data Lake with Azure Data Factory and Data Lake AnalyticsBuilding the Data Lake with Azure Data Factory and Data Lake Analytics
Building the Data Lake with Azure Data Factory and Data Lake Analytics
 
Big Data Analytics Architectural Patterns and Best Practices (ANT201-R1) - AW...
Big Data Analytics Architectural Patterns and Best Practices (ANT201-R1) - AW...Big Data Analytics Architectural Patterns and Best Practices (ANT201-R1) - AW...
Big Data Analytics Architectural Patterns and Best Practices (ANT201-R1) - AW...
 
Ml ops on AWS
Ml ops on AWSMl ops on AWS
Ml ops on AWS
 
Data Platform Architecture Principles and Evaluation Criteria
Data Platform Architecture Principles and Evaluation CriteriaData Platform Architecture Principles and Evaluation Criteria
Data Platform Architecture Principles and Evaluation Criteria
 
Emerging Trends in Data Architecture – What’s the Next Big Thing?
Emerging Trends in Data Architecture – What’s the Next Big Thing?Emerging Trends in Data Architecture – What’s the Next Big Thing?
Emerging Trends in Data Architecture – What’s the Next Big Thing?
 
An Overview of Best Practices for Large Scale Migrations - AWS Transformation...
An Overview of Best Practices for Large Scale Migrations - AWS Transformation...An Overview of Best Practices for Large Scale Migrations - AWS Transformation...
An Overview of Best Practices for Large Scale Migrations - AWS Transformation...
 
Work Backwards to Your Graph Data Model & Queries with Amazon Neptune (DAT330...
Work Backwards to Your Graph Data Model & Queries with Amazon Neptune (DAT330...Work Backwards to Your Graph Data Model & Queries with Amazon Neptune (DAT330...
Work Backwards to Your Graph Data Model & Queries with Amazon Neptune (DAT330...
 
Balance agility and governance with #TrueDataOps and The Data Cloud
Balance agility and governance with #TrueDataOps and The Data CloudBalance agility and governance with #TrueDataOps and The Data Cloud
Balance agility and governance with #TrueDataOps and The Data Cloud
 
Using MLOps to Bring ML to Production/The Promise of MLOps
Using MLOps to Bring ML to Production/The Promise of MLOpsUsing MLOps to Bring ML to Production/The Promise of MLOps
Using MLOps to Bring ML to Production/The Promise of MLOps
 
Best Practices in DataOps: How to Create Agile, Automated Data Pipelines
Best Practices in DataOps: How to Create Agile, Automated Data PipelinesBest Practices in DataOps: How to Create Agile, Automated Data Pipelines
Best Practices in DataOps: How to Create Agile, Automated Data Pipelines
 
Machine Learning on AWS
Machine Learning on AWSMachine Learning on AWS
Machine Learning on AWS
 
Cloud Migration Workshop
Cloud Migration WorkshopCloud Migration Workshop
Cloud Migration Workshop
 
Getting Started with AWS Database Migration Service
Getting Started with AWS Database Migration ServiceGetting Started with AWS Database Migration Service
Getting Started with AWS Database Migration Service
 
Building a Modern Data Platform in the Cloud
Building a Modern Data Platform in the CloudBuilding a Modern Data Platform in the Cloud
Building a Modern Data Platform in the Cloud
 
Free Training: How to Build a Lakehouse
Free Training: How to Build a LakehouseFree Training: How to Build a Lakehouse
Free Training: How to Build a Lakehouse
 
Architect’s Open-Source Guide for a Data Mesh Architecture
Architect’s Open-Source Guide for a Data Mesh ArchitectureArchitect’s Open-Source Guide for a Data Mesh Architecture
Architect’s Open-Source Guide for a Data Mesh Architecture
 

Similar to How to Choose The Right Database on AWS - Berlin Summit - 2019

Choosing the Right Database (Database Freedom)
Choosing the Right Database (Database Freedom)Choosing the Right Database (Database Freedom)
Choosing the Right Database (Database Freedom)Amazon Web Services
 
Building with Purpose-Built Databases: Match Your Workload to the Right Database
Building with Purpose-Built Databases: Match Your Workload to the Right DatabaseBuilding with Purpose-Built Databases: Match Your Workload to the Right Database
Building with Purpose-Built Databases: Match Your Workload to the Right DatabaseAmazon Web Services
 
Build data-driven, high-performance, internet-scale applications with AWS dat...
Build data-driven, high-performance, internet-scale applications with AWS dat...Build data-driven, high-performance, internet-scale applications with AWS dat...
Build data-driven, high-performance, internet-scale applications with AWS dat...Amazon Web Services
 
All Databases Are Equal, But Some Databases Are More Equal than Others: How t...
All Databases Are Equal, But Some Databases Are More Equal than Others: How t...All Databases Are Equal, But Some Databases Are More Equal than Others: How t...
All Databases Are Equal, But Some Databases Are More Equal than Others: How t...javier ramirez
 
Using AWS Purpose-Built Databases to Modernize your Applications
Using AWS Purpose-Built Databases to Modernize your ApplicationsUsing AWS Purpose-Built Databases to Modernize your Applications
Using AWS Purpose-Built Databases to Modernize your ApplicationsAmazon Web Services
 
Databases - Choosing the right Database on AWS
Databases - Choosing the right Database on AWSDatabases - Choosing the right Database on AWS
Databases - Choosing the right Database on AWSAmazon Web Services
 
Building with Purpose - Built Databases: Match Your Workloads to the Right Da...
Building with Purpose - Built Databases: Match Your Workloads to the Right Da...Building with Purpose - Built Databases: Match Your Workloads to the Right Da...
Building with Purpose - Built Databases: Match Your Workloads to the Right Da...Amazon Web Services
 
AWS Database and Analytics State of the Union
AWS Database and Analytics State of the UnionAWS Database and Analytics State of the Union
AWS Database and Analytics State of the UnionAmazon Web Services
 
Everything You Need to Know About Big Data: From Architectural Principles to ...
Everything You Need to Know About Big Data: From Architectural Principles to ...Everything You Need to Know About Big Data: From Architectural Principles to ...
Everything You Need to Know About Big Data: From Architectural Principles to ...Amazon Web Services
 
AWS Database and Analytics State of the Union
AWS Database and Analytics State of the UnionAWS Database and Analytics State of the Union
AWS Database and Analytics State of the UnionAmazon Web Services
 
Enterprise-Database-Migration-Strategies-and-Options-on-AWS
Enterprise-Database-Migration-Strategies-and-Options-on-AWSEnterprise-Database-Migration-Strategies-and-Options-on-AWS
Enterprise-Database-Migration-Strategies-and-Options-on-AWSAmazon Web Services
 
Database su AWS scegliere lo strumento giusto per il giusto obiettivo
Database su AWS scegliere lo strumento giusto per il giusto obiettivoDatabase su AWS scegliere lo strumento giusto per il giusto obiettivo
Database su AWS scegliere lo strumento giusto per il giusto obiettivoAmazon Web Services
 
Building low latency apps with a serverless architecture and in-memory data I...
Building low latency apps with a serverless architecture and in-memory data I...Building low latency apps with a serverless architecture and in-memory data I...
Building low latency apps with a serverless architecture and in-memory data I...AWS Germany
 
AWS SSA Webinar 32 - Getting Started with databases on AWS: Choosing the righ...
AWS SSA Webinar 32 - Getting Started with databases on AWS: Choosing the righ...AWS SSA Webinar 32 - Getting Started with databases on AWS: Choosing the righ...
AWS SSA Webinar 32 - Getting Started with databases on AWS: Choosing the righ...Cobus Bernard
 
Building real-time applications with Amazon ElastiCache - ADB204 - Anaheim AW...
Building real-time applications with Amazon ElastiCache - ADB204 - Anaheim AW...Building real-time applications with Amazon ElastiCache - ADB204 - Anaheim AW...
Building real-time applications with Amazon ElastiCache - ADB204 - Anaheim AW...Amazon Web Services
 
How-to-Choose-the-Right-Database-to-Build-High-Performance-Internet-Scale-App...
How-to-Choose-the-Right-Database-to-Build-High-Performance-Internet-Scale-App...How-to-Choose-the-Right-Database-to-Build-High-Performance-Internet-Scale-App...
How-to-Choose-the-Right-Database-to-Build-High-Performance-Internet-Scale-App...Amazon Web Services
 
Databases - EBC on the road Brazil Edition [Portuguese]
Databases - EBC on the road Brazil Edition [Portuguese]Databases - EBC on the road Brazil Edition [Portuguese]
Databases - EBC on the road Brazil Edition [Portuguese]Amazon Web Services
 
Migrate and Modernize Your Database
Migrate and Modernize Your DatabaseMigrate and Modernize Your Database
Migrate and Modernize Your DatabaseAmazon Web Services
 
Implementing a Data Warehouse on AWS in a Hybrid Environment
Implementing a Data Warehouse on AWS in a Hybrid EnvironmentImplementing a Data Warehouse on AWS in a Hybrid Environment
Implementing a Data Warehouse on AWS in a Hybrid EnvironmentAmazon Web Services
 

Similar to How to Choose The Right Database on AWS - Berlin Summit - 2019 (20)

Choosing the Right Database (Database Freedom)
Choosing the Right Database (Database Freedom)Choosing the Right Database (Database Freedom)
Choosing the Right Database (Database Freedom)
 
Building with Purpose-Built Databases: Match Your Workload to the Right Database
Building with Purpose-Built Databases: Match Your Workload to the Right DatabaseBuilding with Purpose-Built Databases: Match Your Workload to the Right Database
Building with Purpose-Built Databases: Match Your Workload to the Right Database
 
Build data-driven, high-performance, internet-scale applications with AWS dat...
Build data-driven, high-performance, internet-scale applications with AWS dat...Build data-driven, high-performance, internet-scale applications with AWS dat...
Build data-driven, high-performance, internet-scale applications with AWS dat...
 
All Databases Are Equal, But Some Databases Are More Equal than Others: How t...
All Databases Are Equal, But Some Databases Are More Equal than Others: How t...All Databases Are Equal, But Some Databases Are More Equal than Others: How t...
All Databases Are Equal, But Some Databases Are More Equal than Others: How t...
 
Using AWS Purpose-Built Databases to Modernize your Applications
Using AWS Purpose-Built Databases to Modernize your ApplicationsUsing AWS Purpose-Built Databases to Modernize your Applications
Using AWS Purpose-Built Databases to Modernize your Applications
 
Databases - Choosing the right Database on AWS
Databases - Choosing the right Database on AWSDatabases - Choosing the right Database on AWS
Databases - Choosing the right Database on AWS
 
Building with Purpose - Built Databases: Match Your Workloads to the Right Da...
Building with Purpose - Built Databases: Match Your Workloads to the Right Da...Building with Purpose - Built Databases: Match Your Workloads to the Right Da...
Building with Purpose - Built Databases: Match Your Workloads to the Right Da...
 
AWS Database and Analytics State of the Union
AWS Database and Analytics State of the UnionAWS Database and Analytics State of the Union
AWS Database and Analytics State of the Union
 
Everything You Need to Know About Big Data: From Architectural Principles to ...
Everything You Need to Know About Big Data: From Architectural Principles to ...Everything You Need to Know About Big Data: From Architectural Principles to ...
Everything You Need to Know About Big Data: From Architectural Principles to ...
 
AWS Database and Analytics State of the Union
AWS Database and Analytics State of the UnionAWS Database and Analytics State of the Union
AWS Database and Analytics State of the Union
 
Managed NoSQL databases
Managed NoSQL databasesManaged NoSQL databases
Managed NoSQL databases
 
Enterprise-Database-Migration-Strategies-and-Options-on-AWS
Enterprise-Database-Migration-Strategies-and-Options-on-AWSEnterprise-Database-Migration-Strategies-and-Options-on-AWS
Enterprise-Database-Migration-Strategies-and-Options-on-AWS
 
Database su AWS scegliere lo strumento giusto per il giusto obiettivo
Database su AWS scegliere lo strumento giusto per il giusto obiettivoDatabase su AWS scegliere lo strumento giusto per il giusto obiettivo
Database su AWS scegliere lo strumento giusto per il giusto obiettivo
 
Building low latency apps with a serverless architecture and in-memory data I...
Building low latency apps with a serverless architecture and in-memory data I...Building low latency apps with a serverless architecture and in-memory data I...
Building low latency apps with a serverless architecture and in-memory data I...
 
AWS SSA Webinar 32 - Getting Started with databases on AWS: Choosing the righ...
AWS SSA Webinar 32 - Getting Started with databases on AWS: Choosing the righ...AWS SSA Webinar 32 - Getting Started with databases on AWS: Choosing the righ...
AWS SSA Webinar 32 - Getting Started with databases on AWS: Choosing the righ...
 
Building real-time applications with Amazon ElastiCache - ADB204 - Anaheim AW...
Building real-time applications with Amazon ElastiCache - ADB204 - Anaheim AW...Building real-time applications with Amazon ElastiCache - ADB204 - Anaheim AW...
Building real-time applications with Amazon ElastiCache - ADB204 - Anaheim AW...
 
How-to-Choose-the-Right-Database-to-Build-High-Performance-Internet-Scale-App...
How-to-Choose-the-Right-Database-to-Build-High-Performance-Internet-Scale-App...How-to-Choose-the-Right-Database-to-Build-High-Performance-Internet-Scale-App...
How-to-Choose-the-Right-Database-to-Build-High-Performance-Internet-Scale-App...
 
Databases - EBC on the road Brazil Edition [Portuguese]
Databases - EBC on the road Brazil Edition [Portuguese]Databases - EBC on the road Brazil Edition [Portuguese]
Databases - EBC on the road Brazil Edition [Portuguese]
 
Migrate and Modernize Your Database
Migrate and Modernize Your DatabaseMigrate and Modernize Your Database
Migrate and Modernize Your Database
 
Implementing a Data Warehouse on AWS in a Hybrid Environment
Implementing a Data Warehouse on AWS in a Hybrid EnvironmentImplementing a Data Warehouse on AWS in a Hybrid Environment
Implementing a Data Warehouse on AWS in a Hybrid Environment
 

More from Randall Hunt

WhereML a Serverless ML Powered Location Guessing Twitter Bot
WhereML a Serverless ML Powered Location Guessing Twitter BotWhereML a Serverless ML Powered Location Guessing Twitter Bot
WhereML a Serverless ML Powered Location Guessing Twitter BotRandall Hunt
 
Randall's re:Invent Recap
Randall's re:Invent RecapRandall's re:Invent Recap
Randall's re:Invent RecapRandall Hunt
 
Deep Dive: AWS X-Ray London Summit 2017
Deep Dive: AWS X-Ray London Summit 2017Deep Dive: AWS X-Ray London Summit 2017
Deep Dive: AWS X-Ray London Summit 2017Randall Hunt
 
Canada DevOps Conference
Canada DevOps ConferenceCanada DevOps Conference
Canada DevOps ConferenceRandall Hunt
 
TIAD - Is Automation Worth My Time?
TIAD - Is Automation Worth My Time?TIAD - Is Automation Worth My Time?
TIAD - Is Automation Worth My Time?Randall Hunt
 
A Century Of Weather Data - Midwest.io
A Century Of Weather Data - Midwest.ioA Century Of Weather Data - Midwest.io
A Century Of Weather Data - Midwest.ioRandall Hunt
 
MongoDB at LAHacks :)
MongoDB at LAHacks :)MongoDB at LAHacks :)
MongoDB at LAHacks :)Randall Hunt
 
Schema Design in MongoDB - TriMug Meetup North Carolina
Schema Design in MongoDB - TriMug Meetup North CarolinaSchema Design in MongoDB - TriMug Meetup North Carolina
Schema Design in MongoDB - TriMug Meetup North CarolinaRandall Hunt
 
Replication MongoDB Days 2013
Replication MongoDB Days 2013Replication MongoDB Days 2013
Replication MongoDB Days 2013Randall Hunt
 
Sharding in MongoDB Days 2013
Sharding in MongoDB Days 2013Sharding in MongoDB Days 2013
Sharding in MongoDB Days 2013Randall Hunt
 
Replication and replica sets
Replication and replica setsReplication and replica sets
Replication and replica setsRandall Hunt
 

More from Randall Hunt (13)

WhereML a Serverless ML Powered Location Guessing Twitter Bot
WhereML a Serverless ML Powered Location Guessing Twitter BotWhereML a Serverless ML Powered Location Guessing Twitter Bot
WhereML a Serverless ML Powered Location Guessing Twitter Bot
 
Where ml ai_heavy
Where ml ai_heavyWhere ml ai_heavy
Where ml ai_heavy
 
Randall's re:Invent Recap
Randall's re:Invent RecapRandall's re:Invent Recap
Randall's re:Invent Recap
 
Deep Dive: AWS X-Ray London Summit 2017
Deep Dive: AWS X-Ray London Summit 2017Deep Dive: AWS X-Ray London Summit 2017
Deep Dive: AWS X-Ray London Summit 2017
 
Canada DevOps Conference
Canada DevOps ConferenceCanada DevOps Conference
Canada DevOps Conference
 
TIAD - Is Automation Worth My Time?
TIAD - Is Automation Worth My Time?TIAD - Is Automation Worth My Time?
TIAD - Is Automation Worth My Time?
 
Git
GitGit
Git
 
A Century Of Weather Data - Midwest.io
A Century Of Weather Data - Midwest.ioA Century Of Weather Data - Midwest.io
A Century Of Weather Data - Midwest.io
 
MongoDB at LAHacks :)
MongoDB at LAHacks :)MongoDB at LAHacks :)
MongoDB at LAHacks :)
 
Schema Design in MongoDB - TriMug Meetup North Carolina
Schema Design in MongoDB - TriMug Meetup North CarolinaSchema Design in MongoDB - TriMug Meetup North Carolina
Schema Design in MongoDB - TriMug Meetup North Carolina
 
Replication MongoDB Days 2013
Replication MongoDB Days 2013Replication MongoDB Days 2013
Replication MongoDB Days 2013
 
Sharding in MongoDB Days 2013
Sharding in MongoDB Days 2013Sharding in MongoDB Days 2013
Sharding in MongoDB Days 2013
 
Replication and replica sets
Replication and replica setsReplication and replica sets
Replication and replica sets
 

Recently uploaded

The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsPixlogix Infotech
 
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024BookNet Canada
 
Search Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfSearch Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfRankYa
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfAlex Barbosa Coqueiro
 
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo DayH2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo DaySri Ambati
 
Artificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxArtificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxhariprasad279825
 
Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Commit University
 
Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebUiPathCommunity
 
SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024Lorenzo Miniero
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr BaganFwdays
 
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Mark Simos
 
Powerpoint exploring the locations used in television show Time Clash
Powerpoint exploring the locations used in television show Time ClashPowerpoint exploring the locations used in television show Time Clash
Powerpoint exploring the locations used in television show Time Clashcharlottematthew16
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .Alan Dix
 
DSPy a system for AI to Write Prompts and Do Fine Tuning
DSPy a system for AI to Write Prompts and Do Fine TuningDSPy a system for AI to Write Prompts and Do Fine Tuning
DSPy a system for AI to Write Prompts and Do Fine TuningLars Bell
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxLoriGlavin3
 
Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Manik S Magar
 
TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024Lonnie McRorey
 
Take control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteTake control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteDianaGray10
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubKalema Edgar
 
Commit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyCommit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyAlfredo García Lavilla
 

Recently uploaded (20)

The Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and ConsThe Ultimate Guide to Choosing WordPress Pros and Cons
The Ultimate Guide to Choosing WordPress Pros and Cons
 
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
Transcript: New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
 
Search Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfSearch Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdf
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdf
 
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo DayH2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
H2O.ai CEO/Founder: Sri Ambati Keynote at Wells Fargo Day
 
Artificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptxArtificial intelligence in cctv survelliance.pptx
Artificial intelligence in cctv survelliance.pptx
 
Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!
 
Dev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio WebDev Dives: Streamline document processing with UiPath Studio Web
Dev Dives: Streamline document processing with UiPath Studio Web
 
SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan
 
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
 
Powerpoint exploring the locations used in television show Time Clash
Powerpoint exploring the locations used in television show Time ClashPowerpoint exploring the locations used in television show Time Clash
Powerpoint exploring the locations used in television show Time Clash
 
From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .From Family Reminiscence to Scholarly Archive .
From Family Reminiscence to Scholarly Archive .
 
DSPy a system for AI to Write Prompts and Do Fine Tuning
DSPy a system for AI to Write Prompts and Do Fine TuningDSPy a system for AI to Write Prompts and Do Fine Tuning
DSPy a system for AI to Write Prompts and Do Fine Tuning
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
 
Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!
 
TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024
 
Take control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test SuiteTake control of your SAP testing with UiPath Test Suite
Take control of your SAP testing with UiPath Test Suite
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding Club
 
Commit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyCommit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easy
 

How to Choose The Right Database on AWS - Berlin Summit - 2019

  • 1. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Databases onAWS: HowToChoose The Right Database Randall Hunt Software Engineer at AWS @jrhunt randhunt@amazon Markus Ostertag CEO at Team Internet AG @Osterjour Dr. Sebastian Brandt Senior Key Expert - Knowledge Graph at Siemens CT
  • 2. S U M M I T © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
  • 3. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T 1970 1980 1990 2000 Oracle DB2 SQL Server MySQL PostgreSQL DynamoDB Redis MongoDB Elasticsearch Neptune CassandraAccess Aurora 2010 Timestream QLDB DocumentDB
  • 4. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Modern apps havemodern requirements Users: 1 million+ Data volume: TB–PB–EB Locality: Global Performance: Milliseconds–microseconds Request rate: Millions per second Access: Web, Mobile, IoT, Devices Scale: Up-down, Out-in Economics: Pay for what you use Developer access: No assembly requiredSocial mediaRide hailing Media streaming Dating
  • 5. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Common datacategories and usecases Relational Referential integrity, ACID transactions, schema- on-write Lift and shift, ERP, CRM, finance Key-value High throughput, low- latency reads and writes, endless scale Real-time bidding, shopping cart, social, product catalog, customer preferences Document Store documents and quickly query on any attribute Content management, personalization, mobile In-memory Query by key with microsecond latency Leaderboards, real-time analytics, caching Graph Quickly and easily create and navigate relationships between data Fraud detection, social networking, recommendation engine Time-series Collect, store, and process data sequenced by time IoT applications, event tracking Ledger Complete, immutable, and verifiable history of all changes to application data Systems of record, supply chain, health care, registrations, financial
  • 6. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T AWS: Purpose-built databases Relational Key-value Document In-memory Graph Search Amazon DynamoDB Amazon Neptune Amazon RDS Aurora CommercialCommunity Amazon ElastiCache Amazon Elasticsearch Service Amazon DocumentDB Time-series Ledger Amazon Timestream Amazon Quantum Ledger DatabaseMemcachedRedi s
  • 7. S U M M I T © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
  • 8. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Amazon RelationalDatabaseService(RDS) Managed relational database service with a choice of six popular database engines Easy to administer Available and durable Highly scalable Fast and secure No need for infrastructure provisioning, installing, and maintaining DB software Automatic Multi-AZ data replication; automated backup, snapshots, failover Scale database compute and storage with a few clicks with no app downtime SSD storage and guaranteed provisioned I/O; data encryption at rest and in transit
  • 9. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T TraditionalSQL • TCP based wire protocol • Well Known, lots of uses • Common drivers (JDBC) • Frequently used with ORMs • Scale UP individual instances • Scale OUT with read replicas • Sharding at application level • Lots of flavors but very similar language • Joins INSERT INTO table1 (id, first_name, last_name) VALUES (1, Randall’, Hunt’); SELECT col1, col2, col3 FROM table1 WHERE col4 = 1 AND col5 = 2 GROUP BY col1 HAVING count(*) > 1 ORDER BY col2
  • 10. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T AmazonAurora MySQL and PostgreSQL-compatible relational database built for the cloud Performance and availability of commercial-grade databases at 1/10th the cost Performance and scalability Availability and durability Highly secure Fully managed 5x throughput of standard MySQL and 3x of standard PostgreSQL; scale-out up to 15 read replicas Fault-tolerant, self-healing storage; six copies of data across three Availability Zones; continuous backup to Amazon S3 Network isolation, encryption at rest/transit Managed by RDS: No hardware provisioning, software patching, setup, configuration, or backups
  • 11. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T SQLvsNoSQL Optimized for storage Optimized for compute Normalized/relational Denormalized/hierarchical Ad hoc queries Instantiated views Scale vertically Scale horizontally Good for OLAP Built for OLTP at scale SQL NoSQL
  • 12. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Amazon DynamoDB Fast and flexible key value database service for any scale Comprehensive security Encrypts all data by default and fully integrates with AWS Identity and Access Management for robust security Performance at scale Consistent, single-digit millisecond response times at any scale; build applications with virtually unlimited throughput Global database for global users and apps Build global applications with fast access to local data by easily replicating tables across multiple AWS Regions Serverless No server provisioning, software patching, or upgrades; scales up or down automatically; continuously backs up your data
  • 13. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Amazon DynamoDB DataStructure Table Items Attributes Partition Key Sort Key Mandatory Key-value access pattern Determines data distribution Optional Model 1:N relationships Enables rich query capabilities All items for key ==, <, >, >=, <= “begins with” “between” “contains” “in” sorted results counts top/bottom N values
  • 14. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T DynamoDBSchemaandQueries • Connects over HTTP • Global Secondary Indexes and Local Secondary Indexes • Speed up queries with DAX • Global tables (mutli-region-multi-master) • Transactions across multiple tables • Change Streams • Rich query language with expressions • Provision read and write capacity units separately with or without autoscaling • Also supports pay per request model import boto3 votes_table = boto3.resource('dynamodb').Table('votes') resp = votes_table.update_item( Key={'name': editor}, UpdateExpression="ADD votes :incr", ExpressionAttributeValues={":incr": 1}, ReturnValues="ALL_NEW" )
  • 15. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T DynamoDB Advancements over the last 22 months VPC endpoints April 2017 Auto scaling June 2017 DynamoDB Accelerator (DAX) April 2017 Time To Live (TTL) February 2017 Global tables On-demand backup Encryption at rest November 2017 November 2017 November 2017 Point-in-time recovery March 2018 SLA June 2018 99.999% SLA August 2018 Adaptive capacity ACID November 2018 Transactions November 2018 On-demand
  • 16. Amazon DocumentDB: Modern cloud-native architecture Decouple compute and storage Distribute data in smaller partitions Increase the replication of data (6x) What would you do to improve scalability and availability? 1 2 3
  • 17. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T AmazonDocumentDB Fast,scalable,andfullymanagedMongoDB-compatibledatabaseservice Fast Scalable Fully managed MongoDB compatible Millions of requests per second with millisecond latency; twice the throughput of MongoDB Separation of compute and storage enables both layers to scale independently; scale out to 15 read replicas in minutes Managed by AWS: no hardware provisioning; auto patching, quick setup, secure, and automatic backups Compatible with MongoDB 3.6; use the same SDKs, tools, and applications with Amazon DocumentDB
  • 18. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Document databases • Data is stored in JSON-like documents • Documents map naturally to how humans model data • Flexible schema and indexing • Expressive query language built for documents (ad hoc queries and aggregations) JSON documents are first-class objects of the database { id: 1, name: "sue", age: 26, email: "sue@example.com", promotions: ["new user", "5%", "dog lover"], memberDate: 2018-2-22, shoppingCart: [ {product:"abc", quantity:2, cost:19.99}, {product:"edf", quantity:3, cost: 2.99} ] }
  • 19. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Amazon ElastiCache Redis and Memcached compatible, in-memory data store and cache Secure and reliable Network isolation, encryption at rest/transit, HIPAA, PCI, FedRAMP, multi AZ, and automatic failover Redis & Memcached compatible Fully compatible with open source Redis and Memcached Easily scalable Scale writes and reads with sharding and replicas Extreme performance In-memory data store and cache for microsecond response times
  • 20. S U M M I T © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Redis • Redis Serialization Protocol over TCP (RESP) • Supports Strings, Hashes, Lists, Sets, and Sorted Sets • Simple commands for manipulating in memory data structures • Pub/Sub features • Supports clustering via partitions
  • 21. S U M M I T © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Memcached • Text and Binary protocols over TCP • Small number of commands: set, add, replace, append, prepend, cas, get, gets, delete, incr,decr • Client/Application based partitioning
  • 22. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Relationshipsenable newapplications Retail fraud detectionRestaurant recommendationsSocial networks
  • 23. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Usecasesfor highlyconnected data Social networking Life Sciences Network & IT operationsFraud detection Recommendations Knowledge graphs
  • 24. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Different approaches for highlyconnected data Purpose-built for a business process Purpose-built to answer questions about relationships
  • 25. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Amazon Neptune Fully managed graph database Easy Build powerful queries easily with Gremlin and SPARQL Fast Query billions of relationships with millisecond latency Open Supports Apache TinkerPop & W3C RDF graph models Reliable Six replicas of your data across three AZs with full backup and restore
  • 26. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Open Source Apache TinkerPop™ Gremlin Traversal Language W3C Standard SPARQL Query Language R E S O U R C E D E S C R I P T I O N F R A M E W O R K ( R D F ) P R O P E R T Y G R A P H LEADING GRAPH MODELS AND FRAMEWORKS
  • 27. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T AmazonNeptunehighlevelarchitecture Bulk load from S3 Database Mgmt.
  • 28. S U M M I T © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. Gremlin g.addV('person').property(id, 1).property('name', 'randall') g.V('1').property(single, 'age', 27) g.addV('person').property(id, 2).property('name', 'markus') g.addE('knows').from(g.V('1')).to(g.V('2')).property('weight', 1.0) g.V().hasLabel('person') g.V().has('name', 'randall').out('knows').valueMap() http://tinkerpop.apache.org/docs/current/reference/#graph-traversal-steps
  • 29. S U M M I T © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved. SPARQLand RDF https://www.w3.org/TR/sparql11-query/
  • 30. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T THE BENEFIT OF URIs: LINKED DATA Linking across datasets by referencing globally unique URIs GeoNames Wikidata PermID Example: PermID (re)uses <http://sws.geonames.org/6252001/> as a global Identifier for the USA, which is an identifier rooted in GeoNames.
  • 31. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T AmazonTimestream(signup for thepreview) Fast, scalable, fully managed time-series database 1,000x faster and 1/10th the cost of relational databases Collect data at the rate of millions of inserts per second (10M/second) Trillions of daily events Adaptive query processing engine maintains steady, predictable performance Time-series analytics Built-in functions for interpolation, smoothing, and approximation Serverless Automated setup, configuration, server provisioning, software patching
  • 32. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T AmazonQuantumLedger Database(QLDB)(Preview) Fully managed ledger database Track and verify history of all changes made to your application’s data Immutable Maintains a sequenced record of all changes to your data, which cannot be deleted or modified; you have the ability to query and analyze the full history Cryptographically verifiable Uses cryptography to generate a secure output file of your data’s history Easy to use Easy to use, letting you use familiar database capabilities like SQL APIs for querying the data Highly scalable Executes 2–3X as many transactions than ledgers in common blockchain frameworks
  • 33. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T AWS DatabaseMigrationService(AWS DMS) M I G R A T I N G D A T A B A S E S T O A W S Migrate between on-premises and AWS Migrate between databases Automated schema conversion Data replication for zero-downtime migration 100,000+ databases migrated
  • 34. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Customer Premises, EC2, RDS Application Users Aurora PostgreSQLVPN/Network AWS DMS—Logical replication Start a replication instance Connect to source and target databases Select tables, schemas, or databases Let the AWS Database Migration Service create tables and load data Uses change data capture to keep them in sync Switch applications over to the target at your convenience
  • 35. S U M M I T © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
  • 36. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Customers aremoving toAWSDatabases Verizon is migrating over 1,000 business-critical applications and database backend systems to AWS, several of which also include the migration of production databases to Amazon Aurora. Wappa migrated from their Oracle database to Amazon Aurora and improved their reporting time per user by 75 percent. Trimble migrated their Oracle databases to Amazon RDS and project they will pay about 1/4th of what they paid when managing their private infrastructure. Intuit migrated from Microsoft SQL Server to Amazon Redshift to reduce data-processing timelines and get insights to decision makers faster and more frequently. Equinox Fitness migrated its Teradata on-premises data warehouse to Amazon Redshift. They went from static reports to a modern data lake that delivers dynamic reports. Eventbrite moved from Cloudera to Amazon EMR and were able to cut costs dramatically, spinning clusters up/down on-demand and using Spot (saving > 80%) and Reserved Instances. By December 2018, Amazon.com will have migrated 88% of their Oracle DBs (and 97% of critical system DBs) moved to Amazon Aurora and Amazon DynamoDB. They also migrated their 50 PB Oracle Data Warehouse to AWS (Amazon S3, Amazon Redshift, and Amazon EMR). Samsung Electronics migrated their Cassandra clusters to Amazon DynamoDB for their Samsung Cloud workload with 70% cost savings.
  • 37. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Mostenterprisedatabase& analyticscloud customers
  • 38. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Moststartupdatabase&analyticscloud customers
  • 39. S U M M I T © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
  • 40. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Problem: Matching IPs toRanges inReal-Time Bidding Real-Time Bidding Requirements: • High throughput (>25000 req/s) • Response time is critical • Scalability
  • 41. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Possiblechoices Relational Well known, Flexible Indices, Query language, Data scheme obvious Latency? Query volume? Future scale? Key-value High throughput, low-latency reads, endless scale Data scheme? Indices? Sharding? In-memory Query by key with microsecond latency Data scheme? Sharding? Future scale?
  • 42. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Possiblechoices Relational Well known, Flexible Indices, Query language, Data scheme obvious Latency? Query volume? Future scale? Key-value High throughput, low-latency reads, endless scale Data scheme? Indices? Sharding? In-memory Query by key with microsecond latency Data scheme? Sharding? Future scale?
  • 43. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T DynamoDB? But how? • Sharding and Index downsides addressed by transforming data during ingest • Accept multiple entries per range -> sharding & query possible/optimized
  • 44. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T DynamoDB itis!
  • 45. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T DynamoDB itis!
  • 46. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T DynamoDB itis!
  • 47. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Learnings • Focus on “core requirements“ • Be creative about data schemes • Accept minor downsides – very often there is no perfect solution • Use upsides of chosen database to optimize (i.e. adding DAX for caching)
  • 48. Knowledge Graphs @ Siemens Sebastian Brandt Steffen Lamparter Semantics and Reasoning Group Siemens Corporate Technology CT RDA BAM SMR-DE February 2019 Unrestricted © Siemens AG 2019siemens.com
  • 49. Dr. Sebastian Brandt, Dr. Steffen Lamparter, Siemens Corporate Technology Siemens AG 2019 Knowledge graphs become especially powerful for managing complex queries and heterogeneous data Why Knowledge Graphs? • Graphs are a natural way to represent entities and their relationships • Graphs can capture a broad spectrum of data (structured / unstructured) • Graphs can be managed efficiently Intuitive domain modelling Flexibility & performance Low up-front investment Robust data quality assurance Game-changing data integration
  • 50. Dr. Sebastian Brandt, Dr. Steffen Lamparter, Siemens Corporate Technology Siemens AG 2019 Semantic descriptions The semantic description indicates the meaning of an object or relation, e.g. Joe Kaeser is a person What are Graphs? Knowledge representation formalism semantic descriptions of entities and their relationships Relationships Logical connection between two objects e.g. Joe Kaeser is born in Arnbruck Rules make it possible to add further expert knowledge, e.g. "Siemens has to be a company, as a person is working there" Objects Real-world objects (things, places, people) and abstract concepts (genres, religions, professions) Is a Is born on Was founded by Is headquartered in Munich, Berlin Was promoted June 23, 1957 Is a Is working at Roesemarie Kaeser Is married to Werner v. Siemens and Friedrich Halske August 1, 2013 Person Company Has pictures Is born in Arnbruck Joe Kaeser
  • 51. Dr. Sebastian Brandt, Dr. Steffen Lamparter, Siemens Corporate Technology Siemens AG 2019 Knowledge graphs become a powerful addition to traditional data warehouses for managing heterogeneous data with complex relations Dimensions of data relations High number & complex relations, e.g. social networks Limited relations across data, e.g. time series data Variety of questions to be answered "We do not yet know all the different questions, users will ask"  Chatbot "We have a clear scope of user questions"  fixed set of dashboard needed Data heterogeneity Analysis needs to combine he- terogeneous data, e.g. sensor data, text data, product data Only few different data types and data sources e.g. invoicing database in ERP Quality of data Need to handle imperfect, incomplete or inconsistent data Ensured consistency of stored data low high low high low high low high Traditional data warehouses Knowledge graphs
  • 52. Dr. Sebastian Brandt, Dr. Steffen Lamparter, Siemens Corporate Technology Siemens AG 2019 Use cases for knowledge graphs can be clustered into five categories – overview and use case examples Data quality Digital companion Improving data availability and quality by combining and comparing data from various sources to fill in missing data sets or identify potentially wrong data and data duplicates Enhancing features of existing products or services with digital companions that are able to understand and process user questions and providing the needed data insights Data access & dashboarding Maintaining up-to-date meta-data, creating transparency on all available data and making them accessible to users via queries Recommender system Providing users high quality recommendations by identifying similarities in historical data Constraints & planning Enabling autonomous systems to understand data and its dependencies and take own decisions, such as autonomous planning of production proces-ses Degree of complexity
  • 53. Dr. Sebastian Brandt, Dr. Steffen Lamparter, Siemens Corporate Technology Siemens AG 2019 Example: gas-turbine maintenance planning Maintenance Repair Time-series Power turbine Configuration Fleet browsing and search DB Docs CSV, XML, Excel, … Siemens Knowledge Graph Analytics for R&D• Cross-life-cycle data integration at PS DO: gas-turbines, including maintenance, repair, monitoring, and configuration data • Holistic engineering-centric domain model • Intuitive graph queries independent of source data schemata • Virtual integration of time-series data DB Maintenance planning
  • 54. Dr. Sebastian Brandt, Dr. Steffen Lamparter, Siemens Corporate Technology Siemens AG 2019 A semantic data model enables flexible linking and an integrated, intuitive API for applications Semantic Data Model Planning tools Simulation tool Building Operations Security aaS Building Performance Service Portal Location Ba- sed Services Digital Lifecycle Platform Drive applications BT Knowledge Graph Customer Data Weather Data Public Energy Data MindSphereBuilding Structure BIM Product Data www Data APIProduct APIBuilding API
  • 55. Dr. Sebastian Brandt, Dr. Steffen Lamparter, Siemens Corporate Technology Siemens AG 2019 Creating perfect places based on Services – a user-centric holistic approach to the modern workplace … Customer Interest Relevant KPI’s Cost per space unit Employee satisfaction CO2 emissions Employee productivity Optimizing CAPEX and OPEX Energy and asset efficiency Space efficiency Individual efficiency and comfort Workplace Utilization Revenue per space unit Vacancy Rate Asset Performance/Useful Life
  • 56. Dr. Sebastian Brandt, Dr. Steffen Lamparter, Siemens Corporate Technology Siemens AG 2019 Industrializing Knowledge Graphs R&D AreasRelevant technologies Decision Making • Explanation of AI decisions • Data access: Semantic Search • Machine Learning on Graphs for recommendations, quality, etc. Decision Making • Reasoning and Constraint Solving • Machine/Deep Learning • Question Answering Storage and Integration • Reusable Semantic Modelling and Knowledge Graphs • Data integration and cleaning (e.g. entity reconciliation) Storage and Integration • Graph/NoSQL databases • Constraints and Rules • Probabilistic programming • Ontologies Generation • Extraction from unstructured data (inclusive text, audio, image) • Automatic semantic annotation of structured data • Learning of domain-specific rules/patterns Generation • NLP/Text understanding • Machine/Deep Learning • Computer vision • Sound recognition • Virtual data Integration • Information retrieval • … Industrial Knowledge Graph Decision Making Storage and Integration Generation Knowledge Automation HumansMachines ML for Graphs Ontology Library ML for automated annotation Simple RDF + PGM
  • 57. Dr. Sebastian Brandt, Dr. Steffen Lamparter, Siemens Corporate Technology Siemens AG 2019 Get in touch! Dr. Sebastian Brandt Senior Key Expert – Knowledge Graph and Data Management CT RDA BAM SMR-DE Dr. Steffen Lamparter Head of Research Group Semantics & Reasoning CT RDA BAM SMR-DE Siemens AG Corporate Technology Otto-Hahn-Ring 6 81739 München Germany E-mail steffen.lamparter@siemens.com Intranet intranet.siemens.com/ct
  • 58. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I T Thank you! S U M M I T © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.
  • 59. © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.S U M M I TS U M M I T © 2019, Amazon Web Services, Inc. or its affiliates. All rights reserved.

Editor's Notes

  1. Let’s take a look back at how database technology has evolved over the years. In the 1970s and 1980s, we had relational databases and then in the 1990s we added open source options like MySQL and PostgreSQL. Then, around mid 2000s to the current day, we have seen a significant growth of specialized databases that differ from the relational model. I don’t think it is a coincidence that these new database emerged at the same the time cloud was taking off as customers were beginning to build internet-scale apps that demanded functionality, performance, and scale for many different use cases in the same application.
  2. Modern apps create new requirements For example, if you think of some of the largest cloud applications today you will notice some common characteristics Millions of users Located all over the world All expecting instant experiences (e.g., millisecond-microsecond latencies) These systems need to scale on the fly One size fits all approach from the past no longer works, very easy to overburden a database Today, developers are doing what they do best They are breaking down large applications into smaller parts and picking the right tool for the right job So they never have to trade off functionality, performance, or scale
  3. Speaking of right tool for the right job Instead of a listing hundreds, lets pivot and first think about common categories Then consider what the purpose of a tool is within a category And common use cases we hear from customers For example If you were building a health insurance application required strict schema, data accuracy, and consistency, relational is a great choice If you were building a massive online game, with millions of players coming and going, requiring high through put reads and writes, with endless scale, key-value is a great choice If part of your application needed to make a product recommendation, based on highly connected data, graph is a great choice
  4. If you look at our offerings and how they align to these categories, our database strategy is fairly simple, we want to ensure you as developers have the very best purpose-built databases in each of these categories so that you never have to sacrifice on scale, performance, and functionality. Today we’re going to talk about the Amazon DocumentDB - Fully managed MongoDB compatible database service designed from the ground up to be fast, scalable, and highly available
  5. If you have legacy apps you want to migrate Interested in improving scale and performance Want to free up resources to innovate Customers running commercial databases often choose Amazon RDS Run your choice of database engines – open source, Oracle, SQL Server Amazon RDS automates time-consuming administration tasks Hardware provisioning, database setup, patching, and backups Customers can spend time innovating and building new applications
  6. This is why we built Amazon Aurora Aurora = performance and availability of commercial with cost-effectiveness of open source 5x performance of standard MySQL 3x performance of standard PostgreSQL With the Security, availability, and reliability of commercial-grade 1/10th the cost
  7. The first consideration that needs to be made when selecting a database is the characteristics of the data you are looking to leverage. If the data has a simple tabular structure, like an accounting spreadsheet, then the relational model could be adequate. Data such as geo-spatial, engineering parts, or molecular modeling, on the other hand, tends to be very complex.
  8. Let’s talk about one of the types of application use cases: Internet-scale. Amazon DynamoDB is a fast and flexible NoSQL database service for all applications that need consistent, single-digit millisecond latency at any scale. It is a fully managed cloud database and supports both document and key-value store models. Its flexible data model, reliable performance, and automatic scaling of throughput capacity, makes it a great fit for mobile, web, gaming, ad tech, IoT, and many other applications. Backup Notes 1/Performance at scale - Consistent, single-digit millisecond response times at any scale. Build applications with virtually unlimited throughput and storage, backed by a service level agreement for reliability. 2/Serverless - No hardware provisioning, software patching, or upgrades. Scales up or down automatically to accommodate your performance needs. Optimize costs by paying for only the resources you use. Protect your data with on-demand and continuous backups with no downtime. 3/Comprehensive Security- Encrypts all data by default and fully integrates with AWS Identity and Access Management for robust security. Get oversight of your tables by using integrated monitoring on audit logs with AWS CloudTrail, and network isolation with Amazon Virtual Private Cloud. 4/Modern applications - A database for serverless applications that includes AWS Lambda integration. Supports ACID transactions for business-critical applications. Build global applications with fast access to local data by easily replicating tables across multiple AWS Regions.
  9. Again, working backwards from our customers, and taking our collective knowledge of building distributed system and database service as AWS, what could we do differently here to improve these scenarios? 1/Decouple storage and compute 2/Break-up data into smaller partitions, spread the data across multiple AZs, make use of cell-based architectures 3/Distribute those partitions six ways across three AZs We did all this to reduce single points of failures and to reduce blast radiuses – remember, this failure all the time whether it is hardware, software, or networking
  10. Working backwards from our customers and these challenges, we built Amazon DocumentDB to be a fast, scalable, fully-managed, and MongoDB-compatible AWS database service.
  11. Data is stored in JSON-like documents and JSON documents are first-class objects within the database – documents are not a data type or a value, they are the key design point of the database Document databases have flexible schema and make the representation of hierarchical and semi-structured data easy – they also enable powerful indexing to make the querying to such documents fast Documents map naturally to object-oriented programming, which makes the flow of data with your app to persistent layer easier Expressive query languages built for documents that enable ad hoc queries and aggregations across documents Together Document databases help developers build applications faster and iterate quickly
  12. Amazon ElastiCache is a web service that makes it easy to deploy, operate, and scale an in-memory data store and cache in the cloud. The service improves the performance of web applications by allowing you to retrieve information from fast, managed, in-memory data stores, instead of relying entirely on slower disk-based databases. Amazon ElastiCache supports two open-source in-memory engines: Redis - a fast, open source, in-memory data store and cache. Amazon ElastiCache for Redis is a Redis-compatible in-memory service that delivers the ease-of-use and power of Redis along with the availability, reliability and performance suitable for the most demanding applications. Both single-node and up to 15-shard clusters are available, enabling scalability to up to 6.1 TiB of in-memory data. ElastiCache for Redis is fully managed, scalable, and secure - making it an ideal candidate to power high-performance use cases such as Web, Mobile Apps, Gaming, Ad-Tech, and IoT. Memcached - a widely adopted memory object caching system. Amazon ElastiCache for Memcached is protocol compliant with Memcached, so popular tools that you use today with existing Memcached environments will work seamlessly with the service. ElastiCache for Memcached is suitable for caching use cases where performance and concurrency is important. Key benefits of Amazon ElastiCache include: Redis and Memcached Compatible With Amazon ElastiCache, you get native access to Redis or Memcached in-memory environments. This enables compatibility with your existing tools and applications. Extreme Performance Amazon ElastiCache works as an in-memory data store and cache to support the most demanding applications requiring sub-millisecond response times. By utilizing an end-to-end optimized stack running on customer dedicated nodes, Amazon Elasticache provides you secure, blazing fast performance. Fully Managed You no longer need to perform management tasks such as hardware provisioning, software patching, setup, configuration, monitoring, failure recovery, and backups. ElastiCache continuously monitors your clusters to keep your workloads up and running so that you can focus on higher value application development. Easily Scalable Amazon ElastiCache can scale-out, scale-in, and scale-up to meet fluctuating application demands.  Write and memory scaling is supported with sharding. Replicas provide read scaling.
  13. Amazon Neptune is a fast, reliable, fully-managed graph database service that makes it easy to build and run applications that work with highly connected datasets. The core of Amazon Neptune is a purpose-built, high-performance graph database engine optimized for storing billions of relationships and querying the graph with milliseconds latency. Amazon Neptune supports popular graph models Property Graph and W3C's RDF, and their respective query languages Apache TinkerPop Gremlin and SPARQL, allowing you to easily build queries that efficiently navigate highly connected datasets. Neptune powers graph use cases such as recommendation engines, fraud detection, knowledge graphs, drug discovery, and network security. Neptune supports both leading graph models, which are the Property Graph model and W3C’s Resource Description Framework (RDF), and that means Customers can choose the best model for their application needs. Developers like the Property Graphs because it is somewhat familiar to relational models, and they like the open source Apache TinkerPop Gremlin traversal language since it provides a way to quickly traverse property graphs. Also, Developers like RDF as it provides flexibility for modeling complex information and there are a lot of existing public domain data sets available in RDF including Wikidata and PubChem, which is a database of chemical molecules. So with Neptune being fast and scalable, Developers can create graph applications that store billions of relationships and query the graph with milliseconds latency. Neptune is reliable, offering greater than 99.99 percent availability. We store six copies of your data across three Availability Zones and back up your data to Amazon S3. And If there is an instance failover, it typically takes less than 30 seconds.   By using graph query languages like Gremlin or SPARQL, developers can easily execute powerful queries that are easy to write and perform well on connected data, instead of having to write complex SQL queries that are difficult to tune for performance.   So for diverse use cases like social networks to fraud detection to knowledge graphs to drug discovery, with Neptune you can select the best model for your application.
  14. There are two primary graph models that are widely-used. A property graph is a common name for an attributed, multi-relational graph, for those who paid attention in math class. The leading Property Graph API is the open standard Apache TinkerPop™ project. It provides an imperative traversal language, called Gremlin, that can be used to write traversals on property graphs, and it is supported by a number of open source and vendor implementations. Customers like Property Graphs as they are familiar to developers that are used to relational models, and they like the open source Apache TinkerPop Gremlin traversal language as it provides a way to quickly traverse property graphs. The second is the Resource Description Framework, or RDF, standardized by the W3C in a set of standards collectively known as the Semantic Web. The SPARQL query language for RDF allows uses to express declarative graph queries against RDF graph models. The RDF model is also a labeled, directed multi-graph, but it uses the concept of triples, subject, predicate, and object, to encode the graph. Customers like RDF as it provides flexibility for modeling complex information domains and there are a number of existing public domain data sets available in RDF including Wikidata and PubChem, a database of chemical molecules.
  15. RDF data is provided in “triples”
  16. Amazon Timestream is a purpose-built, time-series database designed specifically for collecting, storing, and analyzing time series data.   At the very core of Amazon Timestream, time isn’t just an attribute, but rather the single primary axis of the data model. This allows for simplification and specialization across the database. 1/ 1,000X faster, and at 1/10th the cost of relational databases - Amazon Timestream can collect fast moving time-series data from multiple sources at the rate of millions of inserts per second (10M/second). Amazon Timestream organizes data by time intervals, reducing the amount of data that needs to be scanned to answer a query. Amazon Timestream also executes inserts and queries in separate processing tiers which eliminates resource contention and improves performance. [Note: This claim is based on internal benchmarks and applies to both queries and inserts. The service team is comfortable making the claim].   2/ Trillions of daily events– With its speed and purpose-built architecture, Amazon Timestream is capable of processing trillions of events daily. This opens up the door to more IoT devices, more sensor reads and ultimately a larger data set to make smarter decisions on time series and machine data. Amazon Timestream’s adaptive query processing engine and data retention policies adjust the query performance and storage capacity to maintain steady, predictable performance at the lowest possible cost as your data grows over time.   3/ Analytics optimized for time series data– Analyzing time-series data with Amazon Timestream is easy, with built-in functions for interpolation, smoothing, and approximation that can be used to identify trends, patterns, and anomalies. For example, a smart home device manufacturer can use Amazon Timestream to collect motion or temperature data from the device sensors in a home, interpolate that data to identify the time ranges without any motion in the home, and alert consumers to take actions such as turning off the lights or turning down the heat to save energy during times when no one is in the house.   4/ Serverless – With Amazon Timestream, there are no servers to manage. As your application needs change, Amazon Timestream automatically scales up or down to adjust capacity and performance. Amazon Timestream takes care of the time-consuming tasks such as server provisioning, software patching, setup, and configuration so you can focus on building your applications. In addition, You can set policies to automate the retention and tiering of how data is stored, which can significantly reduce your manual effort, storage requirements, and cost.   Amazon Timestream is available for preview today.
  17. Ledgers are typically used to record a history of economic and financial activity in an organization. Many organizations build applications with ledger-like functionality because they want to maintain an accurate history of their applications' data, for example, tracking the history of credits and debits in banking transactions, verifying the data lineage of an insurance claim, or tracing movement of an item in a supply chain network. Ledger applications are often implemented using custom audit tables or audit trails created in relational databases. However, building audit functionality with relational databases is time-consuming and prone to human error. It requires custom development, and since relational databases are not inherently immutable, any unintended changes to the data are hard to track and verify. Alternatively, blockchain frameworks, such as Hyperledger Fabric and Ethereum, can also be used as a ledger. However, this adds complexity as you need to set-up an entire blockchain network with multiple nodes, manage its infrastructure, and require the nodes to validate each transaction before it can be added to the ledger.
  18. To help customers move their databases to the cloud, we built the AWS Database Migration Service (DMS). DMS helps you move common commercial databases such as Oracle and SQL Server Open source databases such as MySQL and NoSQL databases such as MongoDB to AWS
  19. When you think about this fundamental question 2 fundamental areas of focus emerge
  20. Customers are migrating their workloads to AWS in record numbers Verizon is migrating over 1,000 business-critical applications and database backend systems to AWS, several of which also include the migration of production databases to Amazon Aurora—AWS’s relational database engine that combines the speed and availability of high-end commercial databases with the simplicity and cost-effectiveness of open source databases. Wappa, a market leader in taxi expense management, available in 22 countries, migrated from their Oracle database to Amazon Aurora and improved their reporting time per user by 75 percent. Trimble, a global leader in telematics solutions, had a significant investment in on-premises hardware running Oracle databases. Rather than refresh the hardware and renew their licenses, they opted to migrate the databases to Amazon RDS and project they will pay about 1/4th of what we were paying when managing their private infrastructure. Thomas Publishing is using AWS to save hundreds of thousands of dollars in data center costs, quickly launch new websites by spinning up resources in one day instead of several weeks, and rapidly migrate critical Oracle applications to the cloud. The company links industrial buyers and suppliers by offering an array of information and services online and in print. Thomas runs its primary website on AWS, and worked with AWS Partner Apps Associates to migrate key content management applications to Amazon Aurora on RDS. Thomas Publishing was one of the first customers to use DMS to migrate from Oracle to Aurora. Amazon.com have been moving off Oracle as fast as we can. Today 92% of Amazon’s Fulfillment Centers worldwide have migrated off Oracle to AWS Database services, mostly Aurora PostgreSQL (338 out of 368 FC’s). (Note: By the end of December 2018 Amazon.com will have migrated 170 (56%) of the Critical service databases to DynamoDB and 4316 (58%) of the non-Critical service databases to Aurora and RDS PostgreSQL in a single year). Amazon’s DW: Amazon builds and operates thousands of micro-services to serve millions of customers. These include catalog browsing, order placement, transaction processing, delivery scheduling, video services, and Prime registration. Each service publishes datasets to Amazon’s massive analytics infrastructure, including over 50 petabytes of data and 75,000 tables, processing 600,000 user analytics jobs each day. They migrated this Data Warehouse from Oracle to AWS (S3, EMR, and Redshift). The new analytics infrastructure has one Data Lake with over 100 petabytes of data – almost twice the size of the previous Oracle Data Warehouse. Teams across Amazon are now using over 2700 Amazon Redshift or Amazon EMR clusters to process data from the Data Lake. Samsung - Samsung Electronics migrated their Cassandra cluster to DynamoDB for their Samsung Cloud workload, saving 70%. Intuit - Intuit is a business and financial software company that develops and sells financial, accounting, and tax-preparation software and services for small businesses, accountants, and individuals. Intuit migrated from Microsoft SQL Server to Amazon Redshift to reduce data-processing timelines and get insights to decision makers faster and more frequently. "With Amazon Redshift, our small team has handled a tenfold increase in data volumes while reducing the amount of time spent on system administration—freeing up time for value-add development." -Jason Rhoades, Systems Architect, Intuit Equinox - Equinox Fitness Clubs transitioned from on-premises Data Warehouses and data marts in Teradata to a cloud-based, integrated data platform, built on AWS and Amazon Redshift. They went from static reports, redundant data, and inefficient data integration to a modern and flexible Data Lake and Data Warehouse architecture that delivers dynamic reports based on trusted data. Eventbrite – Eventbrite’s mission is to bring the world together through live experiences. To achieve this goal, Eventbrite relies on data-driven decisions at every level. Previously they had Cloudera on-premises. With Amazon EMR, they are able to intelligent resize clusters, and cut costs considerably. They only pay for what they use and reduce costs further by purchasing Reserved instances and bidding on Spot instances (saving > 80%)
  21. We have >400,000 customer accounts using AWS database and analytics. While the database and analytics markets have been around for a while, with many mature offerings for customers to choose from. We continue to see customers move to the cloud for a number of reasons and our recent growth in the database market is evidence of how rapidly the landscape is changing. In 2016 for the database market, Amazon grew and order or magnitude faster than all other leading commercial database vendors. And this is off an already sizeable revenue base. We see this for a number or reasons Customers move to the cloud to minimize time spent managing infrastructure Customers are choosing the cloud and migrating more and more of their workloads to it. In the next 10 to 15 years, the majority of computing is going to be done in the cloud. In the fullness of time, very few companies will want to own their own data centers, manage infrastructure whether it is compute, storage, databases or analytics. Customers move to the cloud for performance, scale, reliability and cost Increasingly, new applications need to be globally distributed, support millions of users and devices, work with petabytes of data, run 24/7 and be responsive. As customers move to the cloud and to micro-service architectures, developers are increasingly the ones making technology decisions As customers move from monolithic apps to micro-service architectures with loosely coupled components and DevOps cultures. The developers are increasingly making decisions as part of their application development lifecycle on what frameworks and components do they use.
  22. Luckily, we have 2 customers here to speak to us today.
  23. So does it make sense for CIOs to use both relational databases and graph databases? Or should they standardize across the enterprise on one or the other? Today it makes pragmatic sense to use both. Each model has their pros and cons; as the enterprise IT user typically has a wide set of problems it needs to solve, there is no single database or database model that is best at everything. How to choose However, that’s not a very rigorous approach to delineating which is best. How can you tell when the situation is right for graph databases over RDBMS? Start by drawing the domain on a whiteboard. If your domain entities have relationships to other entities, and your queries rely on exploring those relationships, then a graph database is a great fit. Developers find the whiteboard test very convenient to work with because of its adaptability. This ability to adapt is particularly useful as new information about the domain becomes known or changes in requirements cause the model to change. https://sdtimes.com/databases/guest-view-relational-vs-graph-databases-use/ Better performance – highly connected data can cause a lot of joins, which generally are expensive. After over 7 self/recursive joins, the RDMS starts to get really slow compared to Neo4j/native graph databases.  Flexibility – graph data is not forced into a structure like a relational table, and attributes can be added and removed easily. This is especially useful for semi-structured data where a representation in relational database would result in lots of NULL column values. Easier data modelling – since you are not bound by a predefined structure, you can also model your data easily. SQL query pain – query syntax becomes complex and large as the joins increase. https://dzone.com/articles/why-are-native-graph-databases-more-efficient-than
  24. Example: Mona Lisa -> Person or Painting??? Graphs for integration of heterogeneous data Flexible, resource centric data model, use of Linked Data techniques Natural representation for a variety of data models
  25. Asset Tracking Problem Medical staff spends significant time looking for hospital equipment, leading to overstocking and low employee productivity Solution Digital construction twin as calculation, asset type and user interface basis, different tracking technologies and asset data as input for the digital performance twin Space utilization Problem Low efficiencies due to lack of good occupancy and usage data Solution Digital construction twin as calculation and user interface basis, occupancy sensors based on different technologies as input for digital performance twin