SlideShare a Scribd company logo
How to Reduce Your Database
Total Cost of Ownership
with TimescaleDB
November 2019
Presenter
Bob Boule
Product Management
● What’s New in TimescaleDB 1.5?
● Feature Deep Dive: Compression
● Feature Deep Dive: Move Chunks
● Questions
Agenda
THEME: Reduce Total Cost of Ownership (TCO) by optimizing use of storage
● Time-series use cases produce a LOT of data
● Users need cost effective and efficient way to store data
● Storage is usually the largest contributor to operational cost
● Users expects performant solution while delivering storage efficiencies
○ Ingestion and query performance cannot suffer
● Users need to be able to set a long term storage strategy for data
● Store data based on its access & usage pattern
What’s New in TimescaleDB 1.5?
What’s New in TimescaleDB 1.5?
Capability Description Availability
Native Compression Store more data with less actual storage. Community Edition
Data Tiering
(“move chunks”)
Optimizes performance + storage costs by
allowing you to map your data to the
appropriate storage technology.
Enterprise Edition
Reducing Total Cost of Ownership
Time-series
data
Time-series
data
STORAGE
Compression
enables you store
more time-series
data on less actual
storage.
Reducing Total Cost of Ownership
Time-series
data
Time-series
data
STORAGE
Compression
enables you store
more time-series
data on less actual
storage.
Time-series
data
Time-series
data
STORAGE
Time-series
data
Time-series
data
STORAGE
Data Tiering enables you
to target data to different
storage types, optimizing
cost & performance.
Native Compression
COMMUNITY
edition
Native Compression Overview
Compression reduces the overall storage footprint of your data saves $
● User can compression their data manually or by using a policy
● Compression can be applied to specific hypertables
● Once upgraded to 1.5, user can start applying compression policies to their
hypertables
NOTE: Upgrading to TSDB 1.5 does not automatically compress existing
data or start compressing new data
● User drives compression policy based on:
○ Frequency of access
○ Query Characteristics (“Shallow and Wide” vs “Deep and Narrow”)
● Very adaptable:
○ Users can select “order by” and “segment by” to achieve best overall compression
Native Compression: Sample Strategy
Compression: How It Works
Defining how data will be organized
Use an “order by” parameter to move
“like” data closer together to enhance
compressibility and maintain
performance.
order by device_id
Compression: How It Works
Defining how data will be organized
Use “segment by” operator to further
group like data together.
segment by device_id
Using Compression: Manual
Use Case: User wants to compress chunks manually
In this use case we will identify a set of target chunks and compress them manually. Let's start by
identifying chucks more than 3 months old:
SELECT show_chunks('conditions', older_than => interval '3 days');
which will return a list of chunks. User can then compress one of these chunks using the following:
SELECT compress_chunk('_timescaledb_internal._hyper_1_2_chunk');
Using Compression: Policy
Use Case: User wants to compress chunks automatically based on age
In this case, as a user I have decided that I want to compress all chucks older than 60 days, without
any manual intervention.
SELECT add_compress_chunks_policy('cpu', '60d'::interval);
This command will set up a background job that will kick off every 15 minutes, looking for chunks that
are more than 60 days old and compressing them. We can see the job has been added by running:
SELECT * FROM _timescaledb_config.bgw_job WHERE job_type LIKE 'compress%';
which will return the Jobs in the system associated with compression
Compression
Summary
● Compression can be configured based on
user data model
● Compression strategy is driven by the user
● Real savings by reducing storage costs
without removing data
● Why is compression important:
○ Reduce total cost of ownership by
reducing storage costs
○ Allows user to retain more data
without the additional costs
COMMUNITY
edition
Data Tiering
(“move chunks”)
ENTERPRISE
edition
Move Chunks
● Reduce storage costs by leveraging more cost effective storage options
● Tier data based on usage / access pattern
● Optimize storage costs by moving infrequently used data to slower / more
cost effective storage
● Combined with compression will allow for overall reduced TCO based on
efficient use of storage
Move Chunks: Sample Strategy
Move Chunks: How It Works
First step: Create a new tablespace
Create a new tablespace, which will be backed by a new storage mount (in this example I have
mount the new storage to /mnt/postgres):
CREATE TABLESPACE tablespace_2 OWNER postgres LOCATION '/mnt/postgres';
Process of Moving the data:
SELECT move_chunk(chunk=>'_timescaledb_internal._hyper_1_2_chunk',
destination_tablespace=>'tablespace_2',
index_destination_tablespace=>'tablespace_2',
reorder_index=>'_timescaledb_internal._hyper_1_2_chunk_cluster_test_time_id
x', verbose=>TRUE);
Move Chunks: Use Cases
Data Tiering
Summary
● The ability to move data to cost effective
storage
○ As data ages and is less frequently
queried it can be moved to less
expensive storage
● Help spread the data out over multiple
arrays
● Leverages the ability to create tablespaces
in Postgres and back them with different
storage
ENTERPRISE
edition
Learn more
Native Compression Docs: tsdb.co/WhatsNewCompression
Move Chunks Docs: tsdb.co/WhatsNewMoveChunk
All products: tsdb.co/WhatsNewProducts
Get the latest news and join our community
● blog.timescale.com
● twitter.com/TimescaleDB
● github.com/timescale/TimescaleDB
● slack.timescale.com
Questions?

More Related Content

What's hot

(Big) Data Serialization with Avro and Protobuf
(Big) Data Serialization with Avro and Protobuf(Big) Data Serialization with Avro and Protobuf
(Big) Data Serialization with Avro and Protobuf
Guido Schmutz
 
Common Strategies for Improving Performance on Your Delta Lakehouse
Common Strategies for Improving Performance on Your Delta LakehouseCommon Strategies for Improving Performance on Your Delta Lakehouse
Common Strategies for Improving Performance on Your Delta Lakehouse
Databricks
 
Engineering Fast Indexes for Big-Data Applications: Spark Summit East talk by...
Engineering Fast Indexes for Big-Data Applications: Spark Summit East talk by...Engineering Fast Indexes for Big-Data Applications: Spark Summit East talk by...
Engineering Fast Indexes for Big-Data Applications: Spark Summit East talk by...
Spark Summit
 
5 Critical Steps to Clean Your Data Swamp When Migrating Off of Hadoop
5 Critical Steps to Clean Your Data Swamp When Migrating Off of Hadoop5 Critical Steps to Clean Your Data Swamp When Migrating Off of Hadoop
5 Critical Steps to Clean Your Data Swamp When Migrating Off of Hadoop
Databricks
 
Cassandra Data Modeling - Practical Considerations @ Netflix
Cassandra Data Modeling - Practical Considerations @ NetflixCassandra Data Modeling - Practical Considerations @ Netflix
Cassandra Data Modeling - Practical Considerations @ Netflix
nkorla1share
 
Massive Data Processing in Adobe Using Delta Lake
Massive Data Processing in Adobe Using Delta LakeMassive Data Processing in Adobe Using Delta Lake
Massive Data Processing in Adobe Using Delta Lake
Databricks
 
Time series Analytics - a deep dive into ADX Azure Data Explorer @Data Saturd...
Time series Analytics - a deep dive into ADX Azure Data Explorer @Data Saturd...Time series Analytics - a deep dive into ADX Azure Data Explorer @Data Saturd...
Time series Analytics - a deep dive into ADX Azure Data Explorer @Data Saturd...
Riccardo Zamana
 
ClickHouse Data Warehouse 101: The First Billion Rows, by Alexander Zaitsev a...
ClickHouse Data Warehouse 101: The First Billion Rows, by Alexander Zaitsev a...ClickHouse Data Warehouse 101: The First Billion Rows, by Alexander Zaitsev a...
ClickHouse Data Warehouse 101: The First Billion Rows, by Alexander Zaitsev a...
Altinity Ltd
 
InfluxDB IOx Tech Talks: Query Engine Design and the Rust-Based DataFusion in...
InfluxDB IOx Tech Talks: Query Engine Design and the Rust-Based DataFusion in...InfluxDB IOx Tech Talks: Query Engine Design and the Rust-Based DataFusion in...
InfluxDB IOx Tech Talks: Query Engine Design and the Rust-Based DataFusion in...
InfluxData
 
Debugging PySpark: Spark Summit East talk by Holden Karau
Debugging PySpark: Spark Summit East talk by Holden KarauDebugging PySpark: Spark Summit East talk by Holden Karau
Debugging PySpark: Spark Summit East talk by Holden Karau
Spark Summit
 
An Introduction to Elastic Search.
An Introduction to Elastic Search.An Introduction to Elastic Search.
An Introduction to Elastic Search.
Jurriaan Persyn
 
A Tale of Three Apache Spark APIs: RDDs, DataFrames, and Datasets with Jules ...
A Tale of Three Apache Spark APIs: RDDs, DataFrames, and Datasets with Jules ...A Tale of Three Apache Spark APIs: RDDs, DataFrames, and Datasets with Jules ...
A Tale of Three Apache Spark APIs: RDDs, DataFrames, and Datasets with Jules ...
Databricks
 
Apache Kudu: Technical Deep Dive


Apache Kudu: Technical Deep Dive

Apache Kudu: Technical Deep Dive


Apache Kudu: Technical Deep Dive


Cloudera, Inc.
 
Using Apache Arrow, Calcite, and Parquet to Build a Relational Cache
Using Apache Arrow, Calcite, and Parquet to Build a Relational CacheUsing Apache Arrow, Calcite, and Parquet to Build a Relational Cache
Using Apache Arrow, Calcite, and Parquet to Build a Relational Cache
Dremio Corporation
 
Top 5 Mistakes When Writing Spark Applications
Top 5 Mistakes When Writing Spark ApplicationsTop 5 Mistakes When Writing Spark Applications
Top 5 Mistakes When Writing Spark Applications
Spark Summit
 
Building Lakehouses on Delta Lake with SQL Analytics Primer
Building Lakehouses on Delta Lake with SQL Analytics PrimerBuilding Lakehouses on Delta Lake with SQL Analytics Primer
Building Lakehouses on Delta Lake with SQL Analytics Primer
Databricks
 
A Day in the Life of a ClickHouse Query Webinar Slides
A Day in the Life of a ClickHouse Query Webinar Slides A Day in the Life of a ClickHouse Query Webinar Slides
A Day in the Life of a ClickHouse Query Webinar Slides
Altinity Ltd
 
Better than you think: Handling JSON data in ClickHouse
Better than you think: Handling JSON data in ClickHouseBetter than you think: Handling JSON data in ClickHouse
Better than you think: Handling JSON data in ClickHouse
Altinity Ltd
 
PySpark Best Practices
PySpark Best PracticesPySpark Best Practices
PySpark Best Practices
Cloudera, Inc.
 
Iceberg + Alluxio for Fast Data Analytics
Iceberg + Alluxio for Fast Data AnalyticsIceberg + Alluxio for Fast Data Analytics
Iceberg + Alluxio for Fast Data Analytics
Alluxio, Inc.
 

What's hot (20)

(Big) Data Serialization with Avro and Protobuf
(Big) Data Serialization with Avro and Protobuf(Big) Data Serialization with Avro and Protobuf
(Big) Data Serialization with Avro and Protobuf
 
Common Strategies for Improving Performance on Your Delta Lakehouse
Common Strategies for Improving Performance on Your Delta LakehouseCommon Strategies for Improving Performance on Your Delta Lakehouse
Common Strategies for Improving Performance on Your Delta Lakehouse
 
Engineering Fast Indexes for Big-Data Applications: Spark Summit East talk by...
Engineering Fast Indexes for Big-Data Applications: Spark Summit East talk by...Engineering Fast Indexes for Big-Data Applications: Spark Summit East talk by...
Engineering Fast Indexes for Big-Data Applications: Spark Summit East talk by...
 
5 Critical Steps to Clean Your Data Swamp When Migrating Off of Hadoop
5 Critical Steps to Clean Your Data Swamp When Migrating Off of Hadoop5 Critical Steps to Clean Your Data Swamp When Migrating Off of Hadoop
5 Critical Steps to Clean Your Data Swamp When Migrating Off of Hadoop
 
Cassandra Data Modeling - Practical Considerations @ Netflix
Cassandra Data Modeling - Practical Considerations @ NetflixCassandra Data Modeling - Practical Considerations @ Netflix
Cassandra Data Modeling - Practical Considerations @ Netflix
 
Massive Data Processing in Adobe Using Delta Lake
Massive Data Processing in Adobe Using Delta LakeMassive Data Processing in Adobe Using Delta Lake
Massive Data Processing in Adobe Using Delta Lake
 
Time series Analytics - a deep dive into ADX Azure Data Explorer @Data Saturd...
Time series Analytics - a deep dive into ADX Azure Data Explorer @Data Saturd...Time series Analytics - a deep dive into ADX Azure Data Explorer @Data Saturd...
Time series Analytics - a deep dive into ADX Azure Data Explorer @Data Saturd...
 
ClickHouse Data Warehouse 101: The First Billion Rows, by Alexander Zaitsev a...
ClickHouse Data Warehouse 101: The First Billion Rows, by Alexander Zaitsev a...ClickHouse Data Warehouse 101: The First Billion Rows, by Alexander Zaitsev a...
ClickHouse Data Warehouse 101: The First Billion Rows, by Alexander Zaitsev a...
 
InfluxDB IOx Tech Talks: Query Engine Design and the Rust-Based DataFusion in...
InfluxDB IOx Tech Talks: Query Engine Design and the Rust-Based DataFusion in...InfluxDB IOx Tech Talks: Query Engine Design and the Rust-Based DataFusion in...
InfluxDB IOx Tech Talks: Query Engine Design and the Rust-Based DataFusion in...
 
Debugging PySpark: Spark Summit East talk by Holden Karau
Debugging PySpark: Spark Summit East talk by Holden KarauDebugging PySpark: Spark Summit East talk by Holden Karau
Debugging PySpark: Spark Summit East talk by Holden Karau
 
An Introduction to Elastic Search.
An Introduction to Elastic Search.An Introduction to Elastic Search.
An Introduction to Elastic Search.
 
A Tale of Three Apache Spark APIs: RDDs, DataFrames, and Datasets with Jules ...
A Tale of Three Apache Spark APIs: RDDs, DataFrames, and Datasets with Jules ...A Tale of Three Apache Spark APIs: RDDs, DataFrames, and Datasets with Jules ...
A Tale of Three Apache Spark APIs: RDDs, DataFrames, and Datasets with Jules ...
 
Apache Kudu: Technical Deep Dive


Apache Kudu: Technical Deep Dive

Apache Kudu: Technical Deep Dive


Apache Kudu: Technical Deep Dive


 
Using Apache Arrow, Calcite, and Parquet to Build a Relational Cache
Using Apache Arrow, Calcite, and Parquet to Build a Relational CacheUsing Apache Arrow, Calcite, and Parquet to Build a Relational Cache
Using Apache Arrow, Calcite, and Parquet to Build a Relational Cache
 
Top 5 Mistakes When Writing Spark Applications
Top 5 Mistakes When Writing Spark ApplicationsTop 5 Mistakes When Writing Spark Applications
Top 5 Mistakes When Writing Spark Applications
 
Building Lakehouses on Delta Lake with SQL Analytics Primer
Building Lakehouses on Delta Lake with SQL Analytics PrimerBuilding Lakehouses on Delta Lake with SQL Analytics Primer
Building Lakehouses on Delta Lake with SQL Analytics Primer
 
A Day in the Life of a ClickHouse Query Webinar Slides
A Day in the Life of a ClickHouse Query Webinar Slides A Day in the Life of a ClickHouse Query Webinar Slides
A Day in the Life of a ClickHouse Query Webinar Slides
 
Better than you think: Handling JSON data in ClickHouse
Better than you think: Handling JSON data in ClickHouseBetter than you think: Handling JSON data in ClickHouse
Better than you think: Handling JSON data in ClickHouse
 
PySpark Best Practices
PySpark Best PracticesPySpark Best Practices
PySpark Best Practices
 
Iceberg + Alluxio for Fast Data Analytics
Iceberg + Alluxio for Fast Data AnalyticsIceberg + Alluxio for Fast Data Analytics
Iceberg + Alluxio for Fast Data Analytics
 

Similar to How to Reduce Your Database Total Cost of Ownership with TimescaleDB

Combining IBM Real-time Compression and IBM ProtecTIER Deduplication
Combining IBM Real-time Compression and IBM ProtecTIER DeduplicationCombining IBM Real-time Compression and IBM ProtecTIER Deduplication
Combining IBM Real-time Compression and IBM ProtecTIER Deduplication
IBM India Smarter Computing
 
Real time database compression optimization using iterative length compressio...
Real time database compression optimization using iterative length compressio...Real time database compression optimization using iterative length compressio...
Real time database compression optimization using iterative length compressio...
csandit
 
REAL TIME DATABASE COMPRESSION OPTIMIZATION USING ITERATIVE LENGTH COMPRESSIO...
REAL TIME DATABASE COMPRESSION OPTIMIZATION USING ITERATIVE LENGTH COMPRESSIO...REAL TIME DATABASE COMPRESSION OPTIMIZATION USING ITERATIVE LENGTH COMPRESSIO...
REAL TIME DATABASE COMPRESSION OPTIMIZATION USING ITERATIVE LENGTH COMPRESSIO...
cscpconf
 
File system vs DBMS
File system vs DBMSFile system vs DBMS
File system vs DBMS
BHARATH KUMAR
 
Sql data shrink steps
Sql data shrink stepsSql data shrink steps
Sql data shrink steps
Manoj Agnihotri
 
Storage Cost vs. Performance: Which Problem are You Solving?
Storage Cost vs. Performance: Which Problem are You Solving?Storage Cost vs. Performance: Which Problem are You Solving?
Storage Cost vs. Performance: Which Problem are You Solving?
IBM India Smarter Computing
 
new approach of inter cross an efficient multilevel cache management policy
 new approach of inter cross  an efficient multilevel cache management policy new approach of inter cross  an efficient multilevel cache management policy
new approach of inter cross an efficient multilevel cache management policy
INFOGAIN PUBLICATION
 
Generic RLM White Paper
Generic RLM White PaperGeneric RLM White Paper
Generic RLM White Paper
Mike Mendola (mendola@comcast.net)
 
EVALUATE DATABASE COMPRESSION PERFORMANCE AND PARALLEL BACKUP
EVALUATE DATABASE COMPRESSION PERFORMANCE AND PARALLEL BACKUPEVALUATE DATABASE COMPRESSION PERFORMANCE AND PARALLEL BACKUP
EVALUATE DATABASE COMPRESSION PERFORMANCE AND PARALLEL BACKUP
ijdms
 
Intro to Data warehousing lecture 17
Intro to Data warehousing   lecture 17Intro to Data warehousing   lecture 17
Intro to Data warehousing lecture 17
AnwarrChaudary
 
Efficient usage of memory management in big data using “anti caching”
Efficient usage of memory management in big data using “anti caching”Efficient usage of memory management in big data using “anti caching”
Efficient usage of memory management in big data using “anti caching”
eSAT Journals
 
Run your queries 14X faster without any investment!
Run your queries 14X faster without any investment!Run your queries 14X faster without any investment!
Run your queries 14X faster without any investment!
Knoldus Inc.
 
I-Sieve: An inline High Performance Deduplication System Used in cloud storage
I-Sieve: An inline High Performance Deduplication System Used in cloud storageI-Sieve: An inline High Performance Deduplication System Used in cloud storage
I-Sieve: An inline High Performance Deduplication System Used in cloud storage
redpel dot com
 
Selecting efficient and reliable preservation strategies
Selecting efficient and reliable preservation strategiesSelecting efficient and reliable preservation strategies
Selecting efficient and reliable preservation strategies
Micah Altman
 
Amazon Redshift Masterclass
Amazon Redshift MasterclassAmazon Redshift Masterclass
Amazon Redshift Masterclass
Amazon Web Services
 
SiDe Enabled Reliable Replica Optimization
SiDe Enabled Reliable Replica OptimizationSiDe Enabled Reliable Replica Optimization
SiDe Enabled Reliable Replica Optimization
IJCSIS Research Publications
 
TSS03135-USEN-00_HR
TSS03135-USEN-00_HRTSS03135-USEN-00_HR
TSS03135-USEN-00_HR
Ed Ahl
 
Dba tuning
Dba tuningDba tuning
Backing up Microsoft Great Plains / Microsoft Dynamics GP
Backing up Microsoft Great Plains / Microsoft Dynamics GPBacking up Microsoft Great Plains / Microsoft Dynamics GP
Backing up Microsoft Great Plains / Microsoft Dynamics GP
Handy_Backup
 
Deduplication in Open Spurce Cloud
Deduplication in Open Spurce CloudDeduplication in Open Spurce Cloud
Deduplication in Open Spurce Cloud
Mangali Praveen Kumar
 

Similar to How to Reduce Your Database Total Cost of Ownership with TimescaleDB (20)

Combining IBM Real-time Compression and IBM ProtecTIER Deduplication
Combining IBM Real-time Compression and IBM ProtecTIER DeduplicationCombining IBM Real-time Compression and IBM ProtecTIER Deduplication
Combining IBM Real-time Compression and IBM ProtecTIER Deduplication
 
Real time database compression optimization using iterative length compressio...
Real time database compression optimization using iterative length compressio...Real time database compression optimization using iterative length compressio...
Real time database compression optimization using iterative length compressio...
 
REAL TIME DATABASE COMPRESSION OPTIMIZATION USING ITERATIVE LENGTH COMPRESSIO...
REAL TIME DATABASE COMPRESSION OPTIMIZATION USING ITERATIVE LENGTH COMPRESSIO...REAL TIME DATABASE COMPRESSION OPTIMIZATION USING ITERATIVE LENGTH COMPRESSIO...
REAL TIME DATABASE COMPRESSION OPTIMIZATION USING ITERATIVE LENGTH COMPRESSIO...
 
File system vs DBMS
File system vs DBMSFile system vs DBMS
File system vs DBMS
 
Sql data shrink steps
Sql data shrink stepsSql data shrink steps
Sql data shrink steps
 
Storage Cost vs. Performance: Which Problem are You Solving?
Storage Cost vs. Performance: Which Problem are You Solving?Storage Cost vs. Performance: Which Problem are You Solving?
Storage Cost vs. Performance: Which Problem are You Solving?
 
new approach of inter cross an efficient multilevel cache management policy
 new approach of inter cross  an efficient multilevel cache management policy new approach of inter cross  an efficient multilevel cache management policy
new approach of inter cross an efficient multilevel cache management policy
 
Generic RLM White Paper
Generic RLM White PaperGeneric RLM White Paper
Generic RLM White Paper
 
EVALUATE DATABASE COMPRESSION PERFORMANCE AND PARALLEL BACKUP
EVALUATE DATABASE COMPRESSION PERFORMANCE AND PARALLEL BACKUPEVALUATE DATABASE COMPRESSION PERFORMANCE AND PARALLEL BACKUP
EVALUATE DATABASE COMPRESSION PERFORMANCE AND PARALLEL BACKUP
 
Intro to Data warehousing lecture 17
Intro to Data warehousing   lecture 17Intro to Data warehousing   lecture 17
Intro to Data warehousing lecture 17
 
Efficient usage of memory management in big data using “anti caching”
Efficient usage of memory management in big data using “anti caching”Efficient usage of memory management in big data using “anti caching”
Efficient usage of memory management in big data using “anti caching”
 
Run your queries 14X faster without any investment!
Run your queries 14X faster without any investment!Run your queries 14X faster without any investment!
Run your queries 14X faster without any investment!
 
I-Sieve: An inline High Performance Deduplication System Used in cloud storage
I-Sieve: An inline High Performance Deduplication System Used in cloud storageI-Sieve: An inline High Performance Deduplication System Used in cloud storage
I-Sieve: An inline High Performance Deduplication System Used in cloud storage
 
Selecting efficient and reliable preservation strategies
Selecting efficient and reliable preservation strategiesSelecting efficient and reliable preservation strategies
Selecting efficient and reliable preservation strategies
 
Amazon Redshift Masterclass
Amazon Redshift MasterclassAmazon Redshift Masterclass
Amazon Redshift Masterclass
 
SiDe Enabled Reliable Replica Optimization
SiDe Enabled Reliable Replica OptimizationSiDe Enabled Reliable Replica Optimization
SiDe Enabled Reliable Replica Optimization
 
TSS03135-USEN-00_HR
TSS03135-USEN-00_HRTSS03135-USEN-00_HR
TSS03135-USEN-00_HR
 
Dba tuning
Dba tuningDba tuning
Dba tuning
 
Backing up Microsoft Great Plains / Microsoft Dynamics GP
Backing up Microsoft Great Plains / Microsoft Dynamics GPBacking up Microsoft Great Plains / Microsoft Dynamics GP
Backing up Microsoft Great Plains / Microsoft Dynamics GP
 
Deduplication in Open Spurce Cloud
Deduplication in Open Spurce CloudDeduplication in Open Spurce Cloud
Deduplication in Open Spurce Cloud
 

Recently uploaded

"Frontline Battles with DDoS: Best practices and Lessons Learned", Igor Ivaniuk
"Frontline Battles with DDoS: Best practices and Lessons Learned",  Igor Ivaniuk"Frontline Battles with DDoS: Best practices and Lessons Learned",  Igor Ivaniuk
"Frontline Battles with DDoS: Best practices and Lessons Learned", Igor Ivaniuk
Fwdays
 
Choosing The Best AWS Service For Your Website + API.pptx
Choosing The Best AWS Service For Your Website + API.pptxChoosing The Best AWS Service For Your Website + API.pptx
Choosing The Best AWS Service For Your Website + API.pptx
Brandon Minnick, MBA
 
“Temporal Event Neural Networks: A More Efficient Alternative to the Transfor...
“Temporal Event Neural Networks: A More Efficient Alternative to the Transfor...“Temporal Event Neural Networks: A More Efficient Alternative to the Transfor...
“Temporal Event Neural Networks: A More Efficient Alternative to the Transfor...
Edge AI and Vision Alliance
 
How to Interpret Trends in the Kalyan Rajdhani Mix Chart.pdf
How to Interpret Trends in the Kalyan Rajdhani Mix Chart.pdfHow to Interpret Trends in the Kalyan Rajdhani Mix Chart.pdf
How to Interpret Trends in the Kalyan Rajdhani Mix Chart.pdf
Chart Kalyan
 
Dandelion Hashtable: beyond billion requests per second on a commodity server
Dandelion Hashtable: beyond billion requests per second on a commodity serverDandelion Hashtable: beyond billion requests per second on a commodity server
Dandelion Hashtable: beyond billion requests per second on a commodity server
Antonios Katsarakis
 
Programming Foundation Models with DSPy - Meetup Slides
Programming Foundation Models with DSPy - Meetup SlidesProgramming Foundation Models with DSPy - Meetup Slides
Programming Foundation Models with DSPy - Meetup Slides
Zilliz
 
Main news related to the CCS TSI 2023 (2023/1695)
Main news related to the CCS TSI 2023 (2023/1695)Main news related to the CCS TSI 2023 (2023/1695)
Main news related to the CCS TSI 2023 (2023/1695)
Jakub Marek
 
Harnessing the Power of NLP and Knowledge Graphs for Opioid Research
Harnessing the Power of NLP and Knowledge Graphs for Opioid ResearchHarnessing the Power of NLP and Knowledge Graphs for Opioid Research
Harnessing the Power of NLP and Knowledge Graphs for Opioid Research
Neo4j
 
Driving Business Innovation: Latest Generative AI Advancements & Success Story
Driving Business Innovation: Latest Generative AI Advancements & Success StoryDriving Business Innovation: Latest Generative AI Advancements & Success Story
Driving Business Innovation: Latest Generative AI Advancements & Success Story
Safe Software
 
What is an RPA CoE? Session 1 – CoE Vision
What is an RPA CoE?  Session 1 – CoE VisionWhat is an RPA CoE?  Session 1 – CoE Vision
What is an RPA CoE? Session 1 – CoE Vision
DianaGray10
 
JavaLand 2024: Application Development Green Masterplan
JavaLand 2024: Application Development Green MasterplanJavaLand 2024: Application Development Green Masterplan
JavaLand 2024: Application Development Green Masterplan
Miro Wengner
 
Fueling AI with Great Data with Airbyte Webinar
Fueling AI with Great Data with Airbyte WebinarFueling AI with Great Data with Airbyte Webinar
Fueling AI with Great Data with Airbyte Webinar
Zilliz
 
GNSS spoofing via SDR (Criptored Talks 2024)
GNSS spoofing via SDR (Criptored Talks 2024)GNSS spoofing via SDR (Criptored Talks 2024)
GNSS spoofing via SDR (Criptored Talks 2024)
Javier Junquera
 
Taking AI to the Next Level in Manufacturing.pdf
Taking AI to the Next Level in Manufacturing.pdfTaking AI to the Next Level in Manufacturing.pdf
Taking AI to the Next Level in Manufacturing.pdf
ssuserfac0301
 
Northern Engraving | Nameplate Manufacturing Process - 2024
Northern Engraving | Nameplate Manufacturing Process - 2024Northern Engraving | Nameplate Manufacturing Process - 2024
Northern Engraving | Nameplate Manufacturing Process - 2024
Northern Engraving
 
Nordic Marketo Engage User Group_June 13_ 2024.pptx
Nordic Marketo Engage User Group_June 13_ 2024.pptxNordic Marketo Engage User Group_June 13_ 2024.pptx
Nordic Marketo Engage User Group_June 13_ 2024.pptx
MichaelKnudsen27
 
Connector Corner: Seamlessly power UiPath Apps, GenAI with prebuilt connectors
Connector Corner: Seamlessly power UiPath Apps, GenAI with prebuilt connectorsConnector Corner: Seamlessly power UiPath Apps, GenAI with prebuilt connectors
Connector Corner: Seamlessly power UiPath Apps, GenAI with prebuilt connectors
DianaGray10
 
“How Axelera AI Uses Digital Compute-in-memory to Deliver Fast and Energy-eff...
“How Axelera AI Uses Digital Compute-in-memory to Deliver Fast and Energy-eff...“How Axelera AI Uses Digital Compute-in-memory to Deliver Fast and Energy-eff...
“How Axelera AI Uses Digital Compute-in-memory to Deliver Fast and Energy-eff...
Edge AI and Vision Alliance
 
Your One-Stop Shop for Python Success: Top 10 US Python Development Providers
Your One-Stop Shop for Python Success: Top 10 US Python Development ProvidersYour One-Stop Shop for Python Success: Top 10 US Python Development Providers
Your One-Stop Shop for Python Success: Top 10 US Python Development Providers
akankshawande
 

Recently uploaded (20)

"Frontline Battles with DDoS: Best practices and Lessons Learned", Igor Ivaniuk
"Frontline Battles with DDoS: Best practices and Lessons Learned",  Igor Ivaniuk"Frontline Battles with DDoS: Best practices and Lessons Learned",  Igor Ivaniuk
"Frontline Battles with DDoS: Best practices and Lessons Learned", Igor Ivaniuk
 
Choosing The Best AWS Service For Your Website + API.pptx
Choosing The Best AWS Service For Your Website + API.pptxChoosing The Best AWS Service For Your Website + API.pptx
Choosing The Best AWS Service For Your Website + API.pptx
 
“Temporal Event Neural Networks: A More Efficient Alternative to the Transfor...
“Temporal Event Neural Networks: A More Efficient Alternative to the Transfor...“Temporal Event Neural Networks: A More Efficient Alternative to the Transfor...
“Temporal Event Neural Networks: A More Efficient Alternative to the Transfor...
 
How to Interpret Trends in the Kalyan Rajdhani Mix Chart.pdf
How to Interpret Trends in the Kalyan Rajdhani Mix Chart.pdfHow to Interpret Trends in the Kalyan Rajdhani Mix Chart.pdf
How to Interpret Trends in the Kalyan Rajdhani Mix Chart.pdf
 
Dandelion Hashtable: beyond billion requests per second on a commodity server
Dandelion Hashtable: beyond billion requests per second on a commodity serverDandelion Hashtable: beyond billion requests per second on a commodity server
Dandelion Hashtable: beyond billion requests per second on a commodity server
 
Programming Foundation Models with DSPy - Meetup Slides
Programming Foundation Models with DSPy - Meetup SlidesProgramming Foundation Models with DSPy - Meetup Slides
Programming Foundation Models with DSPy - Meetup Slides
 
Main news related to the CCS TSI 2023 (2023/1695)
Main news related to the CCS TSI 2023 (2023/1695)Main news related to the CCS TSI 2023 (2023/1695)
Main news related to the CCS TSI 2023 (2023/1695)
 
Harnessing the Power of NLP and Knowledge Graphs for Opioid Research
Harnessing the Power of NLP and Knowledge Graphs for Opioid ResearchHarnessing the Power of NLP and Knowledge Graphs for Opioid Research
Harnessing the Power of NLP and Knowledge Graphs for Opioid Research
 
Driving Business Innovation: Latest Generative AI Advancements & Success Story
Driving Business Innovation: Latest Generative AI Advancements & Success StoryDriving Business Innovation: Latest Generative AI Advancements & Success Story
Driving Business Innovation: Latest Generative AI Advancements & Success Story
 
Artificial Intelligence and Electronic Warfare
Artificial Intelligence and Electronic WarfareArtificial Intelligence and Electronic Warfare
Artificial Intelligence and Electronic Warfare
 
What is an RPA CoE? Session 1 – CoE Vision
What is an RPA CoE?  Session 1 – CoE VisionWhat is an RPA CoE?  Session 1 – CoE Vision
What is an RPA CoE? Session 1 – CoE Vision
 
JavaLand 2024: Application Development Green Masterplan
JavaLand 2024: Application Development Green MasterplanJavaLand 2024: Application Development Green Masterplan
JavaLand 2024: Application Development Green Masterplan
 
Fueling AI with Great Data with Airbyte Webinar
Fueling AI with Great Data with Airbyte WebinarFueling AI with Great Data with Airbyte Webinar
Fueling AI with Great Data with Airbyte Webinar
 
GNSS spoofing via SDR (Criptored Talks 2024)
GNSS spoofing via SDR (Criptored Talks 2024)GNSS spoofing via SDR (Criptored Talks 2024)
GNSS spoofing via SDR (Criptored Talks 2024)
 
Taking AI to the Next Level in Manufacturing.pdf
Taking AI to the Next Level in Manufacturing.pdfTaking AI to the Next Level in Manufacturing.pdf
Taking AI to the Next Level in Manufacturing.pdf
 
Northern Engraving | Nameplate Manufacturing Process - 2024
Northern Engraving | Nameplate Manufacturing Process - 2024Northern Engraving | Nameplate Manufacturing Process - 2024
Northern Engraving | Nameplate Manufacturing Process - 2024
 
Nordic Marketo Engage User Group_June 13_ 2024.pptx
Nordic Marketo Engage User Group_June 13_ 2024.pptxNordic Marketo Engage User Group_June 13_ 2024.pptx
Nordic Marketo Engage User Group_June 13_ 2024.pptx
 
Connector Corner: Seamlessly power UiPath Apps, GenAI with prebuilt connectors
Connector Corner: Seamlessly power UiPath Apps, GenAI with prebuilt connectorsConnector Corner: Seamlessly power UiPath Apps, GenAI with prebuilt connectors
Connector Corner: Seamlessly power UiPath Apps, GenAI with prebuilt connectors
 
“How Axelera AI Uses Digital Compute-in-memory to Deliver Fast and Energy-eff...
“How Axelera AI Uses Digital Compute-in-memory to Deliver Fast and Energy-eff...“How Axelera AI Uses Digital Compute-in-memory to Deliver Fast and Energy-eff...
“How Axelera AI Uses Digital Compute-in-memory to Deliver Fast and Energy-eff...
 
Your One-Stop Shop for Python Success: Top 10 US Python Development Providers
Your One-Stop Shop for Python Success: Top 10 US Python Development ProvidersYour One-Stop Shop for Python Success: Top 10 US Python Development Providers
Your One-Stop Shop for Python Success: Top 10 US Python Development Providers
 

How to Reduce Your Database Total Cost of Ownership with TimescaleDB

  • 1. How to Reduce Your Database Total Cost of Ownership with TimescaleDB November 2019
  • 3. ● What’s New in TimescaleDB 1.5? ● Feature Deep Dive: Compression ● Feature Deep Dive: Move Chunks ● Questions Agenda
  • 4. THEME: Reduce Total Cost of Ownership (TCO) by optimizing use of storage ● Time-series use cases produce a LOT of data ● Users need cost effective and efficient way to store data ● Storage is usually the largest contributor to operational cost ● Users expects performant solution while delivering storage efficiencies ○ Ingestion and query performance cannot suffer ● Users need to be able to set a long term storage strategy for data ● Store data based on its access & usage pattern What’s New in TimescaleDB 1.5?
  • 5. What’s New in TimescaleDB 1.5? Capability Description Availability Native Compression Store more data with less actual storage. Community Edition Data Tiering (“move chunks”) Optimizes performance + storage costs by allowing you to map your data to the appropriate storage technology. Enterprise Edition
  • 6. Reducing Total Cost of Ownership Time-series data Time-series data STORAGE Compression enables you store more time-series data on less actual storage.
  • 7. Reducing Total Cost of Ownership Time-series data Time-series data STORAGE Compression enables you store more time-series data on less actual storage. Time-series data Time-series data STORAGE Time-series data Time-series data STORAGE Data Tiering enables you to target data to different storage types, optimizing cost & performance.
  • 9. Native Compression Overview Compression reduces the overall storage footprint of your data saves $ ● User can compression their data manually or by using a policy ● Compression can be applied to specific hypertables ● Once upgraded to 1.5, user can start applying compression policies to their hypertables NOTE: Upgrading to TSDB 1.5 does not automatically compress existing data or start compressing new data
  • 10. ● User drives compression policy based on: ○ Frequency of access ○ Query Characteristics (“Shallow and Wide” vs “Deep and Narrow”) ● Very adaptable: ○ Users can select “order by” and “segment by” to achieve best overall compression Native Compression: Sample Strategy
  • 11. Compression: How It Works Defining how data will be organized Use an “order by” parameter to move “like” data closer together to enhance compressibility and maintain performance. order by device_id
  • 12. Compression: How It Works Defining how data will be organized Use “segment by” operator to further group like data together. segment by device_id
  • 13. Using Compression: Manual Use Case: User wants to compress chunks manually In this use case we will identify a set of target chunks and compress them manually. Let's start by identifying chucks more than 3 months old: SELECT show_chunks('conditions', older_than => interval '3 days'); which will return a list of chunks. User can then compress one of these chunks using the following: SELECT compress_chunk('_timescaledb_internal._hyper_1_2_chunk');
  • 14. Using Compression: Policy Use Case: User wants to compress chunks automatically based on age In this case, as a user I have decided that I want to compress all chucks older than 60 days, without any manual intervention. SELECT add_compress_chunks_policy('cpu', '60d'::interval); This command will set up a background job that will kick off every 15 minutes, looking for chunks that are more than 60 days old and compressing them. We can see the job has been added by running: SELECT * FROM _timescaledb_config.bgw_job WHERE job_type LIKE 'compress%'; which will return the Jobs in the system associated with compression
  • 15. Compression Summary ● Compression can be configured based on user data model ● Compression strategy is driven by the user ● Real savings by reducing storage costs without removing data ● Why is compression important: ○ Reduce total cost of ownership by reducing storage costs ○ Allows user to retain more data without the additional costs COMMUNITY edition
  • 17. Move Chunks ● Reduce storage costs by leveraging more cost effective storage options ● Tier data based on usage / access pattern ● Optimize storage costs by moving infrequently used data to slower / more cost effective storage ● Combined with compression will allow for overall reduced TCO based on efficient use of storage
  • 19. Move Chunks: How It Works First step: Create a new tablespace Create a new tablespace, which will be backed by a new storage mount (in this example I have mount the new storage to /mnt/postgres): CREATE TABLESPACE tablespace_2 OWNER postgres LOCATION '/mnt/postgres'; Process of Moving the data: SELECT move_chunk(chunk=>'_timescaledb_internal._hyper_1_2_chunk', destination_tablespace=>'tablespace_2', index_destination_tablespace=>'tablespace_2', reorder_index=>'_timescaledb_internal._hyper_1_2_chunk_cluster_test_time_id x', verbose=>TRUE);
  • 21. Data Tiering Summary ● The ability to move data to cost effective storage ○ As data ages and is less frequently queried it can be moved to less expensive storage ● Help spread the data out over multiple arrays ● Leverages the ability to create tablespaces in Postgres and back them with different storage ENTERPRISE edition
  • 22. Learn more Native Compression Docs: tsdb.co/WhatsNewCompression Move Chunks Docs: tsdb.co/WhatsNewMoveChunk All products: tsdb.co/WhatsNewProducts Get the latest news and join our community ● blog.timescale.com ● twitter.com/TimescaleDB ● github.com/timescale/TimescaleDB ● slack.timescale.com