Using a Field Programmable Gate Array to Accelerate Application PerformanceOdinot Stanislas
Intel s'intéresse tout particulièrement aux FPGA et notamment au potentiel qu'ils apportent lorsque les ISV et développeurs ont des besoins très spécifiques en Génomique, traitement d'images, traitement de bases de données, et même dans le Cloud. Dans ce document vous aurez l'occasion d'en savoir plus sur notre stratégie, et sur un programme de recherche lancé par Intel et Altera impliquant des Xeon E5 équipés... de FPGA
Intel is looking at FPGA and what they bring to ISVs and developers and their very specific needs in genomics, image processing, databases, and even in the cloud. In this document you will have the opportunity to learn more about our strategy, and a research program initiated by Intel and Altera involving Xeon E5 with... FPGA inside.
Auteur(s)/Author(s):
P. K. Gupta, Director of Cloud Platform Technology, Intel Corporation
Short Survey on the current state of Field-programmable gate array usage in Deep learning by several companies like Intel Nervana and Google's TPU (tensor processing units) vs GPU usage in terms of energy consumption and performance.
DB Time, Average Active Sessions, and ASH Math - Oracle performance fundamentalsJohn Beresniewicz
RMOUG 2020 abstract:
This session will cover core concepts for Oracle performance analysis first introduced in Oracle 10g and forming the backbone of many features in the Diagnostic and Tuning packs. The presentation will cover the theoretical basis and meaning of these concepts, as well as illustrate how they are fundamental to many user-facing features in both the database itself and Enterprise Manager.
Backup Options for IBM PureData for Analytics powered by NetezzaTony Pearson
Confused about what options there are to backup your Netezza or IBM PureData for Analytics solution? This presentation provides alternatives related to file system and external backup software approaches using IBM Storwize V7000 Unified and IBM Tivoli Storage Manager
Using a Field Programmable Gate Array to Accelerate Application PerformanceOdinot Stanislas
Intel s'intéresse tout particulièrement aux FPGA et notamment au potentiel qu'ils apportent lorsque les ISV et développeurs ont des besoins très spécifiques en Génomique, traitement d'images, traitement de bases de données, et même dans le Cloud. Dans ce document vous aurez l'occasion d'en savoir plus sur notre stratégie, et sur un programme de recherche lancé par Intel et Altera impliquant des Xeon E5 équipés... de FPGA
Intel is looking at FPGA and what they bring to ISVs and developers and their very specific needs in genomics, image processing, databases, and even in the cloud. In this document you will have the opportunity to learn more about our strategy, and a research program initiated by Intel and Altera involving Xeon E5 with... FPGA inside.
Auteur(s)/Author(s):
P. K. Gupta, Director of Cloud Platform Technology, Intel Corporation
Short Survey on the current state of Field-programmable gate array usage in Deep learning by several companies like Intel Nervana and Google's TPU (tensor processing units) vs GPU usage in terms of energy consumption and performance.
DB Time, Average Active Sessions, and ASH Math - Oracle performance fundamentalsJohn Beresniewicz
RMOUG 2020 abstract:
This session will cover core concepts for Oracle performance analysis first introduced in Oracle 10g and forming the backbone of many features in the Diagnostic and Tuning packs. The presentation will cover the theoretical basis and meaning of these concepts, as well as illustrate how they are fundamental to many user-facing features in both the database itself and Enterprise Manager.
Backup Options for IBM PureData for Analytics powered by NetezzaTony Pearson
Confused about what options there are to backup your Netezza or IBM PureData for Analytics solution? This presentation provides alternatives related to file system and external backup software approaches using IBM Storwize V7000 Unified and IBM Tivoli Storage Manager
Approximation techniques used for general purpose algorithmsSabidur Rahman
Survey on approximation techniques used for general purpose algorithms, data parallel applications ans solid-state memories. It is interesting to see how approximation algorithms can contribute to solve real-life problems with better efficiency and lower cost!
Questions? krahman@ucdavis.edu.
Small updates for this version presented at OakTableWorld 2018
Discusses Oracle time-based performance instrumentation as presented in AWR reports and inconsistencies between instrumentation sources that can cause confusion as conflicting information is presented. The cognitive load of investigating and reasoning about such conundrums is very high, discouraging even senior performance experts. A program (AWR1page) is discussed that consumes an AWR report and produces a 1-page normalized time summary by instrumentation source, precisely designed for reasoning about instrumentation inconsistencies in AWR reports.
This presentation from 2008 is a good summary of Design by Contract and its application to PL/SQL as I have adopted and recommend others to try as well.
The IBM Netezza Data Warehouse ApplianceIBM Sverige
Netezza - Ett enklare sätt till smart analys.
Denna presentation hölls på IBM Data Server Day den 22 maj i Stockholm av Jacques Milman, Datawarehouse Architecture Leader, IBM
Gary Paek from Intel presented this deck at the HPC User Forum in Tucson.
Learn more: https://software.intel.com/en-us/tags/18892
and
http://hpcuserforum.com
Watch the video presentation: http://wp.me/p3RLHQ-fdt
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Vectorized Deep Learning Acceleration from Preprocessing to Inference and Tra...Databricks
This talk presents how we accelerated deep learning processing from preprocessing to inference and training on Apache Spark in SK Telecom. In SK Telecom, we have half a Korean population as our customers. To support them, we have 400,000 cell towers, which generates logs with geospatial tags.
Approximation techniques used for general purpose algorithmsSabidur Rahman
Survey on approximation techniques used for general purpose algorithms, data parallel applications ans solid-state memories. It is interesting to see how approximation algorithms can contribute to solve real-life problems with better efficiency and lower cost!
Questions? krahman@ucdavis.edu.
Small updates for this version presented at OakTableWorld 2018
Discusses Oracle time-based performance instrumentation as presented in AWR reports and inconsistencies between instrumentation sources that can cause confusion as conflicting information is presented. The cognitive load of investigating and reasoning about such conundrums is very high, discouraging even senior performance experts. A program (AWR1page) is discussed that consumes an AWR report and produces a 1-page normalized time summary by instrumentation source, precisely designed for reasoning about instrumentation inconsistencies in AWR reports.
This presentation from 2008 is a good summary of Design by Contract and its application to PL/SQL as I have adopted and recommend others to try as well.
The IBM Netezza Data Warehouse ApplianceIBM Sverige
Netezza - Ett enklare sätt till smart analys.
Denna presentation hölls på IBM Data Server Day den 22 maj i Stockholm av Jacques Milman, Datawarehouse Architecture Leader, IBM
Gary Paek from Intel presented this deck at the HPC User Forum in Tucson.
Learn more: https://software.intel.com/en-us/tags/18892
and
http://hpcuserforum.com
Watch the video presentation: http://wp.me/p3RLHQ-fdt
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Vectorized Deep Learning Acceleration from Preprocessing to Inference and Tra...Databricks
This talk presents how we accelerated deep learning processing from preprocessing to inference and training on Apache Spark in SK Telecom. In SK Telecom, we have half a Korean population as our customers. To support them, we have 400,000 cell towers, which generates logs with geospatial tags.
Accelerate Your Apache Spark with Intel Optane DC Persistent MemoryDatabricks
The capacity of data grows rapidly in big data area, more and more memory are consumed either in the computation or holding the intermediate data for analytic jobs. For those memory intensive workloads, end-point users have to scale out the computation cluster or extend memory with storage like HDD or SSD to meet the requirement of computing tasks. For scaling out the cluster, the extra cost from cluster management, operation and maintenance will increase the total cost if the extra CPU resources are not fully utilized. To address the shortcoming above, Intel Optane DC persistent memory (Optane DCPM) breaks the traditional memory/storage hierarchy and scale up the computing server with higher capacity persistent memory. Also it brings higher bandwidth & lower latency than storage like SSD or HDD. And Apache Spark is widely used in the analytics like SQL and Machine Learning on the cloud environment. For cloud environment, low performance of remote data access is typical a stop gap for users especially for some I/O intensive queries. For the ML workload, it's an iterative model which I/O bandwidth is the key to the end-2-end performance. In this talk, we will introduce how to accelerate Spark SQL with OAP (https://github.com/Intel-bigdata/OAP) to accelerate SQL performance on Cloud to archive 8X performance gain and RDD cache to improve K-means performance with 2.5X performance gain leveraging Intel Optane DCPM. Also we will have a deep dive how Optane DCPM for these performance gains.
Speakers: Cheng Xu, Piotr Balcer
DAOS - Scale-Out Software-Defined Storage for HPC/Big Data/AI Convergenceinside-BigData.com
In this deck, Johann Lombardi from Intel presents: DAOS - Scale-Out Software-Defined Storage for HPC/Big Data/AI Convergence.
"Intel has been building an entirely open source software ecosystem for data-centric computing, fully optimized for Intel® architecture and non-volatile memory (NVM) technologies, including Intel Optane DC persistent memory and Intel Optane DC SSDs. Distributed Asynchronous Object Storage (DAOS) is the foundation of the Intel exascale storage stack. DAOS is an open source software-defined scale-out object store that provides high bandwidth, low latency, and high I/O operations per second (IOPS) storage containers to HPC applications. It enables next-generation data-centric workflows that combine simulation, data analytics, and AI."
Unlike traditional storage stacks that were primarily designed for rotating media, DAOS is architected from the ground up to make use of new NVM technologies, and it is extremely lightweight because it operates end-to-end in user space with full operating system bypass. DAOS offers a shift away from an I/O model designed for block-based, high-latency storage to one that inherently supports fine- grained data access and unlocks the performance of next- generation storage technologies.
Watch the video: https://youtu.be/wnGBW31yhLM
Learn more: https://www.intel.com/content/www/us/en/high-performance-computing/daos-high-performance-storage-brief.html
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Even though there have been a large number of proposals to accelerate databases using specialized hardware, often the opinion of the community is pessimistic: the performance and energy efficiency benefits of specialization are seen to be outweighed by the limitations of the proposed solutions and the additional complexity of including specialized hardware, such as field programmable gate arrays (FPGAs), in servers. Recently, however, as an effect of stagnating CPU performance, server architectures started to incorporate various programmable hardware and the availability of such components brings opportunities to databases. In the light of a shifting hardware landscape and emerging analytics workloads, it is time to revisit our stance on hardware acceleration. In this talk we highlight several challenges that have traditionally hindered the deployment of hardware acceleration in databases and explain how they have been alleviated or removed altogether by recent research results and the changing hardware landscape. We also highlight a new set of questions that emerge around deep integration of heterogeneous programmable hardware in tomorrow’s databases.
Spring Hill (NNP-I 1000): Intel's Data Center Inference Chipinside-BigData.com
Today at Hot Chips 2019, Intel revealed new details of upcoming high-performance AI accelerators: Intel Nervana neural network processors, with the NNP-T for training and the NNP-I for inference. Intel engineers also presented technical details on hybrid chip packaging technology, Intel Optane DC persistent memory and chiplet technology for optical I/O.
"To get to a future state of ‘AI everywhere,’ we’ll need to address the crush of data being generated and ensure enterprises are empowered to make efficient use of their data, processing it where it’s collected when it makes sense and making smarter use of their upstream resources," said Naveen Rao, Intel vice president and GM, Artificial Intelligence Products Group. "Data centers and the cloud need to have access to performant and scalable general purpose computing and specialized acceleration for complex AI applications. In this future vision of AI everywhere, a holistic approach is needed—from hardware to software to applications.”
Learn more: https://www.intel.ai/accelerating-for-ai/?elq_cid=1192980
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Cómo se diseña una base de datos que pueda ingerir más de cuatro millones de ...javier ramirez
En esta sesión voy a contar las decisiones técnicas que tomamos al desarrollar QuestDB, una base de datos Open Source para series temporales compatible con Postgres, y cómo conseguimos escribir más de cuatro millones de filas por segundo sin bloquear o enlentecer las consultas.
Hablaré de cosas como (zero) Garbage Collection, vectorización de instrucciones usando SIMD, reescribir en lugar de reutilizar para arañar microsegundos, aprovecharse de los avances en procesadores, discos duros y sistemas operativos, como por ejemplo el soporte de io_uring, o del balance entre experiencia de usuario y rendimiento cuando se plantean nuevas funcionalidades.
Best Practice of Compression/Decompression Codes in Apache Spark with Sophia...Databricks
Nowadays, people are creating, sharing and storing data at a faster pace than ever before, effective data compression / decompression could significantly reduce the cost of data usage. Apache Spark is a general distributed computing engine for big data analytics, and it has large amount of data storing and shuffling across cluster in runtime, the data compression/decompression codecs can impact the end to end application performance in many ways.
However, there’s a trade-off between the storage size and compression/decompression throughput (CPU computation). Balancing the data compress speed and ratio is a very interesting topic, particularly while both software algorithms and the CPU instruction set keep evolving. Apache Spark provides a very flexible compression codecs interface with default implementations like GZip, Snappy, LZ4, ZSTD etc. and Intel Big Data Technologies team also implemented more codecs based on latest Intel platform like ISA-L(igzip), LZ4-IPP, Zlib-IPP and ZSTD for Apache Spark; in this session, we’d like to compare the characteristics of those algorithms and implementations, by running different micro workloads as well as end to end workloads, based on different generations of Intel x86 platform and disk.
It’s supposedly to be the best practice for big data software engineers to choose the proper compression/decompression codecs for their applications, and we also will present the methodologies of measuring and tuning the performance bottlenecks for typical Apache Spark workloads.
Best Practices for Building Robust Data Platform with Apache Spark and DeltaDatabricks
This talk will focus on Journey of technical challenges, trade offs and ground-breaking achievements for building performant and scalable pipelines from the experience working with our customers.
Innovating Inference - Remote Triggering of Large Language Models on HPC Clus...Globus
Large Language Models (LLMs) are currently the center of attention in the tech world, particularly for their potential to advance research. In this presentation, we'll explore a straightforward and effective method for quickly initiating inference runs on supercomputers using the vLLM tool with Globus Compute, specifically on the Polaris system at ALCF. We'll begin by briefly discussing the popularity and applications of LLMs in various fields. Following this, we will introduce the vLLM tool, and explain how it integrates with Globus Compute to efficiently manage LLM operations on Polaris. Attendees will learn the practical aspects of setting up and remotely triggering LLMs from local machines, focusing on ease of use and efficiency. This talk is ideal for researchers and practitioners looking to leverage the power of LLMs in their work, offering a clear guide to harnessing supercomputing resources for quick and effective LLM inference.
Exploring Innovations in Data Repository Solutions - Insights from the U.S. G...Globus
The U.S. Geological Survey (USGS) has made substantial investments in meeting evolving scientific, technical, and policy driven demands on storing, managing, and delivering data. As these demands continue to grow in complexity and scale, the USGS must continue to explore innovative solutions to improve its management, curation, sharing, delivering, and preservation approaches for large-scale research data. Supporting these needs, the USGS has partnered with the University of Chicago-Globus to research and develop advanced repository components and workflows leveraging its current investment in Globus. The primary outcome of this partnership includes the development of a prototype enterprise repository, driven by USGS Data Release requirements, through exploration and implementation of the entire suite of the Globus platform offerings, including Globus Flow, Globus Auth, Globus Transfer, and Globus Search. This presentation will provide insights into this research partnership, introduce the unique requirements and challenges being addressed and provide relevant project progress.
Understanding Globus Data Transfers with NetSageGlobus
NetSage is an open privacy-aware network measurement, analysis, and visualization service designed to help end-users visualize and reason about large data transfers. NetSage traditionally has used a combination of passive measurements, including SNMP and flow data, as well as active measurements, mainly perfSONAR, to provide longitudinal network performance data visualization. It has been deployed by dozens of networks world wide, and is supported domestically by the Engagement and Performance Operations Center (EPOC), NSF #2328479. We have recently expanded the NetSage data sources to include logs for Globus data transfers, following the same privacy-preserving approach as for Flow data. Using the logs for the Texas Advanced Computing Center (TACC) as an example, this talk will walk through several different example use cases that NetSage can answer, including: Who is using Globus to share data with my institution, and what kind of performance are they able to achieve? How many transfers has Globus supported for us? Which sites are we sharing the most data with, and how is that changing over time? How is my site using Globus to move data internally, and what kind of performance do we see for those transfers? What percentage of data transfers at my institution used Globus, and how did the overall data transfer performance compare to the Globus users?
OpenFOAM solver for Helmholtz equation, helmholtzFoam / helmholtzBubbleFoamtakuyayamamoto1800
In this slide, we show the simulation example and the way to compile this solver.
In this solver, the Helmholtz equation can be solved by helmholtzFoam. Also, the Helmholtz equation with uniformly dispersed bubbles can be simulated by helmholtzBubbleFoam.
Enhancing Project Management Efficiency_ Leveraging AI Tools like ChatGPT.pdfJay Das
With the advent of artificial intelligence or AI tools, project management processes are undergoing a transformative shift. By using tools like ChatGPT, and Bard organizations can empower their leaders and managers to plan, execute, and monitor projects more effectively.
Enhancing Research Orchestration Capabilities at ORNL.pdfGlobus
Cross-facility research orchestration comes with ever-changing constraints regarding the availability and suitability of various compute and data resources. In short, a flexible data and processing fabric is needed to enable the dynamic redirection of data and compute tasks throughout the lifecycle of an experiment. In this talk, we illustrate how we easily leveraged Globus services to instrument the ACE research testbed at the Oak Ridge Leadership Computing Facility with flexible data and task orchestration capabilities.
Software Engineering, Software Consulting, Tech Lead.
Spring Boot, Spring Cloud, Spring Core, Spring JDBC, Spring Security,
Spring Transaction, Spring MVC,
Log4j, REST/SOAP WEB-SERVICES.
top nidhi software solution freedownloadvrstrong314
This presentation emphasizes the importance of data security and legal compliance for Nidhi companies in India. It highlights how online Nidhi software solutions, like Vector Nidhi Software, offer advanced features tailored to these needs. Key aspects include encryption, access controls, and audit trails to ensure data security. The software complies with regulatory guidelines from the MCA and RBI and adheres to Nidhi Rules, 2014. With customizable, user-friendly interfaces and real-time features, these Nidhi software solutions enhance efficiency, support growth, and provide exceptional member services. The presentation concludes with contact information for further inquiries.
Large Language Models and the End of ProgrammingMatt Welsh
Talk by Matt Welsh at Craft Conference 2024 on the impact that Large Language Models will have on the future of software development. In this talk, I discuss the ways in which LLMs will impact the software industry, from replacing human software developers with AI, to replacing conventional software with models that perform reasoning, computation, and problem-solving.
May Marketo Masterclass, London MUG May 22 2024.pdfAdele Miller
Can't make Adobe Summit in Vegas? No sweat because the EMEA Marketo Engage Champions are coming to London to share their Summit sessions, insights and more!
This is a MUG with a twist you don't want to miss.
Listen to the keynote address and hear about the latest developments from Rachana Ananthakrishnan and Ian Foster who review the updates to the Globus Platform and Service, and the relevance of Globus to the scientific community as an automation platform to accelerate scientific discovery.
How to Position Your Globus Data Portal for Success Ten Good PracticesGlobus
Science gateways allow science and engineering communities to access shared data, software, computing services, and instruments. Science gateways have gained a lot of traction in the last twenty years, as evidenced by projects such as the Science Gateways Community Institute (SGCI) and the Center of Excellence on Science Gateways (SGX3) in the US, The Australian Research Data Commons (ARDC) and its platforms in Australia, and the projects around Virtual Research Environments in Europe. A few mature frameworks have evolved with their different strengths and foci and have been taken up by a larger community such as the Globus Data Portal, Hubzero, Tapis, and Galaxy. However, even when gateways are built on successful frameworks, they continue to face the challenges of ongoing maintenance costs and how to meet the ever-expanding needs of the community they serve with enhanced features. It is not uncommon that gateways with compelling use cases are nonetheless unable to get past the prototype phase and become a full production service, or if they do, they don't survive more than a couple of years. While there is no guaranteed pathway to success, it seems likely that for any gateway there is a need for a strong community and/or solid funding streams to create and sustain its success. With over twenty years of examples to draw from, this presentation goes into detail for ten factors common to successful and enduring gateways that effectively serve as best practices for any new or developing gateway.
Globus Connect Server Deep Dive - GlobusWorld 2024Globus
We explore the Globus Connect Server (GCS) architecture and experiment with advanced configuration options and use cases. This content is targeted at system administrators who are familiar with GCS and currently operate—or are planning to operate—broader deployments at their institution.
First Steps with Globus Compute Multi-User EndpointsGlobus
In this presentation we will share our experiences around getting started with the Globus Compute multi-user endpoint. Working with the Pharmacology group at the University of Auckland, we have previously written an application using Globus Compute that can offload computationally expensive steps in the researcher's workflows, which they wish to manage from their familiar Windows environments, onto the NeSI (New Zealand eScience Infrastructure) cluster. Some of the challenges we have encountered were that each researcher had to set up and manage their own single-user globus compute endpoint and that the workloads had varying resource requirements (CPUs, memory and wall time) between different runs. We hope that the multi-user endpoint will help to address these challenges and share an update on our progress here.
Field Employee Tracking System| MiTrack App| Best Employee Tracking Solution|...informapgpstrackings
Keep tabs on your field staff effortlessly with Informap Technology Centre LLC. Real-time tracking, task assignment, and smart features for efficient management. Request a live demo today!
For more details, visit us : https://informapuae.com/field-staff-tracking/
In software engineering, the right architecture is essential for robust, scalable platforms. Wix has undergone a pivotal shift from event sourcing to a CRUD-based model for its microservices. This talk will chart the course of this pivotal journey.
Event sourcing, which records state changes as immutable events, provided robust auditing and "time travel" debugging for Wix Stores' microservices. Despite its benefits, the complexity it introduced in state management slowed development. Wix responded by adopting a simpler, unified CRUD model. This talk will explore the challenges of event sourcing and the advantages of Wix's new "CRUD on steroids" approach, which streamlines API integration and domain event management while preserving data integrity and system resilience.
Participants will gain valuable insights into Wix's strategies for ensuring atomicity in database updates and event production, as well as caching, materialization, and performance optimization techniques within a distributed system.
Join us to discover how Wix has mastered the art of balancing simplicity and extensibility, and learn how the re-adoption of the modest CRUD has turbocharged their development velocity, resilience, and scalability in a high-growth environment.
2. System Technologies & Optimization (STO) 2
Agenda
Quick Overview of Impala
Design Challenges of an Impala Deployment
Case Study: Use Simulation-Based Approach to Design and Optimize an Impala
Cluster
What’s in side: Intel Cofluent Technology for Big Data
3. System Technologies & Optimization (STO) 3
Impala Overview
Open-ource MPP query execution engine
Built natively for Hadoop
Efficiently access data stored in Hadoop using SQL
Piplined execution mode enables fast data processing speed
4. System Technologies & Optimization (STO) 4
Design Challenges of an Impala Cluster – H/W
Meet Performance
Requirements
Plan For
the Future
Not Over
Provisioning
10 GB
50GB
1TB
5TB
10TB
5. System Technologies & Optimization (STO) 5
Example: Cluster Sizing
Requirements: a deep data analytic query over historical data should response
within 10 seconds
6. System Technologies & Optimization (STO) 6
Example: Storage Choice of One Use Case
~0.0448%
In general, SSD is faster than HDD, but there’re exceptions
7. System Technologies & Optimization (STO)
• No impact on the illustrated workload running on the Text formatted table
• Scaling well when running on the Parquet formatted table
7
Example: CPU Frequency
8. System Technologies & Optimization (STO) 8
Design Challenges of an Impala Cluster – S/W
HDFS Cache
HDFS Block Size
Parquet Row Group Size
Software Configuration
Options....
....
....
10. System Technologies & Optimization (STO) 10
Design Challenge Summary
We have talked about deployment
challenges, in terms of:
• hardware selections and settings
• software configuration choices
There’s NO ONE SIZE FIT-ALL
solution to the design challenges one
would face with when deploying a
system for production.
Efficient Way to Predict
System Performance?
Current Approach
11. System Technologies & Optimization (STO)
Simulation Approach
Deploy on
Experimental
Cluster
Generate
Simulation
Report
Collect and
Analyze
System Log Simulation Plan
Change H/W config
Change H/W knobs
Adjust WL setting
12. System Technologies & Optimization (STO) 12
• Impala Query Execution Simulation
• Query Planning Flow
• Plan Nodes, Plan Fragments, Execution
Nodes Geneation
• Task Scheduling and Distribution
• Data Processing Flow (Pull & Push)
• Data Distribution (Data Skew and
Partitioning)
• Disk IO Scheduling and Scan Operations
• Execution nodes
Impala Simulator Overview
13. System Technologies & Optimization (STO) 13
One Banking Use Case Study
• Offline Customer Account Historical Data Analysis
– Complex and Deep Analytic Queries
– Low Latency Interactive Queries
– Reporting Queries
• Initially evaluated on Hive, now Impala
14. System Technologies & Optimization (STO) 14
Step1: Deploy an Experimental Cluster
• Deploy a 4-node cluster
• Small scale of the data
17. System Technologies & Optimization (STO) 17
Step 3: Baseline Validation on Experimental Cluster
18. System Technologies & Optimization (STO) 18
Not just query execution time.
We also compare with Impala Log File to
check the duration of each stage
• disk-io-mgr.cc: disk id (1) reading for ....
• exchange.cc: #rows ... instance_id = ...
HashJoin Build Phase
HashJoin Probe Phase
Aggregation
Hdfs Scan Operation
Exchange Execution Node
Disk Worker 4
Disk Worker 0
19. System Technologies & Optimization (STO) 19
Step 4: From Experimental Cluster to Production Cluster
• We have completed baseline verification on an experimental cluster
• Performance prediction for the production cluster
• Simulation assumptions:
• upper- and lower- data distribution boundaries
• small scale of the data
20. System Technologies & Optimization (STO) 20
Step 5: Simulation Plan for Production Cluster
File Format Compression Partition
Text
GZIP
PartitionedAvro Snappy
No Partition
Cached
No Cache
Cache
Parquet
No
Compression
... ...
CPU Freq Netw ork Cluster Size
2.7Gz
...
42.4Gz 10GbE
2
SDD
HDD
Disk Type
2.1Gz
1GbE
6 ...
Software Configuration
Matrix
Hardware Configuration
Matrix
22. System Technologies & Optimization (STO) 22
Software Performance Predication
> 40GB
data to cache
23. System Technologies & Optimization (STO) 23
Cache Impact on Text Formatted Data
With Cache Without Cache
HdfsScanNode finishes
at around 6 sec
HdfsScanNode finishes at
around 12 sec
24. System Technologies & Optimization (STO) 24
Cache Impact on Text Formatted Data
Block for a short period
waiting for RowBatches Execution nodes are busy
processing RowBatches
25. System Technologies & Optimization (STO) 25
Cache Impact on Parquet Formatted Data
With Cache Without Cache
Fast Scan, CPU Bound
26. System Technologies & Optimization (STO) 26
Cache Impact on Parquet Formatted Data
CPU Bound,Scan Speed Does Not Have Impact on Overall
Performance of Query Execution.
27. System Technologies & Optimization (STO)
GZIPParquet Partitioned Cached
14.45% 2.74% 0.49%-9.22%
27
Reporting
Workload
Deep Analytic
Workload
Baseline
10x Files to
Scan
CPU
Intensive
Avro Snappy Partitioned Cached
1.1% 7.37% 7.94% 4.62%
Text
No
Compression
No Partition No Cache
Software Configuration Recommandation
33. System Technologies & Optimization (STO)
SCALE UP WITH CONFIDENCE:
Simulate to determine the minimum cost to meet
your future demand
FASTER CLUSTER DEPLOYMENT:
Explore deployment options and meet performance goals
OPTIMIZE CLUSTERS:
Find performance bottlenecks and optimize
software operation
Intel® CoFluent™ Technology for Big Data
34. System Technologies & Optimization (STO)
Intel® CoFluent™ Studio Based Simulation
Enables fast “What if?” analysis
with a virtual system
35. System Technologies & Optimization (STO)
Layered Simulation Architecture
H/W Resource Monitoring and Performance Library
CPU Memory Storage Ethernet …
Discrete Events Simulation Kernel on SystemC
Dynamic
S/W &
H/W
Mapping
S/W Stack
HBaseSpark M/R HDFS Impala
OS
JVM
…
System Topology
Role Assignment
Build a cluster
37. System Technologies & Optimization (STO)
Hardware Coverage
Validated: 50 Nodes SSD & HDD
1GbE & 10GbE
Rack Scale Architecture
Pooled
Compute
Pooled
Memory
Pooled
I/O
38. System Technologies & Optimization (STO) 38
Simulation Accuracy
High Simulation Accuracy is achieved for Big Data applications running on
different cluster size, hardware configurations and software stacks.
39. System Technologies & Optimization (STO)
Fast Simulation
7
18
36
71
2
6
14
29
20 50 100 200
NUMBER OF CONCURRENT UPLOADING REQUESTS
Simulation vs. Real Time in minutes
Hardware - 4 node Cluster (min)
Simulation Speed - Lenovo T420 (min)
Abstract
Modeling
Event Driven
Simulation
41. System Technologies & Optimization (STO) 41
Call to Actions
Visit cofluent.intel.com for more information
Request white papers
Various customer success stories and use cases available
– Optimize a 50-node Hive/MR Cluster
– Predict the scalability of a large HBase Cluster
– Software Parameter tunings for Spark Applications
– …
Demo in the showcase – Intel booth
44. System Technologies & Optimization (STO)
Risk FactorsThe above statements and any others in this document that refer to plans and expectations for the second quarter, the year and the future are forward-
looking statements that involve a number of risks and uncertainties. Words such as "anticipates," "expects," "intends," "plans," "believes," "seeks," "estimates,"
"may," "will," "should" and their variations identify forward-looking statements. Statements that refer to or are based on projections, uncertain events or
assumptions also identify forward-looking statements. Many factors could affect Intel's actual results, and variances from Intel's current expectations
regarding such factors could cause actual results to differ materially from those expressed in these forward-looking statements. Intel presently considers the
following to be important factors that could cause actual results to differ materially from the company's expectations. Demand for Intel's products is highly
variable and could differ from expectations due to factors including changes in business and economic conditions; consumer confidence or income levels;
the introduction, availability and market acceptance of Intel's products, products used together with Intel products and competitors' products; competitive
and pricing pressures, including actions taken by competitors; supply constraints and other disruptions affecting customers; changes in customer order
patterns including order cancellations; and changes in the level of inventory at customers. Intel's gross margin percentage could vary significantly from
expectations based on capacity utilization; variations in inventory valuation, including variations related to the timing of qualifying products for sale; changes
in revenue levels; segment product mix; the timing and execution of the manufacturing ramp and associated costs; excess or obsolete inventory; changes in
unit costs; defects or disruptions in the supply of materials or resources; and product manufacturing quality/yields. Variations in gross margin may also be
caused by the timing of Intel product introductions and related expenses, including marketing expenses, and Intel's ability to respond quickly to
technological developments and to introduce new products or incorporate new features into existing products, which may result in restructuring and asset
impairment charges. Intel's results could be affected by adverse economic, social, political and physical/infrastructure conditions in countries where Intel, its
customers or its suppliers operate, including military conflict and other security risks, natural disasters, infrastructure disruptions, health concerns and
fluctuations in currency exchange rates. Results may also be affected by the formal or informal imposition by countries of new or revised export and/or
import and doing-business regulations, which could be changed without prior notice. Intel operates in highly competitive industries and its operations have
high costs that are either fixed or difficult to reduce in the short term. The amount, timing and execution of Intel's stock repurchase program could be
affected by changes in Intel's priorities for the use of cash, such as operational spending, capital spending, acquisitions, and as a result of changes to Intel's
cash flows or changes in tax laws. Product defects or errata (deviations from published specifications) may adversely impact our expenses, revenues and
reputation. Intel's results could be affected by litigation or regulatory matters involving intellectual property, stockholder, consumer, antitrust, disclosure
and other issues. An unfavorable ruling could include monetary damages or an injunction prohibiting Intel from manufacturing or selling one or more
products, precluding particular business practices, impacting Intel's ability to design its products, or requiring other remedies such as compulsory licensing
of intellectual property. Intel's results may be affected by the timing of closing of acquisitions, divestitures and other significant transactions. A detailed
discussion of these and other factors that could affect Intel's results is included in Intel's SEC filings, including the company's most recent reports on Form
10-Q, Form 10-K and earnings release.
Rev. 4/14/15