Pileus is a cloud storage system that provides consistency-based service level agreements (SLAs) to applications. It offers a range of consistency choices between strong and eventual consistency. Pileus exports APIs that allow applications to specify desired consistency levels and latency targets via SLAs. The system enforces SLAs by selecting storage nodes that can meet the consistency and latency guarantees defined in the SLA. It uses client-side monitoring and adaptive techniques to satisfy SLAs even when network conditions vary. Evaluation results demonstrate that Pileus can better meet application-defined SLAs compared to fixed consistency approaches.
Beyond REST and RPC: Asynchronous Eventing and Messaging PatternsClemens Vasters
In this session you will learn about when and why to use asynchronous communication with and between services, what kind of eventing/messaging infrastructure you can use in the cloud and on the edge, and how to make it all work together.
High-Speed Reactive Microservices - trials and tribulationsRick Hightower
Covers how we built a set of high-speed reactive microservices and maximized cloud/hardware costs while meeting objectives in resilience and scalability. This has more notes attached as it is based on the ppt not the PDF.
Hannover Messe 2017 - Systems Federation in industrie 4.0Clemens Vasters
In the last couple of years, we have observed a rapid and, for many, quite surprising consolidation of standards in the industrial automation arena. Built on proven existing base standards and with an extensible model, the OPC Foundation's Unified Architecture has emerged as the clear winner for horizontal integration of production environments and vertical integration of such environments with their attached IT systems. In this talk, we will discuss the next set of challenges that lie ahead given this common foundation: Each production line may have dozens of different stakeholders, from sensor to system, who each want to provide predictive-maintenance, remote-management, or X-as-a-service within that one production line as their I4.0 business objective. How does a sensor manufacturer get access to what telemetry from their product inside a component that is part of a machine installed in a production line? And through what path might they make a machine-learned optimizing correction?
Whether you are developing a greenfield data project or migrating a legacy system,
there are many critical design decisions to be made. Often, it is advantageous to not only
consider immediate requirements, but also the future requirements and technologies you may
want to support. Your project may start out supporting batch analytics with the vision of adding
realtime support. Or your data pipeline may feed data to one technology today, but tomorrow
an entirely new system needs to be integrated. Apache Kafka can help decouple these
decisions and provide a flexible core to your data architecture. This talk will show how building
Kafka into your pipeline can provide the flexibility to experiment, evolve and grow. It will also
cover a brief overview of Kafka, its architecture, and terminology.
Data processing use cases, from transformation to analytics, perform tasks that require various combinations of queuing, streaming & lightweight processing steps. Until now, supporting all of those needs has required different systems for each task--stream processing engines, messaging queuing middleware, & streaming messaging systems. That has led to increased complexity for development & operations.
In this session, well discuss the need to unify these capabilities in a single system & how Apache Pulsar was designed to address that. Apache Pulsar is a next generation distributed pub-sub system that was developed & deployed at Yahoo. Streamlios Karthik Ramasamy, will explain how the architecture & design of Pulsar provides the flexibility to support developers & applications needing any combination of queuing, messaging, streaming & lightweight compute.
This session endeavors to explain high-speed reactive microservice architecture, a set of patterns for building services that can readily back mobile and web applications at scale. It uses a scale-up and -out versus a scale-out model to do more with less hardware. A scale-up model uses in-memory operational data, efficient queue handoff, and microbatch streaming, plus async calls to handle more calls on a single node. High-speed microservice architecture endeavors to get back to OOP roots, where data and logic live together in a cohesive, understandable representation of the problem domain, and away from separation of data and logic, because data lives with the service logic that operates on it.
Beyond REST and RPC: Asynchronous Eventing and Messaging PatternsClemens Vasters
In this session you will learn about when and why to use asynchronous communication with and between services, what kind of eventing/messaging infrastructure you can use in the cloud and on the edge, and how to make it all work together.
High-Speed Reactive Microservices - trials and tribulationsRick Hightower
Covers how we built a set of high-speed reactive microservices and maximized cloud/hardware costs while meeting objectives in resilience and scalability. This has more notes attached as it is based on the ppt not the PDF.
Hannover Messe 2017 - Systems Federation in industrie 4.0Clemens Vasters
In the last couple of years, we have observed a rapid and, for many, quite surprising consolidation of standards in the industrial automation arena. Built on proven existing base standards and with an extensible model, the OPC Foundation's Unified Architecture has emerged as the clear winner for horizontal integration of production environments and vertical integration of such environments with their attached IT systems. In this talk, we will discuss the next set of challenges that lie ahead given this common foundation: Each production line may have dozens of different stakeholders, from sensor to system, who each want to provide predictive-maintenance, remote-management, or X-as-a-service within that one production line as their I4.0 business objective. How does a sensor manufacturer get access to what telemetry from their product inside a component that is part of a machine installed in a production line? And through what path might they make a machine-learned optimizing correction?
Whether you are developing a greenfield data project or migrating a legacy system,
there are many critical design decisions to be made. Often, it is advantageous to not only
consider immediate requirements, but also the future requirements and technologies you may
want to support. Your project may start out supporting batch analytics with the vision of adding
realtime support. Or your data pipeline may feed data to one technology today, but tomorrow
an entirely new system needs to be integrated. Apache Kafka can help decouple these
decisions and provide a flexible core to your data architecture. This talk will show how building
Kafka into your pipeline can provide the flexibility to experiment, evolve and grow. It will also
cover a brief overview of Kafka, its architecture, and terminology.
Data processing use cases, from transformation to analytics, perform tasks that require various combinations of queuing, streaming & lightweight processing steps. Until now, supporting all of those needs has required different systems for each task--stream processing engines, messaging queuing middleware, & streaming messaging systems. That has led to increased complexity for development & operations.
In this session, well discuss the need to unify these capabilities in a single system & how Apache Pulsar was designed to address that. Apache Pulsar is a next generation distributed pub-sub system that was developed & deployed at Yahoo. Streamlios Karthik Ramasamy, will explain how the architecture & design of Pulsar provides the flexibility to support developers & applications needing any combination of queuing, messaging, streaming & lightweight compute.
This session endeavors to explain high-speed reactive microservice architecture, a set of patterns for building services that can readily back mobile and web applications at scale. It uses a scale-up and -out versus a scale-out model to do more with less hardware. A scale-up model uses in-memory operational data, efficient queue handoff, and microbatch streaming, plus async calls to handle more calls on a single node. High-speed microservice architecture endeavors to get back to OOP roots, where data and logic live together in a cohesive, understandable representation of the problem domain, and away from separation of data and logic, because data lives with the service logic that operates on it.
Kafka is a real-time, fault-tolerant, scalable messaging system.
It is a publish-subscribe system that connects various applications with the help of messages - producers and consumers of information.
Real-time Data Ingestion from Kafka to ClickHouse with Deterministic Re-tries...HostedbyConfluent
In a real-time data ingestion pipeline for analytical processing, efficient and fast data loading to a columnar database such as ClickHouse favors large blocks over individual rows. Therefore, applications often rely on some buffering mechanism such as Kafka to store data temporarily, and having a message processing engine to aggregate Kafka messages into large blocks which then get loaded to the backend database. Due to various failures in this pipeline, a naive block aggregator that forms blocks without additional measures, would cause data duplication or data loss. We have developed a solution to avoid these issues, thereby achieving exactly-once delivery from Kafka to ClickHouse. Our solution utilizes Kafka’s metadata to keep track of blocks that we intend to send to ClickHouse, and later uses this metadata information to deterministically re-produce ClickHouse blocks for re-tries in case of failures. The identical blocks are guaranteed to be deduplicated by ClickHouse. We have also developed a run-time verification tool that monitors Kafka’s internal metadata topic, and raises alerts when the required invariants for exactly-once delivery are violated. Our solution has been developed and deployed to the production clusters that span multiple datacenters at eBay.
Introduction and Overview of Apache Kafka, TriHUG July 23, 2013mumrah
Apache Kafka is a new breed of messaging system built for the "big data" world. Coming out of LinkedIn (and donated to Apache), it is a distributed pub/sub system built in Scala. It has been an Apache TLP now for several months with the first Apache release imminent. Built for speed, scalability, and robustness, Kafka should definitely be one of the data tools you consider when designing distributed data-oriented applications.
The talk will cover a general overview of the project and technology, with some use cases, and a demo.
Reactive Java: Promises and Streams with Reakt (JavaOne talk 2016)Rick Hightower
see labs at https://github.com/advantageous/j1-talks-2016
Import based on PDF. This is from our JavaOne Talk 2016 on Reakt, reactive Java programming with promises, circuit breakers, and streams. Reakt is a reactive Java lib that provides promises, streams, and a reactor to handle asynchronous call coordination. It was influenced by the design of promises in ES6. You want to async-call serviceA and then serviceB, take the results of serviceA and serviceB, and then call serviceC. Then, based on the results of call C, call D or E and then return the results to the original caller. Calls to A, B, C, D, and E are all async calls, and none should take longer than 10 seconds. If they do, then return a timeout to the original caller. The whole async call sequence should time out in 20 seconds if it does not complete and should also check for circuit breakers and provide back pressure feedback so the system does not have cascading failures. Learn more in this session.
KafkaConsumer - Decoupling Consumption and Processing for Better Resource Uti...confluent
When working with KafkaConsumer, we usually employ single thread both for reading and processing of messages. KafkaConsumer is not thread-safe, so using single thread fits in well. Downside of this approach is that you are limited to single thread for processing messages.
By decoupling consumption and processing, we can achieve processing parallelization with single consumer and get the most out of multi-core CPU architectures available today. While this can be very useful in certain use-case scenarios, it's not trivial to implement.
How do we use multiple threads with KafkaConsumer which is not thread safe? How do we react to consumer group rebalancing? Can we get desired processing and ordering guarantees? In this talk we 'll try to answer these questions and explore challenges we face on our path.
Kafka is most popular messaging queue.
Key Areas:
What is Messgaing Queue?
Why Messaging Queue?
Kafka- basic terminologies
Kafka- Architecture (Message Flow)
AWS SQS vs Apache Kafka
Strata+Hadoop 2017 San Jose: Lessons from a year of supporting Apache Kafkaconfluent
The number of deployments of Apache Kafka at enterprise scale has greatly increased in the years since Kafka’s original development in 2010. Along with this rapid growth has come a wide variety of use cases and deployment strategies that transcend what Kafka’s creators imagined when they originally developed the technology. As the scope and reach of streaming data platforms based on Apache Kafka has grown, the need to understand monitoring and troubleshooting strategies has as well.
Dustin Cote and Ryan Pridgeon share their experience supporting Apache Kafka at enterprise-scale and explore monitoring and troubleshooting techniques to help you avoid pitfalls when scaling large-scale Kafka deployments.
Topics include:
- Effective use of JMX for Kafka
- Tools for preventing small problems from becoming big ones
- Efficient architectures proven in the wild
- Finding and storing the right information when it all goes wrong
Visit www.confluent.io for more information.
Deep Dive into the Pulsar Binary Protocol - Pulsar Virtual Summit Europe 2021StreamNative
To achieve maximum performance, some important choices have been made when designing the Pulsar binary protocol.
This session will explain how Pulsar implements all the features of a high quality streaming protocol such as frame multiplexing, session establishment, keep-alive, flow control, authentication and authorisation, encoding, zero-copy capabilities and more.
RabbitMQ vs Kafka
Messaging is at the core of many architectures and two giants in the messaging space are RabbitMQ and Apache Kafka. In this webinar we'll take a look at RabbitMQ and Kafka within the context of real-time event-driven architectures.
In this session we’re joined by guest speaker Jack Vanlightly who will explore what RabbitMQ and Apache Kafka are and their approach to messaging. Each technology has made very different decisions regarding every aspect of their design, each with strengths and weaknesses, enabling different architectural patterns.
WEBINAR LIVE DATE: Wednesday 23 May 2018 | 17:30 CEST / 16:30 BST / 11:30 EDT / 08:30 PDT
Link to video: https://www.youtube.com/watch?v=sjDnqrnnYNM
———————————————————————
SPEAKER CONTACT DETAILS
JACK VANLIGHTLY - Jack Vanlightly is a software architect based in Barcelona specialising in event-driven architectures, data processing pipelines and data stores both relational and non-relational.
Twitter: https://twitter.com/vanlightly
———————————————————————
COMPANY CONTACT DETAILS
ERLANG SOLUTIONS
- Website: https://www.erlang-solutions.com
- Twitter: https://www.twitter.com/ErlangSolutions
- LinkedIn: http://www.linkedin.com/company/erlan…
- GitHub: https://github.com/esl
This talk covers Kafka cluster sizing, instance type selections, scaling operations, replication throttling and more. Don’t forget to check out the Kafka-Kit repository.
https://www.youtube.com/watch?time_continue=2613&v=7uN-Vlf7W5E
William Brander and Sean Farmar show how the monitoring game changes when a system becomes distributed and you start delving into the world of microservices.
Learn:
* Why monitoring changes in distributed systems
* A monitoring philosophy that ensures all bases are covered
* The aspects of monitoring that affect asynchronous messaging systems
Topic: Speedtest: Benchmark Your Apache Kafka®️
Abstract: In this session, Mark will talk about running benchmarking utilities for Apache Kafka; to determine how much MB/sec a cluster can handle; how to set up automated benchmark runs (including the repo), and using this to find and optimize client-side producer configuration properties
Hierarchical SLA-based Service Selection for Multi-Cloud EnvironmentsSoodeh Farokhi
Cloud computing popularity is growing rapidly and consequently the number of companies offering their services in the form of Software-as-a-Service (SaaS) or Infrastructure-as-a-Service (IaaS) is increasing. The diversity and usage benefits of IaaS offers are encouraging SaaS providers to lease resources from the Cloud instead of operating their own data centers. However, the question remains for them how to, on the one hand, exploit Cloud benefits to gain less maintenance overheads and on the other hand, maximize the satisfactions of customers with a wide range of requirements. The complexity of addressing these issues prevent many SaaS providers to benefit from the Cloud infrastructures. In this paper, we propose HS4MC approach for automatic service selection by considering SLA claims of SaaS providers. The novelty of our approach lies
in the utilization of prospect theory for the service ranking that represents a natural choice for scoring of comparable services due to the users preferences. The HS4MC approach first constructs a set of SLAs based on the given accumulated SaaS provider requirements. Then, it selects a set of services that best fulfills the SLAs. We evaluate our approach in a simulated environment by comparing it with a state-of-the-art utility based algorithm. The evaluation results show that our approach selects services that more effectively satisfy the SLAs.
Kafka is a real-time, fault-tolerant, scalable messaging system.
It is a publish-subscribe system that connects various applications with the help of messages - producers and consumers of information.
Real-time Data Ingestion from Kafka to ClickHouse with Deterministic Re-tries...HostedbyConfluent
In a real-time data ingestion pipeline for analytical processing, efficient and fast data loading to a columnar database such as ClickHouse favors large blocks over individual rows. Therefore, applications often rely on some buffering mechanism such as Kafka to store data temporarily, and having a message processing engine to aggregate Kafka messages into large blocks which then get loaded to the backend database. Due to various failures in this pipeline, a naive block aggregator that forms blocks without additional measures, would cause data duplication or data loss. We have developed a solution to avoid these issues, thereby achieving exactly-once delivery from Kafka to ClickHouse. Our solution utilizes Kafka’s metadata to keep track of blocks that we intend to send to ClickHouse, and later uses this metadata information to deterministically re-produce ClickHouse blocks for re-tries in case of failures. The identical blocks are guaranteed to be deduplicated by ClickHouse. We have also developed a run-time verification tool that monitors Kafka’s internal metadata topic, and raises alerts when the required invariants for exactly-once delivery are violated. Our solution has been developed and deployed to the production clusters that span multiple datacenters at eBay.
Introduction and Overview of Apache Kafka, TriHUG July 23, 2013mumrah
Apache Kafka is a new breed of messaging system built for the "big data" world. Coming out of LinkedIn (and donated to Apache), it is a distributed pub/sub system built in Scala. It has been an Apache TLP now for several months with the first Apache release imminent. Built for speed, scalability, and robustness, Kafka should definitely be one of the data tools you consider when designing distributed data-oriented applications.
The talk will cover a general overview of the project and technology, with some use cases, and a demo.
Reactive Java: Promises and Streams with Reakt (JavaOne talk 2016)Rick Hightower
see labs at https://github.com/advantageous/j1-talks-2016
Import based on PDF. This is from our JavaOne Talk 2016 on Reakt, reactive Java programming with promises, circuit breakers, and streams. Reakt is a reactive Java lib that provides promises, streams, and a reactor to handle asynchronous call coordination. It was influenced by the design of promises in ES6. You want to async-call serviceA and then serviceB, take the results of serviceA and serviceB, and then call serviceC. Then, based on the results of call C, call D or E and then return the results to the original caller. Calls to A, B, C, D, and E are all async calls, and none should take longer than 10 seconds. If they do, then return a timeout to the original caller. The whole async call sequence should time out in 20 seconds if it does not complete and should also check for circuit breakers and provide back pressure feedback so the system does not have cascading failures. Learn more in this session.
KafkaConsumer - Decoupling Consumption and Processing for Better Resource Uti...confluent
When working with KafkaConsumer, we usually employ single thread both for reading and processing of messages. KafkaConsumer is not thread-safe, so using single thread fits in well. Downside of this approach is that you are limited to single thread for processing messages.
By decoupling consumption and processing, we can achieve processing parallelization with single consumer and get the most out of multi-core CPU architectures available today. While this can be very useful in certain use-case scenarios, it's not trivial to implement.
How do we use multiple threads with KafkaConsumer which is not thread safe? How do we react to consumer group rebalancing? Can we get desired processing and ordering guarantees? In this talk we 'll try to answer these questions and explore challenges we face on our path.
Kafka is most popular messaging queue.
Key Areas:
What is Messgaing Queue?
Why Messaging Queue?
Kafka- basic terminologies
Kafka- Architecture (Message Flow)
AWS SQS vs Apache Kafka
Strata+Hadoop 2017 San Jose: Lessons from a year of supporting Apache Kafkaconfluent
The number of deployments of Apache Kafka at enterprise scale has greatly increased in the years since Kafka’s original development in 2010. Along with this rapid growth has come a wide variety of use cases and deployment strategies that transcend what Kafka’s creators imagined when they originally developed the technology. As the scope and reach of streaming data platforms based on Apache Kafka has grown, the need to understand monitoring and troubleshooting strategies has as well.
Dustin Cote and Ryan Pridgeon share their experience supporting Apache Kafka at enterprise-scale and explore monitoring and troubleshooting techniques to help you avoid pitfalls when scaling large-scale Kafka deployments.
Topics include:
- Effective use of JMX for Kafka
- Tools for preventing small problems from becoming big ones
- Efficient architectures proven in the wild
- Finding and storing the right information when it all goes wrong
Visit www.confluent.io for more information.
Deep Dive into the Pulsar Binary Protocol - Pulsar Virtual Summit Europe 2021StreamNative
To achieve maximum performance, some important choices have been made when designing the Pulsar binary protocol.
This session will explain how Pulsar implements all the features of a high quality streaming protocol such as frame multiplexing, session establishment, keep-alive, flow control, authentication and authorisation, encoding, zero-copy capabilities and more.
RabbitMQ vs Kafka
Messaging is at the core of many architectures and two giants in the messaging space are RabbitMQ and Apache Kafka. In this webinar we'll take a look at RabbitMQ and Kafka within the context of real-time event-driven architectures.
In this session we’re joined by guest speaker Jack Vanlightly who will explore what RabbitMQ and Apache Kafka are and their approach to messaging. Each technology has made very different decisions regarding every aspect of their design, each with strengths and weaknesses, enabling different architectural patterns.
WEBINAR LIVE DATE: Wednesday 23 May 2018 | 17:30 CEST / 16:30 BST / 11:30 EDT / 08:30 PDT
Link to video: https://www.youtube.com/watch?v=sjDnqrnnYNM
———————————————————————
SPEAKER CONTACT DETAILS
JACK VANLIGHTLY - Jack Vanlightly is a software architect based in Barcelona specialising in event-driven architectures, data processing pipelines and data stores both relational and non-relational.
Twitter: https://twitter.com/vanlightly
———————————————————————
COMPANY CONTACT DETAILS
ERLANG SOLUTIONS
- Website: https://www.erlang-solutions.com
- Twitter: https://www.twitter.com/ErlangSolutions
- LinkedIn: http://www.linkedin.com/company/erlan…
- GitHub: https://github.com/esl
This talk covers Kafka cluster sizing, instance type selections, scaling operations, replication throttling and more. Don’t forget to check out the Kafka-Kit repository.
https://www.youtube.com/watch?time_continue=2613&v=7uN-Vlf7W5E
William Brander and Sean Farmar show how the monitoring game changes when a system becomes distributed and you start delving into the world of microservices.
Learn:
* Why monitoring changes in distributed systems
* A monitoring philosophy that ensures all bases are covered
* The aspects of monitoring that affect asynchronous messaging systems
Topic: Speedtest: Benchmark Your Apache Kafka®️
Abstract: In this session, Mark will talk about running benchmarking utilities for Apache Kafka; to determine how much MB/sec a cluster can handle; how to set up automated benchmark runs (including the repo), and using this to find and optimize client-side producer configuration properties
Hierarchical SLA-based Service Selection for Multi-Cloud EnvironmentsSoodeh Farokhi
Cloud computing popularity is growing rapidly and consequently the number of companies offering their services in the form of Software-as-a-Service (SaaS) or Infrastructure-as-a-Service (IaaS) is increasing. The diversity and usage benefits of IaaS offers are encouraging SaaS providers to lease resources from the Cloud instead of operating their own data centers. However, the question remains for them how to, on the one hand, exploit Cloud benefits to gain less maintenance overheads and on the other hand, maximize the satisfactions of customers with a wide range of requirements. The complexity of addressing these issues prevent many SaaS providers to benefit from the Cloud infrastructures. In this paper, we propose HS4MC approach for automatic service selection by considering SLA claims of SaaS providers. The novelty of our approach lies
in the utilization of prospect theory for the service ranking that represents a natural choice for scoring of comparable services due to the users preferences. The HS4MC approach first constructs a set of SLAs based on the given accumulated SaaS provider requirements. Then, it selects a set of services that best fulfills the SLAs. We evaluate our approach in a simulated environment by comparing it with a state-of-the-art utility based algorithm. The evaluation results show that our approach selects services that more effectively satisfy the SLAs.
Semantic Validation: Enforcing Kafka Data Quality Through Schema-Driven Verif...HostedbyConfluent
"Incorrect data produced into Kafka can be a poison pill that has the potential to disrupt businesses built upon Kafka. The “Semantic Validation” feature is designed to address the challenges posed by incorrect or unexpected data in Kafka’s data processing pipelines, with the goal of mitigating such disruptions. By allowing users to define robust field constraints directly within schemas, such as Avro, we aim to enhance data quality and minimize the downstream impacts of inaccurate data in Kafka.
Furthermore, this feature can be expanded to include offline data processing, in addition to Kafka and Flink real-time processing. By combining real-time processing, batch analytics, and AI data pipelines, a global semantic validation system can be built.
In our upcoming talk, we will delve into the use cases of this feature, discuss its architecture, provide examples of defining rules, and explain how we enforce these rules. Ultimately, we will demonstrate how this feature can significantly enhance reliability and trustworthiness in Uber’s data processing pipelines."
Chill, Distill, No Overkill: Best Practices to Stress Test Kafka with Siva Ku...HostedbyConfluent
"So, you have built/inherited/discovered one of your many Kafka clusters. How now do you know that it is good enough to sustain and grow your applications? Do you stress test it as a data store, a messaging system, as middleware, or like a REST API? Or are you in production and worried about the next unprecedented surge? Find out from those who have asked and answered before.
Repeatable, and recordable stress testing for Kafka is a challenge for novices and some legends. Real supplies like storage, compute, network, threads etc. do not naturally map to demands of messages, bytes, and milliseconds. In the session, we will cover ways to:
* Define parameters and variables before beginning
* Accommodate for changing conditions - brokers, applications, config, network
* Overlap infrastructure, test design, latency, and throughput
* Meet cost, service level agreements, and multi-tenancy needs while testing
* Do it all without entirely relying on estimation, and extrapolation
We will also discuss common and innovative practices observed in the industry to meet this challenge. At the end of the session, you would walk away with the knowledge needed to set up a repeatable stress test suite without stress."
Video and slides synchronized, mp3 and slide download available at URL http://bit.ly/2qoUklo.
Mark Price talks about techniques for making performance testing a first-class citizen in a Continuous Delivery pipeline. He covers a number of war stories experienced by the team building one of the world's most advanced trading exchanges. Filmed at qconlondon.com.
Mark Price is a Senior Performance Engineer at Improbable.io, working on optimizing and scaling reality-scale simulations. Previously, he worked as Lead Performance Engineer at LMAX Exchange, where he helped to optimize the platform to become one of the world's fastest FX exchanges.
AWS re:Invent 2016: Beeswax: Building a Real-Time Streaming Data Platform on ...Amazon Web Services
Amazon Kinesis is a platform of services for building real-time, streaming data applications in the cloud. Customers can use Amazon Kinesis to collect, stream, and process real-time data such as website clickstreams, financial transactions, social media feeds, application logs, location-tracking events, and more. In this session, we first cover best practices for building an end-to-end streaming data applications using Amazon Kinesis. Next, Beeswax, which provides real-time Bidder as a Service for programmatic digital advertising, will talk about how they built a feature-rich, real-time streaming data solution on AWS using Amazon Kinesis, Amazon Redshift, Amazon S3, Amazon EMR, and Apache Spark. Beeswax will discuss key components of their solution including scalable data capture, messaging hub for archival, data warehousing, near real-time analytics, and real-time alerting.
Blockchain Testing Strategy - Testing is crucial in Blockchain as the technology ledger is immutable. The cost of a defect is very high in production. This paper explained what all changes faced in blockchain testing, and how can we resolve those challenges. What needs to be tested and testing approach. How performance testing can be done and what KPI's to be monitored.
Learn about the various approaches to sharding your data with MongoDB. This presentation will help you answer questions such as when to shard and how to choose a shard key.
Simple robot pets with three emotions (uC/OS III)YongraeJo
Implement a silly simple pet having switches, wheels and leds to represent his emotion, reacting to their environment. The system is implemented using uc/OS III (2015. 6)
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...Ramesh Iyer
In today's fast-changing business world, Companies that adapt and embrace new ideas often need help to keep up with the competition. However, fostering a culture of innovation takes much work. It takes vision, leadership and willingness to take risks in the right proportion. Sachin Dev Duggal, co-founder of Builder.ai, has perfected the art of this balance, creating a company culture where creativity and growth are nurtured at each stage.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on the notifications, alerts, and approval requests using Slack for Bonterra Impact Management. The solutions covered in this webinar can also be deployed for Microsoft Teams.
Interested in deploying notification automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
Key Trends Shaping the Future of Infrastructure.pdf
Pileus
1. Consistency-Based
Service Level Agreements
for Cloud Storage
Douglas B. Terry, Vijayan Prabhakaran, Ramakrishna Kotla, Mahesh Balakrishnan, Marcos K.
Aguilera, Hussam Abu-Libdeh†
Microsoft Research Silicon Valley †Cornell University
ACM SOSP 2013
Presented by Yongrae Jo, System software lab. at POSTECH
2. Motivation
2
Single ideal consistency
(weak ~ strong)
Multiple alternative consistency
(depending on app demand)
Cloud service provider
Guaranteed
by SLA
Guaranteed
by system design(DB)
Consistency vs Availability / Performance Single & Fixed consistency
to developer
Multiple & alternative
consistency to developer
3. Pileus
• Storage system for cloud
• Replicated key/value store with consistency-based SLAs
• Provides a broad set of consistency choices
• that lie between strong and eventual consistency
• avoiding single-ideal consistency
• Satisfies application-specific consistency /
latency demand
• Latency-favoring applications (e.g., shopping cart)
• Consistency-favoring applications (e.g., bank)
• Exports system APIs to application 3
4. Consistency Levels
Strong
consistency
Eventual
consistency
Latency-favoring applications
• Shopping app
• Real-time Multiplayer games
• Computer-supported collaborative work
• Data analytics
Consistency-favoring applications
• Bank
• Calendar
• web-based e-mail
Applications with Trade-offs
• Web-browser
• Display local cache first,
and then load accurate
data as it arrives
4
5. System API
Interface with traditional
key-value cloud storage
(e.g., Table op., Get, Put)
Consistency Guarantees
• Consistency choices
• Service level agreements
Default
sla
condition code
(Consistency + SLA met?)
5
6. System API
Interface with traditional
key-value cloud storage
(e.g., Table op., Get, Put)
Consistency Guarantees
• Consistency choices
• Service level agreements
Default
sla
condition code
(Consistency + SLA met?)
6
7. Consistency Guarantees
on Get(s, key, sla)
Types Return value of Get(key)
Strong Consistency
Return value of last preceding Put(key)
performed by any client
Casual Consistency
Return value of latest casually preceding
Put(key)
Bounded Staleness(t) Return value that is stale at most t seconds
Read My Writes
Return value of latest Put(key) in client session
or a later value
Monotonic Reads
Return same or later value as earlier Get(key)
in client session
Eventual Consistency
Return value written by any Put(key)
(but, expected to return the latest value later)
7
8. Example evaluation on different
consistency guarantees
8
one client in each
country
(Secondary) (Secondary) (N/A)
9. Service Level Agreement
• An ordered list of subSLAs
• subSLA
• A pair of Consistency / Latency target with application-specific
utility
Most preferable
Relative
importance
Less preferable
9
11. SLA Failure & Checking
• Satisfying SLA can fail due to
• configuration of replicas and network conditions
• poor decision based on inaccurate information
• Checking SLA Failure from the return value(i.e.,
condition code) of Get
• Application can take different actions based on the
consistency of the returned data.
Def. Unavailability of Pileus
The inability to retrieve the desired data with acceptable consistency
and latency as defined by the SLA
11
12. Design and Implementation
• Architecture
• Client-side
• Consistency-specific node selection
• Monitoring storage nodes
• Client-side SLA enforcement
• Choosing a target subSLA
• Determining which subSLA was met
12
14. Components
• Storage Node (or Secondary Node)
• Periodically fetches update results from primary node
• Primary Node
• One or some of storage node
• Holds master data(i.e., up-to-date data)
• Run replication protocol (e.g., consensus)
• (Client-side) Monitor
• Tracks the amount that storage nodes lag behind the primary node
• Measures roundtrip latencies between clients and storage nodes
• Client library
• Exports Pilleus APIs to application
14
15. Consistency-specific node selection
• How to select a node which a Get operation should be
sent?
• For desired consistency guarantee (Is it sufficiently up-to-date?)
• Minimum acceptable read timestamp
• Serves as a decision point between consistency guarantees
• Indicates an amount of lag of each node
15
Previous Object Versions
(in current session)
Key
(being read)
Minimum Acceptable
Read Timestamp
Consistency
guarantee
(Per) Node
high timestamp
Node Selection
16. Consistency-specific node selection
16
Types Minimum Acceptable Read Timestamp
Strong Consistency
At least as large as the update timestamp of
the latest Put to the key that is being Get
Casual Consistency
the maximum timestamp of any object that was
previously read or written in this session
(already casual-ordered by primary)
Bounded Staleness(t) Current time – bound time t
Read My Writes
Maximum timestamp of any previous Puts to
the key being accessed in current session
Monotonic Reads
The recorded timestamp for the key being
accessed in the Get of current session
Eventual Consistency 0
19. Monitoring Storage Nodes
• (client-side) Monitor probes latency / timestamp of
each node
• Monitor
• collects measurements in a sliding window (last few minutes)
• returns three probability estimates based on the recorded
information
PNodeCons (node, consistency, key)
Return probability that a node follows a
sufficiently up-to-date value
PNodeLat (node, latency)
Return probability that a node responds
within a given time
PNodeSla (node, consistency, latency,
key)
Return probability that a node satisfy
SLA (=PNodeCons * PNodeLat
19
20. Client-side SLA Enforcement
• How can we satisfy SLA effectively?
• Simple, but flawed method for SLA enforcement
• Broadcast Get op. to all replicas
• Incurs high cost (e.g., network resource, charging per byte)
• SLA Enforcement by Client Library
• Chooses a node group that can meet SLA
• Responsible for maximizing the expected utility
• Methods
• Choosing a target subSLA
• Determining which subSLA was met
20
21. // Choosing a target SLA and nodes
// subSLA that maximizes the expected utility
// Node group that clients will contact
21
22. Find targetSLA and node
group that best satisfies
SLA with maximum utility
// subSLA that maximizes the expected utility
// Node group that clients will contact
// Choosing a target SLA and nodes
22
23. // Node group that clients will contact
Find a node with
minimum latency
// subSLA that maximizes the expected utility
// Choosing a target SLA and nodes
23
24. Evaluation:
Experimental Setup
• Goal
• Evaluate Pileus in a globally distributed datacenter
environment
• Verify that adapting consistency is better than a fixed
consistency
• Measure how well the client’s Get operations meet a given consistency-
based SLA
• Evaluation
• Shopping cart SLA (weak consistency)
• Password checking SLA (strong consistency)
• Adaptability to network delays
• Sensitivity to utility values
24
25. Evaluation:
Experimental Setup
• YCSB benchmark with one client in each country
• Total 10,000 Put / Get, 400 Put / Get per session (20) by a client
• U.S West(Secondary), England(Primary), India(Secondary),
China(N/A)
• Comparisons with Pileus (different selection method)
• Primary
• always performs Gets at the primary node
• Random
• performs each Get at a randomly selected node
• Closest
• always performs Gets at the node with the lowest average latency
25
29. Evaluation:
Adaptability to network delays
• Injecting artificial delays into Get op in Password
checking SLA experiment
client (in the U.S.)
primary (in England)
Injecting 300ms delay
to primary
29
30. Evaluation:
Adaptability to network delays
• Injecting artificial delays into Get op in Password
checking SLA experiment
client (in the U.S.)
primary (in England)
Client -> (Primary, Rank1-SLA)
Clienit learns primary is far away,
switching to second subSLA with U.S node
; Client -> (U.S node, Rank2-SLA)
Client realizes that first subSLA is
cannot be met, switching to third one
; Client -> (Primary, Rank3-SLA)
Injecting 300ms delay
to primary
30
31. Evaluation:
Adaptability to network delays
• Injecting artificial delays into Get op in Password
checking SLA experiment
client (in the U.S.)
primary (in England)
Adding additional latency,
no SLA can be met
Client realized that only the third SLA
can be met, switching to primary
; Client -> (Primary, Rank3-SLA)
Injecting 300ms delay
to U.S. node(local node)
31
32. Evaluation:
Adaptability to network delays
• Injecting artificial delays into Get op in Password
checking SLA experiment
client (in the U.S.)
primary (in England)
Client discovers, through periodic
probes, that it can regularly access its
local site with low delay
switching back to local node
; Client -> (U.S. Node, Rank2-SLA)
Reducing delay(a millesecond)
in U.S. node(local node)
32
33. Evaluation:
Adaptability to network delays
• Injecting artificial delays into Get op in Password
checking SLA experiment
client (in the U.S.)
primary (in England)
Client figures primary is
normal,
switching back to primary
; Client ->
(Primary, Rank1-SLA)
Restoring the avg. latency to the
primary to the usual (149ms)
33
34. Extensions and future work
• Enhanced monitoring
• Sharing monitoring information between clients for accurate
decision
• (i.e., client-centric distributed monitoring service)
• SLA-driven reconfiguration
• Reconfiguring replicas according to SLA
• (e.g., moving primary replica nearby client)
• Parallel Gets
• Multi-site Puts
34
35. Conclusion
• Pilieus is a storage system with consistency-based SLA
• Consistency-based SLAs allow applications that were
written to tolerate eventual consistency to benefit
from increased consistency
• Adaptive to varying system condition
• (e.g., nodes fail, overloaded, performance variation)
• Avoiding single ideal consistency
• Pileus can improve application-specific consistency
levels of service
• application’s SLA indicates how best to adapt
35
36. Research Implications
• Structural similarities to Hyperledger/Fabric
• SLA ?= Endorsement policy
• Primary node ?= Ordering Service
• Storage node ?= Peer
• Simple monitoring & decision technique
• Defining a sliding window and prob. functions
• Collect metrics -> Calculate prob. -> Decision an action
• Quorum based SLA
• 2f+1 <= Q <= (2/3)n
• From low importance to high importance
36
37. Many form of consistency
Consistency models in distributed systems: A survey on definitions, disciplines, challenges and applications (2019)
37
38. Casual precedence relationship
op1 < op2 if either,
• (a) op1 occurs before op2 in the same session
• (b) op1 is a Put(key) and op2 is a Get(key) that returns the
version put in op1
• (c) for some op3, op1 < op3 and op3 < op2.
38