Key-value stores are widely used in applications that only require primary key data access, which is common in many web applications. Because developing an industrial grade key value store is expensive, the conventional solution is to use one of the existing key-value stores and layer application semantics on top of the primitives provided by the store. This approach leads to potential inefficiencies, because application specific semantics can often allow optimizations in the implementation of the store. We present an alternative approach, using the TACC platform to provide a key-value store implementation that is both performant and easily customizable. The TACC programming model separates state from logic: state is stored in a collection of distributed in-memory database instances, while logic is performed by distributed agents that react asynchronously to changes in objects stored in the database instances. Agents can selectively subscribe to updates using a fine-grain hierarchical directory system to mount objects into a local namespace. TACC provides performance comparable to hand-coded C while reducing the actual source code size to a fraction of that. We describe the implementation and performance of a scalable and fault tolerant key-value store using TACC, pointing out the benefits realized by using TACC's strong, user-defined types and triggering/notification.Key-value stores are widely used in applications that only require primary key data access, which is common in many web applications. Because developing an industrial grade key value store is expensive, the conventional solution is to use one of the existing key-value stores and layer application semantics on top of the primitives provided by the store. This approach leads to potential inefficiencies, because application specific semantics can often allow optimizations in the implementation of the store. We present an alternative approach, using the TACC platform to provide a key-value store implementation that is both performant and easily customizable. The TACC programming model separates state from logic: state is stored in a collection of distributed in-memory database instances, while logic is performed by distributed agents that react asynchronously to changes in objects stored in the database instances. Agents can selectively subscribe to updates using a fine-grain hierarchical directory system to mount objects into a local namespace. TACC provides performance comparable to hand-coded C while reducing the actual source code size to a fraction of that. We describe the implementation and performance of a scalable and fault tolerant key-value store using TACC, pointing out the benefits realized by using TACC's strong, user-defined types and triggering/notification.
http://sdec.kr/
Overcoming the Top Four Challenges to Real-Time Performance in Large-Scale, D...SL Corporation
The most critical large-scale applications today, regardless of industry, involve a demand for real-time data transfer and visualization of potentially large volumes of data. With this demand comes numerous challenges and limiting factors, especially if these applications are deployed in virtual or cloud environments. In this session, SL’s CEO, Tom Lubinski, explains how to overcome the top four challenges to real-time application performance: database performance, network data transfer bandwidth limitations, processor performance and lack of real-time predictability. Solutions discussed will include design of the proper data model for the application data, along with design patterns that facilitate optimal and minimal data transfer across networks.
Overcoming the Top Four Challenges to Real-Time Performance in Large-Scale, D...SL Corporation
The most critical large-scale applications today, regardless of industry, involve a demand for real-time data transfer and visualization of potentially large volumes of data. With this demand comes numerous challenges and limiting factors, especially if these applications are deployed in virtual or cloud environments. In this session, SL’s CEO, Tom Lubinski, explains how to overcome the top four challenges to real-time application performance: database performance, network data transfer bandwidth limitations, processor performance and lack of real-time predictability. Solutions discussed will include design of the proper data model for the application data, along with design patterns that facilitate optimal and minimal data transfer across networks.
Jos van den Oord, senior Oracle Database Specialist bij Transfer Solutions laat zien hoe je kunt ‘spelen’ met Oracle clusterware infrastructuur componenten om de hoogbeschikbaarheid binnen een RAC omgeving te verbeteren.
Presentation at FreedomHEC 2012 Conference. 0xlab extends DMTCP (Distributed Multi-Threaded CheckPointing) to enable Android checkpointing, which leads to resume to stored state for faster Android boot time and make better product field trial experience.
this is the slides from the talk i gave at DevGeekWeek2014
further details are in my blog: http://blogs.microsoft.co.il/iblogger/2014/06/25/devgeekweek-2014-slides-and-demos/
[NetPonto] Arquitectura dos Serviços da plataforma Windows AzureVitor Tomaz
O Windows Azure é uma plataforma que fornece serviços de alta disponibilidade e escalabilidade. Nesta sessão iremos abordar a arquitectura dos serviços base desta plataforma (Compute, Storage e SQLAzure) de modo a entendermos de que forma é que a escalabilidade e alta disponibilidade são conseguidas. Iremos ver as diferenças para as plataformas "tradicionais" e algumas consequências no desenvolvimento de soluções para este ambiente.
Apresentação do Vitor Tomaz sobre a Arquitectura dos Serviços da plataforma Windows Azure na 4a Reunião Presencial da Comunidade NetPonto em Coimbra (http://netponto.org).
Presentation given to the #lspe meetup (Large Systems Performance Engineering) on February 21, 2013 by Steve Shah. Topic for the night was Dynamic Scaling. This presentation is titled "Shock Absorbers and APIs" and covers features typical of ADCs (modern load balancers) that can help in managing scale as well as give a quick overview of what to expect from an API in an ADC.
Session presented at the 2nd IndicThreads.com Conference on Cloud Computing held in Pune, India on 3-4 June 2011.
http://CloudComputing.IndicThreads.com
Abstract:“With increasing demand, ever-growing datasets, unpredictable traffic patterns and need for faster response times, “scalable architecture” has become a necessity. Here, we will see how the traditional concepts and best practices for scalability have to be adopted for the cloud. Further, we will go through the unique advantages that Amazon AWS cloud offers for architecting scalable applications. As an architect, you need to identify the components and bottlenecks in your architecture and modify your application to leverage the underlying scalability.
We will cover the following topics:
Scalability principles for the cloud
Leveraging AWS services for application components
Shared nothing architecture
Asynchronous work queues for loosely coupled applications
Database scalability
Tools, connectors and enablers to help build, deploy and monitor your cloud environment
Scalability using Platform-as-a-Service offerings on top of AWS
An example of a horizontally scalable architecture for an enterprise application on Amazon AWS
This talk will act as a primer for a cloud architect to achieve an auto-scalable, highly available, fully-monitored edge-cached application.”
Speaker:
Kalpak Shah is the Founder & CEO of Clogeny Technologies Pvt. Ltd. and guides the overall strategic direction of the company. Clogeny is focused on niche software and product development in cloud computing and scalable applications domains. He is passionate about the ground-breaking economics and technology afforded by the cloud computing platforms. He has been leading and architecting cutting-edge product development across the cloud stack including IaaS, PaaS and SaaS vendors.
He has previously worked at organizations like Sun Microsystems and Symantec in the storage domain primarily distributed and disk filesystems. Kalpak has a Bachelors’ of Engineering degree in computer engineering from PICT, University of Pune.
PPTV is using CloudStack 3.0.2 in its production environment. Currently there are more than 150 hosts, and migrate their apps to cloud everyday (10 host per day). At the end of 2013, there will be more than 1000 hosts in a CloudStack environment.
Jos van den Oord, senior Oracle Database Specialist bij Transfer Solutions laat zien hoe je kunt ‘spelen’ met Oracle clusterware infrastructuur componenten om de hoogbeschikbaarheid binnen een RAC omgeving te verbeteren.
Presentation at FreedomHEC 2012 Conference. 0xlab extends DMTCP (Distributed Multi-Threaded CheckPointing) to enable Android checkpointing, which leads to resume to stored state for faster Android boot time and make better product field trial experience.
this is the slides from the talk i gave at DevGeekWeek2014
further details are in my blog: http://blogs.microsoft.co.il/iblogger/2014/06/25/devgeekweek-2014-slides-and-demos/
[NetPonto] Arquitectura dos Serviços da plataforma Windows AzureVitor Tomaz
O Windows Azure é uma plataforma que fornece serviços de alta disponibilidade e escalabilidade. Nesta sessão iremos abordar a arquitectura dos serviços base desta plataforma (Compute, Storage e SQLAzure) de modo a entendermos de que forma é que a escalabilidade e alta disponibilidade são conseguidas. Iremos ver as diferenças para as plataformas "tradicionais" e algumas consequências no desenvolvimento de soluções para este ambiente.
Apresentação do Vitor Tomaz sobre a Arquitectura dos Serviços da plataforma Windows Azure na 4a Reunião Presencial da Comunidade NetPonto em Coimbra (http://netponto.org).
Presentation given to the #lspe meetup (Large Systems Performance Engineering) on February 21, 2013 by Steve Shah. Topic for the night was Dynamic Scaling. This presentation is titled "Shock Absorbers and APIs" and covers features typical of ADCs (modern load balancers) that can help in managing scale as well as give a quick overview of what to expect from an API in an ADC.
Session presented at the 2nd IndicThreads.com Conference on Cloud Computing held in Pune, India on 3-4 June 2011.
http://CloudComputing.IndicThreads.com
Abstract:“With increasing demand, ever-growing datasets, unpredictable traffic patterns and need for faster response times, “scalable architecture” has become a necessity. Here, we will see how the traditional concepts and best practices for scalability have to be adopted for the cloud. Further, we will go through the unique advantages that Amazon AWS cloud offers for architecting scalable applications. As an architect, you need to identify the components and bottlenecks in your architecture and modify your application to leverage the underlying scalability.
We will cover the following topics:
Scalability principles for the cloud
Leveraging AWS services for application components
Shared nothing architecture
Asynchronous work queues for loosely coupled applications
Database scalability
Tools, connectors and enablers to help build, deploy and monitor your cloud environment
Scalability using Platform-as-a-Service offerings on top of AWS
An example of a horizontally scalable architecture for an enterprise application on Amazon AWS
This talk will act as a primer for a cloud architect to achieve an auto-scalable, highly available, fully-monitored edge-cached application.”
Speaker:
Kalpak Shah is the Founder & CEO of Clogeny Technologies Pvt. Ltd. and guides the overall strategic direction of the company. Clogeny is focused on niche software and product development in cloud computing and scalable applications domains. He is passionate about the ground-breaking economics and technology afforded by the cloud computing platforms. He has been leading and architecting cutting-edge product development across the cloud stack including IaaS, PaaS and SaaS vendors.
He has previously worked at organizations like Sun Microsystems and Symantec in the storage domain primarily distributed and disk filesystems. Kalpak has a Bachelors’ of Engineering degree in computer engineering from PICT, University of Pune.
PPTV is using CloudStack 3.0.2 in its production environment. Currently there are more than 150 hosts, and migrate their apps to cloud everyday (10 host per day). At the end of 2013, there will be more than 1000 hosts in a CloudStack environment.
StreamBase - Embedded Erjang - Erlang User Group London - 20th April 2011darach
A presentation delivered to the Erlang User Group in London demonstrating how to embed the erjang implementation of erlang into the StreamBase CEP engine, enabling extending StreamBase with erlang based extensions.
With AWS you can choose the right database for the right job. Given the myriad of choices, from relational databases to non-relational stores, this session will profile details and examples of some of the choices available to you (MySQL, RDS, Elasticache, Redis, Cassandra, MongoDB and DynamoDB), with details on real world deployments from customers using Amazon RDS, ElastiCache and DynamoDB.
The Good, the Bad and the Ugly of Migrating Hundreds of Legacy Applications ...Josef Adersberger
Running applications on Kubernetes can provide a lot of benefits: more dev speed, lower ops costs, and a higher elasticity & resiliency in production. Kubernetes is the place to be for cloud native apps. But what to do if you’ve no shiny new cloud native apps but a whole bunch of JEE legacy systems? No chance to leverage the advantages of Kubernetes? Yes you can!
We’re facing the challenge of migrating hundreds of JEE legacy applications of a major German insurance company onto a Kubernetes cluster within one year. We're now close to the finish line and it worked pretty well so far.
The talk will be about the lessons we've learned - the best practices and pitfalls we've discovered along our way. We'll provide our answers to life, the universe and a cloud native journey like:
- What technical constraints of Kubernetes can be obstacles for applications and how to tackle these?
- How to architect a landscape of hundreds of containerized applications with their surrounding infrastructure like DBs MQs and IAM and heavy requirements on security?
- How to industrialize and govern the migration process?
- How to leverage the possibilities of a cloud native platform like Kubernetes without challenging the tight timeline?
Migrating Hundreds of Legacy Applications to Kubernetes - The Good, the Bad, ...QAware GmbH
CloudNativeCon North America 2017, Austin (Texas, USA): Talk by Josef Adersberger (@adersberger, CTO at QAware)
Abstract:
Running applications on Kubernetes can provide a lot of benefits: more dev speed, lower ops costs, and a higher elasticity & resiliency in production. Kubernetes is the place to be for cloud native apps. But what to do if you’ve no shiny new cloud native apps but a whole bunch of JEE legacy systems? No chance to leverage the advantages of Kubernetes? Yes you can!
We’re facing the challenge of migrating hundreds of JEE legacy applications of a major German insurance company onto a Kubernetes cluster within one year. We're now close to the finish line and it worked pretty well so far.
The talk will be about the lessons we've learned - the best practices and pitfalls we've discovered along our way. We'll provide our answers to life, the universe and a cloud native journey like:
- What technical constraints of Kubernetes can be obstacles for applications and how to tackle these?
- How to architect a landscape of hundreds of containerized applications with their surrounding infrastructure like DBs MQs and IAM and heavy requirements on security?
- How to industrialize and govern the migration process?
- How to leverage the possibilities of a cloud native platform like Kubernetes without challenging the tight timeline?
게임을 위한 Cloud Native on AWS
IT 기술이 변화하며 클라우드를 보다 적극적으로 사용하는 게임사가 늘어나는 추세입니다. 게임 고객분들이 다양한 시각에서 AWS Cloud Service를 보다 효과적으로 잘 사용할 수 있는 방법을 소개합니다. 또한, 고객분들께서 개발에 집중하고 효율적으로 운영할 수 있도록 AWS가 어떠한 도움을 드리는지에 대해 말씀드리고자 합니다.
Towards an Architectural Style for Multi-tenant Software ApplicationsHeiko Koziolek
Multi-tenant software applications serve different organizations from a single instance and help to save development, maintenance, and administration costs. The architectural concepts of these applications and their relation to emerging platform-as-a-service (PaaS) environments are still not well understood, so that it is hard for many developers to design and implement such an application. Existing attempts at a structured documentation of the underlying concepts are either technology-specific or restricted to certain details. We propose documenting the concepts as a new architectural style. This paper initially describes the architectural properties, elements, views, and constraints of this style. We illustrate how the architectural elements are implemented in current PaaS environments, such as Force.com, Windows Azure, and Google App Engine.
In the SNS domain, Response time of Friend Suggestion Algorithms and several SNA algorithms is in direct proportion to square of relationship number. In addition, increasing rate of relationship number is bigger and bigger. But existing usage pattern of Relational DB is suffering poor performance. To guarantee performance and scalability, we have developed such methods for Friend Suggestion and SNA
Relation Pruning using intimacy value
No Join & Keeping all Data in-Memory Strategy
Distributed Graph Structure
SDEC2011 Replacing legacy Telco DB/DW to Hadoop and HiveKorea Sdec
Currently telecom companies store their data in database or data warehouse, treating them through ETL process and working on statistics and analysis by using OLAP tools or data mining engines. However, due to the data explosion along with the spread of Smart Phones traditional data storages like DB and DW aren’t sufficient to cope with these “Big Data”. As an alternative the method of storing data in Hadoop and performing ETL process and Ad-hoc Query with Hive is being introduced, and China Mobile is being mentioned as the most representative example. But, they are adopted mainly by new projects, which have low barriers in applying the new Hive data model and HQL. On the other hand, it is extremely difficult to replace the existing database with the combination of Hadoop and Hive if there are already a number of tables and SQL queries. NexR is migrating the telecom company’s data from Oracle DB to Hadoop, and converting a lot of existing Oracle SQL queries to Hive HQL queries. Though HQL supports a similar syntax to ANSI-SQL, it lacks a large portion of basic functions and hardly supports Oracle analytic functions like rank() which are utilized mainly in statistical analysis. Furthermore, the difference of data types like null value is also blocking the application of it. In this presentation, we will share the experience converting Oracle SQL to Hive HQL and developing additional functions with MapReduce. Also, we will introduce several ideas and trials to improve Hive performance.
http://sdec.kr/
SDEC2011 Mahout - the what, the how and the whyKorea Sdec
Mahout is an open source machine learning library from Apache. From its humble beginnings at Apache Lucene, the project has grown into a active community of developers, machine learning experts and enthusiasts. With v0.5 released recently, the project has been focussing full steam on developing stable APIs with an eye on our major milestone of v1.0. The speaker has been with Mahout from his days in college as a computer science student. The talk will focus on the major use cases of Mahout. The design decisions, things that worked, things that didn't, and things to expect in the future releases.
http://sdec.kr/
SDEC2011 Using Couchbase for social game scaling and speedKorea Sdec
A social game, by it's very nature, can spread very quickly to a large user population. Because the game is typically interactive, the speed of retrieving information needed for the user's interactions with the system is critical. When building their new game Animal Party, the developers at Tribal Crossing needed to get away from the complexity of sharding an SQL database. They also were looking for a solution to the administration cost associated with the operation of traditional data stores. When evaluating multiple different NoSQL solutions, they realized that Couchbase's Membase server meets most of their critical requirements in developing their game software. Simple to use, Couchbase's model allows Tribal Crossing to easily model their game interactions with the key/value data store. Fast read and write performance is required with interactive, social games, and they found that support in Membase as well. Elastic scalability is easily achieved by simply adding more nodes to the Couchbase cluster without any modifications required to the application. Relying on Couchbase's technology Tribal Crossing has been able to quickly build and scale Animal Party with a small team and no dedicated system administrators.
http://sdec.kr/
Arcus is data caching cloud developed on Memcached which is a memory-based Key-Value repository. It was developed meeting requirements from various NHN services. Arcus is capable of having up-to-date cache server list of clients using ZooKeeper and also able to possess cloud architecture flexible to add or delete caching servers while eliminating extra network latency between client and server. Furthermore, to support save/ computation function of data collection which is becoming a requirement among many services, it provides list, set, b+tree structures.
http://sdec.kr/
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Securing your Kubernetes cluster_ a step-by-step guide to success !
SDEC2011 Going by TACC
1. Going
by
TACC:
Beyond
Key-‐Value
to
Fault-‐Tolerant
Stores
with
Easily
Customizable
Semantics
Henk
Goosen,
CEO
goosen@optumsoft.com
2. Key-‐value
stores
rule
the
Web
Many
applications
only
need
primary
key
data
access
Examples:
catalogs,
shopping
carts,
web
session
state
No
need
for
the
complexity,
performance
overhead,
and
lack
of
scalability
of
a
full
database
Hence:
Key-‐value
stores
are
everywhere
Dynamo,
CouchDB,
Cassandra,
Project
Voldemort,
Riak,
Redis,
memcached,
MongoDB,
…
Improving key-value stores is important
OptumSoft, Inc. Proprietary and 2
Confidential
3. Key-‐value
stores
in
practice
Developing
a
key-‐value
store
from
scratch
using
conventional
languages
is
expensive:
scalability,
performance,
and
fault
tolerance
Conventional
solution:
use
existing
key-‐value
store
Layer
on
get()
and
put()
semantics
Mismatches
between
application
requirements
and
library:
either
accept
or
extensively
modify
library
code
Applications are more complex,
performance suffers
OptumSoft, Inc. Proprietary and 3
Confidential
4. TACC
provides
a
different
model
Use
a
very
high-‐level
language
to
specify
the
key-‐value
store
Then
customize
the
store,
applying
application-‐specific
semantics
Benefits:
Simplifies
the
application
business
logic
Improves
the
performance
of
both
store
and
application
TACC model is better!
OptumSoft, Inc. Proprietary and 4
Confidential
5. TACC
is
an
object-‐oriented,
strongly
typed
language
User-‐defined
type:
a
list
of
attributes
(nouns)
Read
or
write
attributes
(there
are
no
methods/verbs)
Logic
primarily
implemented
via
constraints
imperative
code
is
also
supported
Compact
code
First
class
high
level
data
types
(eg,
queues,
hash
tables)
Several
design
patterns
directly
supported
in
language
(eg
observer
pattern)
Compact code fewer bugs, quicker to market
5
6. TACC:
efficient
development
of
distributed
systems
Reduce
development
time
by
a
factor
of
2x
to
3x
Reduce
lines
of
code
by
10x
or
more
Eliminate
most
synchronization
and
concurrency
bugs
High,
predictable
performance
using
optimized
code
generation
Fault-‐Tolerance
built
into
the
model,
and
easy
to
implement
TACC is a general purpose language,
focused on distributed systems
6
7. Stateful
remote
proxy
objects
LR
1
LR
2
Agents 1
Proxy:
local
copy
of
data
1
Writes
are
asynchronously
object added copied
to
SysDB
to collection
SysDB
changes
are
copied
to
“interested”
agents
R/W
access
is
local,
fast
SysDB 1
No
remote
access
collection
exceptions
Simple semantics, and fast
OptumSoft, Inc. Proprietary and 7
Confidential
8. SysDB:
a
hierarchical
in-‐memory
object
database
Stores
state
(ideally
no
logic)
Minimizes
risk
of
program
logic
bugs,
hence
reliable
Agents
Concise
specification
of
user-‐
defined
types
TACC
compiler
automatically
generates
all
required
code
for
remote
access
SysDB
Agents
receive
automatic
notification
when
values
change
OptumSoft, Inc. Proprietary and 8
Confidential
9. Distributed,
hierarchical
name
space
SysDB
defines
and
exports
an
hierarchical
name
space
(similar
to
a
distributed
file
system)
Remote
agents
can
“mount”
remote
directories
into
a
local
namespace
Each
object
is
instantiated
into
a
directory,
state
is
made
available
remotely
via
proxy
objects
Updates
propagate
asynchronously,
notifications
are
delivered
on
changes
Simple, powerful, proven way to
provide large, structured name space
OptumSoft, Inc. Proprietary and 9
Confidential
10. Fault-‐tolerance
is
built
in
When
an
agent
restarts,
it
recovers
its
state
from
SysDB
A1
A2
A3
A4
Agents
implement
invariants,
therefore
can
be
restarted
at
any
time,
on
any
server
Any
number
of
backup
SP
SB
SysDBs
are
supported
Fast
recovery
for
high
availability
10
11. Example:
Location
Service
as
customized
key-‐value
store
Application
needs
to
track
real-‐time
location
of
user
User
allowed
in
only
one
location
at
a
time
Three
operations:
ENTER
<user
id>
<session
id>
<location
id>
LEAVE
<user
id>
QUERY
<user
id>
Throughput
>
10,000
requests/sec,
latency
<
1
ms
High throughput, low latency required
OptumSoft, Inc. Proprietary and 11
Confidential
12. Location
Service
Overview
Load
balancer
Get
GS
location
LR
HTTP
access
to
service
GS
Application
(GS)
contacts
Leave
any
LR
server
via
load
LR
balancer
GS
LR
servers
replicated
for
GS
LR
scalability
and
for
fault
Enter
tolerance
GS
LR
Challenge: ensure responses from
GS
Enter
multiple LR servers are handled
correctly
OptumSoft, Inc. Proprietary and 12
Confidential
13. Key-‐value
store
tracks
location
for
each
user
Load
Key-‐value
balancer
store
GS
LR
Shard
GS
A-‐J
LR
Smith,1
GS
Has
to
be
Enter
atomic
Shard
Smith,1
K-‐R
GS
get(),
LR
put()
GS
Shard
Enter
LR
Smith,2
get(),
Smith
S-‐Z
GS
Smith,2
put()
OptumSoft, Inc. Proprietary and 13
Confidential
14. TACC
allows
easy
customization
of
key-‐value
update
semantics
Each
partition
stores
a
unique
subset
of
the
user
state
We
directly
implement
ENTER,
LEAVE,
and
QUERY
semantics,
using
a
TACC
Constrainer
No
locking
or
inter-‐agent
synchronization
required
Requests
and
responses
sent
asynchronously
High
performance:
there
is
no
waiting
or
blocking
Specializing the key-value store semantics
simplifies the application and improves performance
OptumSoft, Inc. Proprietary and 14
Confidential
15. Single-‐writer
collections:
no
need
for
synchronization
R
S
R
LR
R S
S
R Shard
R Request
Collection
A-‐J
S
Response
Collection
S
R R
S
S
LR
R
S
R
S
R R Shard
S
S
K-‐R
LR
R R
S
S
OptumSoft, Inc. Proprietary and 15
Confidential
16. The
Serializer
Constrainer
Logic
Notify
Update user
Write result status
Request
Collection
Response
Collection
A
Enter
U1,
R5
A
OK
Status
Collection
K
Enter
U1,
R5
K
NOT
ALLOWED
U1
R5
D
Enter
U8,
R9
D
OK
U8
R9
Really simple!
OptumSoft, Inc. Proprietary and 16
Confidential
17. Details
of
Constrainer
implementation
Code
for
the
Serializer
constrainer
defines
three
collections:
Input
collection:
requests
Output
collections:
responses
and
user
status
A
dependency
constraint
causes
imperative
code
to
be
executed
when
a
new
request
arrives
from
LR
server
The
imperative
code
in
the
constrainer
implements
the
application
specific
semantics
This code is a minor tweak on put() implementation
OptumSoft, Inc. Proprietary and 17
Confidential
18. Constraints,
strong
typing
improves
event
handling
code
Constraint
handling
code
automatically
inserted
by
compiler
No
need
to
manually
maintain
invariants
in
many
call
sites
User-‐defined
types
organize
constraint
handling
code
and
protect
against
mistakes
TACC
coroutine
further
simplifies
event
handling
TACC changes event-handling spaghetti into
well-structured, type-safe code
OptumSoft, Inc. Proprietary and 18
Confidential
19. Instrumentation
and
Measurements
Stress
Agent
and
SysDB
instrumented
to
collect
timestamps
(stored
in
memory,
I/O
after
test)
tcpdump
run
on
Stress
Agent
and
SysDB
servers
Correlate
timestamps
with
tcpdump
OptumSoft, Inc. Proprietary and 19
Confidential
20. Low
latency
pitfalls
to
avoid
Network
and
TCP
behavior
Many
TCP
settings
have
a
dramatic
and
non-‐linear
performance
impact
Memory
management
Memory
allocation/deallocation
Avoid
garbage
collection
“The devil is in the details”
OptumSoft, Inc. Proprietary and 20
Confidential
21. Zero-‐load
Latency
(μs)
End-‐to-‐end
Time
Latency
Request
0
created
1
Request
48
48
SysDB
Time
Latency
packet
2
Receive
request
3
0.0
Response
248
200
Notification
4
42.3
42.3
packet
7
Response
75.1
32.8
Notification
8
288
40
enqueued
5
Response
packet
6
108.5
33.4
Latencies are low and predictable
OptumSoft, Inc. Proprietary and 21
Confidential
22. Latency,
throughput
vs
SysDBs
High scalability under Latency converges to
strict latency bound zero-load latency
OptumSoft, Inc. Proprietary and 22
Confidential
23. Summary
Tacc
enables
developers
to
efficiently
create
predictably
high
performance,
scalable,
fault-‐tolerant
distributed
applications
Eliminates
synchronization
and
locking
bugs
Fewer
lines
of
code
Faster
to
develop,
shorter
time
to
market
Easier
to
maintain
Fewer
bugs
23
24. Contact
me
for
more
information
about
TACC
and
OptumSoft!
goosen@optumsoft.com
OptumSoft, Inc. Proprietary and 24
Confidential