The document discusses key concepts for designing IT infrastructure to ensure high performance. It covers perceived performance from a user perspective, benchmarking systems, profiling users to predict load, identifying and managing bottlenecks, scaling systems horizontally and vertically, load balancing, caching frequently used data, and designing systems based on their intended use to optimize performance. The overall goal is to design infrastructure that can meet performance requirements under all conditions, both currently and as load increases over time.
INTRODUCTIONTO OPERATING SYSTEM
What is an Operating System?
Mainframe Systems
Desktop Systems
Multiprocessor Systems
Distributed Systems
Clustered System
Real -Time Systems
Handheld Systems
Computing Environments
this slides explains basics of system administration by raising these question, what is system administration? who can become syss admin? how to become sys admin? and discusses about tools used by sys admins.
INTRODUCTIONTO OPERATING SYSTEM
What is an Operating System?
Mainframe Systems
Desktop Systems
Multiprocessor Systems
Distributed Systems
Clustered System
Real -Time Systems
Handheld Systems
Computing Environments
this slides explains basics of system administration by raising these question, what is system administration? who can become syss admin? how to become sys admin? and discusses about tools used by sys admins.
Grails has great performance characteristics but as with all full stack frameworks, attention must be paid to optimize performance. In this talk Lari will discuss common missteps that can easily be avoided and share tips and tricks which help profile and tune Grails applications.
Performance is a critical factor for success of any packaged application implementations. The presentation discusses performance assurance for packaged applications on example of Oracle Enterprise Performance Management. While details are related to this particular set of applications, many approaches discussed would be applicable to most packaged applications. The presentation will discuss a holistic performance assurance approach, top-down approach to performance troubleshooting, potential performance issues and ways to address them.
University of Alberta migrated their central Learning Management System from Blackboard Vista on Oracle to Moodle on Postgresql 9.0. We went from a pilot project of 13 courses in January 2011 to running all centrally supported courses (3600+) in Moodle in September 2012. Our central Moodle instance has seen more than 500,000 page loads and 24,000 unique visitors in a single day. Over the last two years we have learned a few hard lessons and overcome a few challenges in running Postgresql in a 24x7 production environment.
Adding Value in the Cloud with Performance TestRodolfo Kohn
System quality attributes such performance, scalability, and availability are among the main concerns for cloud application developers and product managers. There are many examples of notable system failures that show how a company business can be affected during key events like a Cyber Monday. However, many difficulties come up when a team intends to consciously manage these type of quality attributes during development and operations. It is possible to group these difficulties in two main aspects: human aspects and technical aspects. During this presentation, I will share main technical difficulties we had to deal with in the last seven years working with different cloud services as well as key technical performance, scalability, and availability issues we were able to find and solve. It is about cases that are relevant through different products, technologies, and teams.
MongoDB Days Silicon Valley: Best Practices for Upgrading to MongoDBMongoDB
Presented by Achille Brighton, Principal Consulting Engineer, MongoDB
Experience level: Deep dive
MongoDB 3.2 brings major enhancements. New pluggable storage engines optimized for in-memory computing and the most security-sensitive applications. Simplified data governance with document validation, coupled with GUI-based schema discovery and visualization. Improved operational efficiency with enhanced management platforms, continuous uptime across distributed, multi-region deployments, and zero-downtime upgrades. To take advantage of these features, your team needs an upgrade plan. In this session, we’ll walk you through how to build an upgrade plan. We’ll show you how to validate your existing deployment, build a test environment with a representative workload, and detail how to carry out the upgrade. You’ll walk away confident that you're prepared to upgrade.
MongoDB 3.2 introduces a host of new features and benefits, including encryption at rest, document validation, MongoDB Compass, numerous improvements to queries and the aggregation framework, and more. To take advantage of these features, your team needs an upgrade plan.
In this session, we’ll walk you through how to build an upgrade plan. We’ll show you how to validate your existing deployment, build a test environment with a representative workload, and detail how to carry out the upgrade. By the end, you should be prepared to start developing an upgrade plan for your deployment.
Similar to 05. performance-concepts-26-slides (20)
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on the notifications, alerts, and approval requests using Slack for Bonterra Impact Management. The solutions covered in this webinar can also be deployed for Microsoft Teams.
Interested in deploying notification automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Generating a custom Ruby SDK for your web service or Rails API using Smithyg2nightmarescribd
Have you ever wanted a Ruby client API to communicate with your web service? Smithy is a protocol-agnostic language for defining services and SDKs. Smithy Ruby is an implementation of Smithy that generates a Ruby SDK using a Smithy model. In this talk, we will explore Smithy and Smithy Ruby to learn how to generate custom feature-rich SDKs that can communicate with any web service, such as a Rails JSON API.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
2. Introduction
• Performance is a
typical hygiene
factor.
• Nobody notices a
highly performing
system.
• But when a system is
not performing well
enough, users quickly
start complaining.
3. Perceived performance
• Perceived performance refers to how quickly a
system appears to perform its task.
• In general, people tend to overestimate their
own patience.
• People tend to value predictability in
performance
– When the performance of a system is fluctuating,
they remember a bad experience, even if the
fluctuation is relatively rare.
6. Performance during infrastructure
design
• Designing for performance ensures that a solution is
designed, implemented, and supported to meet the
performance requirements, even under increasing load.
• When designing a system, performance must be considered
not only when the system works as expected, but also
when the system is in a special state.
– Failing parts
– Maintenance state
– Performing backup
– Running batch jobs
• Calculating performance of a system in the design phase is
extremely difficult and very unreliable.
7. Benchmarking
• A benchmark uses a specific test program to assess the
relative performance of an infrastructure component
• Benchmarks provide a method of comparing the
performance of various subsystems across different
system architectures.
• Often used for computer hardware
– Floating Point Operations Per Second – FLOPS
– Million Instructions Per Second – MIPS
• Only useful for comparing the raw speed of parts of an
infrastructure
– Like the speed difference between processors or between
disk drives
8. Vendor experience
• The best way to determine the performance
of a system in the design phase is to use the
experience of vendors
• They have a lot of experience running their
products in various infrastructure
configurations
9. Prototyping
• Also known as proof of concept (PoC)
• To measure the performance of a system at an
early stage
– Hiring equipment from suppliers
– Using datacenter capacity at a vendor’s premise
– Using cloud computing resources
• Focus on those parts of the system that pose
the highest risk, early in the design process
10. User profiling
• Predict the load a new software system will
pose on the infrastructure before the software
is actually built
• It is important to have a good indication of the
expected usage of the system
– Defining a number of typical user groups of the
new system (personas)
– Creating a list of tasks personas will perform on
the new system.
11. User profiling personas/tasks
Persona Number
of users
per
persona
System task Infrastructure load
as a result of the
system task
Frequency
Data
entry
officer
100 Start
application
Read 100 MB data
from SAN
Once a day
Data
entry
officer
100 Start
application
Transport 100 MB
data to workstation
Once a day
Data
entry
officer
100 Enter new
data
Transport 50 KB data
from workstation to
server
40 per
hour
Data
entry
officer
100 Enter new
data
Store 50 KB data to
SAN
40 per
hour
Data
entry
officer
100 Change
existing data
Read 50 KB data
from SAN
10 per
hour
12. User profiling Infrastructure load
Infrastructure load Per day
Per
second
Data transport from server to workstation (KB) 10,400,000 361.1
Data transport from workstation to server (KB) 2,050,000 71.2
Data read from SAN (KB) 10,400,000 361.1
Data written to SAN (KB) 2,050,000 71.2
14. Managing bottlenecks
• The performance of a system is based on the
performance of all its components, and the
interoperability of various components
• Every system, regardless of how well it works, has at
least one constraint (a bottleneck) that limits its
performance (Bottleneck law)
• A component causing the system to reach some limit is
referred to as the bottleneck of the system
• If the bottleneck does not negatively influence
performance of the complete system under the highest
expected load, it is OK
15. Performance testing
• Load testing - This test shows how a system
performs under the expected load
• Stress testing - This test shows how a system
reacts when it is under extreme load
• Endurance testing - This test shows how a
system behaves when it is used at the
expected load for a long period of time
18. Increasing performance on upper
layers
• 80% of the performance issues are due to badly
behaving applications
• Database and application tuning typically
provides much more opportunity for
performance increase than installing more
computing power
• Application performance can benefit from:
– Prioritizing tasks
– Working from memory as much as possible (as
opposed to working with data on disk)
– Making good use of queues and schedulers
19. Caching
• Caching improves performance by retaining frequently used
data in high speed memory, reducing access times to data.
– Disk caching
– Web proxies
– Operational Data Store
– Front-end servers
– In-memory databases
Component
Time it takes to fetch 1 MB of
data (ms)
Network, 1 Gbit/s 675
Hard disk, 15k rpm, 4 KB disk blocks 105
Main memory DDR3 RAM 0.2
CPU L1 cache 0.016
20. Scalability
• Scalability indicates the ease in with which a system
can be modified, or components can be added, to
handle increasing load
• Two ways to increase the scalability of a system:
– Vertical scaling (scale up) - adding resources to a single
component
– Horizontal scaling (scale out) - adding more components to
the infrastructure
22. Load balancing
• Load balancing spreads the load over various machines
• It checks the current load on each server in the farm and
sends incoming requests to the least busy server.
23. High performance clusters
• High performance clusters provide a vast amount of
computing power by combining many computer systems.
• Usually a large number of cheap off the-shelf servers are used
• A combination of relatively small computers can create one
large supercomputer
• Used for calculation-intensive systems
– Weather forecasts
– Geological research
– Nuclear research
– Pharmaceutical research
• TOP500.ORG
24. Grid Computing
• A computer grid is a high performance cluster that consists of
systems that are spread geographically
• The limited bandwidth is the bottleneck
• Examples:
– SETI@HOME
– CERN LHC Computing Grid (140 computing centers in 35 countries)
• Broker firms exist for commercial exploitation of grids
• Security is a concern!
25. Design for use
• Performance critical applications should be designed as such
• Tips:
– Know what the system will be used for. A large data warehouse needs
a different infrastructure design than an online transaction processing
system or a web application
– In some cases, special products must be used for certain systems (real-
time operating systems, in-memory databases, specially designed file
systems)
– Use standard implementation plans that are proven in practice
– Have the vendors check the design you created.
– When possible, try to spread the load of the system over the available
time
– Move rarely used data from the main systems to other systems
26. Capacity management
• Capacity management guarantees high performance of a
system in the long term
• Performance of the system is monitored on a continuous
base, to ensure performance stays within acceptable limits
• Trend analyses can be used to predict performance
degradation
• Anticipate on business changes (like forthcoming marketing
campaigns)