#Interactive Session by Vivek Patle and Jahnavi Umarji, "Empowering Functional Testing with Support Vector Machines: An Experimental Journey" at #ATAGTR2023.
#Interactive Session by Vivek Patle and Jahnavi Umarji, "Empowering Functional Testing with Support Vector Machines: An Experimental Journey" at #ATAGTR2023.
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
MLOps and Data Quality: Deploying Reliable ML Models in ProductionProvectus
Looking to build a robust machine learning infrastructure to streamline MLOps? Learn from Provectus experts how to ensure the success of your MLOps initiative by implementing Data QA components in your ML infrastructure.
For most organizations, the development of multiple machine learning models, their deployment and maintenance in production are relatively new tasks. Join Provectus as we explain how to build an end-to-end infrastructure for machine learning, with a focus on data quality and metadata management, to standardize and streamline machine learning life cycle management (MLOps).
Agenda
- Data Quality and why it matters
- Challenges and solutions of Data Testing
- Challenges and solutions of Model Testing
- MLOps pipelines and why they matter
- How to expand validation pipelines for Data Quality
Identifying and classifying unknown Network Disruptionjagan477830
Since the evolution of modern technology and with the drastic increase in the scale of network communication more and more network disruptions in traffic and private protocols have been taking place. Identifying and classifying the unknown network disruptions can provide support and even help to maintain the backup systems.
Dataset: Gather a large dataset of laptops and their features, including processor speed, RAM, storage, and display size, along with their corresponding prices.
Feature engineering: Extracting meaningful features from the dataset, such as brand, model, and year, and transforming them into a format that machine learning algorithms can use.
Model selection: Choosing the most appropriate machine learning algorithm, such as linear regression, decision tree, or random forest, based on the type of data and desired level of accuracy.
Model training: Splitting the dataset into training and testing sets, and using the training data to train the machine learning model.
Model evaluation: Testing the model's performance on the testing data and evaluating its accuracy using metrics such as mean squared error or R-squared.
Hyperparameter tuning: Optimizing the model's hyperparameters, such as learning rate or regularization strength, to achieve the best performance.
Reliability is concerned with decreasing faults and their impact. The earlier the faults are detected the better. That's why this presentation talks about automated techniques using machine learning to detect faults as early as possible.
A presentation covers how data science is connected to build effective machine learning solutions. How to build end to end solutions in Azure ML. How to build, model, and evaluate algorithms in Azure ML.
What are the Unique Challenges and Opportunities in Systems for ML?Matei Zaharia
Presentation by Matei Zaharia at the SOSP 2019 AI Systems workshop about the systems research challenges specific to machine learning systems, including debugging and performance optimization for ML. Covers research from Stanford DAWN and an industry perspective from Databricks.
MLOps and Data Quality: Deploying Reliable ML Models in ProductionProvectus
Looking to build a robust machine learning infrastructure to streamline MLOps? Learn from Provectus experts how to ensure the success of your MLOps initiative by implementing Data QA components in your ML infrastructure.
For most organizations, the development of multiple machine learning models, their deployment and maintenance in production are relatively new tasks. Join Provectus as we explain how to build an end-to-end infrastructure for machine learning, with a focus on data quality and metadata management, to standardize and streamline machine learning life cycle management (MLOps).
Agenda
- Data Quality and why it matters
- Challenges and solutions of Data Testing
- Challenges and solutions of Model Testing
- MLOps pipelines and why they matter
- How to expand validation pipelines for Data Quality
Identifying and classifying unknown Network Disruptionjagan477830
Since the evolution of modern technology and with the drastic increase in the scale of network communication more and more network disruptions in traffic and private protocols have been taking place. Identifying and classifying the unknown network disruptions can provide support and even help to maintain the backup systems.
Dataset: Gather a large dataset of laptops and their features, including processor speed, RAM, storage, and display size, along with their corresponding prices.
Feature engineering: Extracting meaningful features from the dataset, such as brand, model, and year, and transforming them into a format that machine learning algorithms can use.
Model selection: Choosing the most appropriate machine learning algorithm, such as linear regression, decision tree, or random forest, based on the type of data and desired level of accuracy.
Model training: Splitting the dataset into training and testing sets, and using the training data to train the machine learning model.
Model evaluation: Testing the model's performance on the testing data and evaluating its accuracy using metrics such as mean squared error or R-squared.
Hyperparameter tuning: Optimizing the model's hyperparameters, such as learning rate or regularization strength, to achieve the best performance.
Reliability is concerned with decreasing faults and their impact. The earlier the faults are detected the better. That's why this presentation talks about automated techniques using machine learning to detect faults as early as possible.
A presentation covers how data science is connected to build effective machine learning solutions. How to build end to end solutions in Azure ML. How to build, model, and evaluate algorithms in Azure ML.
What are the Unique Challenges and Opportunities in Systems for ML?Matei Zaharia
Presentation by Matei Zaharia at the SOSP 2019 AI Systems workshop about the systems research challenges specific to machine learning systems, including debugging and performance optimization for ML. Covers research from Stanford DAWN and an industry perspective from Databricks.
This talk includes the following items:
1) discussion of the various stages of ML application life cycle - problem formulation, data definitions, modeling, production system design & implementation, testing, deployment & maintenance, online evaluation & evolution.
2) getting the ML problem formulation right
3) key tenets for different stages of application cycle.
Audio for the talk:
https://youtu.be/oBR8flk2TjQ?t=19207
Multi-modal sources for predictive modeling using deep learningSanghamitra Deb
Using Vision Language models : Is it possible to prompt them similar to LLMs? when to use out of the box and when to pre-train? General multi-modal models --- deeplearning. Machine learning metrics, feature engineering and setting up an ML problem.
Sample Codes: https://github.com/davegautam/dotnetconfsamplecodes
Presentation on How you can get started with ML.NET. If you are existing .NET Stack Developer and Wanna use the same technology into Machine Learning, this slide focuses on how you can use ML.NET for Machine Learning.
Introduction to Machine Learning and Data Science using Autonomous Database ...Sandesh Rao
This session will focus on basics of what Machine Learning is , different types of Machine Learning and Neural Networks , supervised and unsupervised machine learning , autoML for training models and this ends with an example of how to predict workloads using Average Active sessions and different algorithms as an example and also how to predict maintenance windows for your databases. We will also use different open source frameworks as well as some of the tools in the Autonomous Database cloud to do this. If you are a DBA and want to learn something about machine learning and use the tools to perform your tasks more efficiently and automaticall
Machine Learning in Autonomous Data WarehouseSandesh Rao
Machine Learning in Autonomous Data Warehouse: One can use Oracle Autonomous Data Warehouse for machine learning. There are several ways to do this. This presentation explores these different but related options for performing machine learning. Each of these options enables people with different backgrounds to engage with building machine learning solutions on their data. At the end of the session, you will know which option will work best for you
This is from the Bay area Cloud Computing event https://www.meetup.com/All-Things-Cloud-Computing-Bay-Area/events/271017950/
Keynote presentation from ECBS conference. The talk is about how to use machine learning and AI in improving software engineering. Experiences from our project in Software Center (www.software-center.se).
Introduction to Machine Learning and Data Science using the Autonomous databa...Sandesh Rao
This session will focus on basics of what Machine Learning is, different types of Machine Learning and Neural Networks, supervised and unsupervised machine learning, AutoML for training models and this ends with an example of how to predict workloads using Average Active sessions and different algorithms as an example and also how to predict maintenance windows for your databases. We will also use many examples from the ADW Oracle Autonomous Database offering, Oracle Machine Learning library to make this a session with lots of code examples in addition to the theory of Machine Learning and you will walk out having a definitive path to being a data scientist
Law firms & lawyers - rid the manual review of text documents, correspondence, etc. Text Analytics of unstructured documents signals potential knowledge that brings relevance & helps win cases. Moreover, use of text analytics helps offer small firms the same advantage that big firms have. As the information can be used to strengthen solutions and provide advice to attorneys, courtrooms will also benefit from more informed, better prepared legal teams and swift action, keeping long years of litigation away!
Automated EDW Assessment and Actionable Recommendations - Impetus WebinarImpetus Technologies
Assessing analytical workloads is the first step towards successful cloud migration. However, an assessment typically provides a non-actionable list of inventories.
An intelligent automation-based workload assessment offered by Impetus’ Workload Transformation Solution can help you get actionable insights. It profiles workloads and maps their compatibility with your target cloud environment. As a result, you are prepared to avoid common pitfalls and ensure a successful cloud transition of your ETL and analytics workloads.
In this session, our experts will share insights on how this solution can help you:
Identify workload complexities, patterns, and technical debt
Map existing workloads to your target cloud stack
Create a blueprint for future-state architecture based on an automation-based intelligent assessment
Implement best practices to de-risk your cloud transition
We will also share success stories of how Impetus has helped Fortune 500 enterprises make the right decisions for a seamless EDW transformation.
To learn more view our webinar here - https://bit.ly/37zSwML
Introduction to Machine learning - DBA's to data scientists - Oct 2020 - OGBEmeaSandesh Rao
This session will focus on basics of what Machine Learning is , different types of Machine Learning and Neural Networks , supervised and unsupervised machine learning with examples, AutoML for training models and this ends with an example of how to predict fraud , to determining shopping patterns to Wine picking and different algorithms as an example and also how to predict workload for your databases. We will also use OML in the Autonomous Database cloud to do this. If you are a DBA and want to learn something about machine learning and use the tools to perform your tasks more efficiently and automatically
Introduction to Machine Learning - From DBA's to Data Scientists - OGBEMEASandesh Rao
This session will focus on basics of what Machine Learning is , different types of Machine Learning and Neural Networks , supervised and unsupervised machine learning with examples, AutoML for training models and this ends with an example of how to predict fraud , to determining shopping patterns to Wine picking and different algorithms as an example and also how to predict workload for your databases. We will also use OML in the Autonomous Database cloud to do this. If you are a DBA and want to learn something about machine learning and use the tools to perform your tasks more efficiently and automatically
Guiding through a typical Machine Learning PipelineMichael Gerke
Many People are talking about AI and Machine Learning. Here's a quick guideline how to manage ML Projects and what to consider in order to implement machine learning use cases.
#Interactive Session by Anindita Rath and Mahathee Dandibhotla, "From Good to...Agile Testing Alliance
#Interactive Session by Anindita Rath and Mahathee Dandibhotla, "From Good to Great: Enhancing Testability in Software Testing " at ATAGTR2023
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
#Interactive Session by Ajay Balamurugadas, "Where Are The Real Testers In T...Agile Testing Alliance
#Interactive Session by Ajay Balamurugadas, "Where Are The Real Testers In The Age of AI? " at ATAGTR2023
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
More Related Content
Similar to #Interactive Session by Vivek Patle and Jahnavi Umarji, "Empowering Functional Testing with Support Vector Machines: An Experimental Journey" at #ATAGTR2023.
This talk includes the following items:
1) discussion of the various stages of ML application life cycle - problem formulation, data definitions, modeling, production system design & implementation, testing, deployment & maintenance, online evaluation & evolution.
2) getting the ML problem formulation right
3) key tenets for different stages of application cycle.
Audio for the talk:
https://youtu.be/oBR8flk2TjQ?t=19207
Multi-modal sources for predictive modeling using deep learningSanghamitra Deb
Using Vision Language models : Is it possible to prompt them similar to LLMs? when to use out of the box and when to pre-train? General multi-modal models --- deeplearning. Machine learning metrics, feature engineering and setting up an ML problem.
Sample Codes: https://github.com/davegautam/dotnetconfsamplecodes
Presentation on How you can get started with ML.NET. If you are existing .NET Stack Developer and Wanna use the same technology into Machine Learning, this slide focuses on how you can use ML.NET for Machine Learning.
Introduction to Machine Learning and Data Science using Autonomous Database ...Sandesh Rao
This session will focus on basics of what Machine Learning is , different types of Machine Learning and Neural Networks , supervised and unsupervised machine learning , autoML for training models and this ends with an example of how to predict workloads using Average Active sessions and different algorithms as an example and also how to predict maintenance windows for your databases. We will also use different open source frameworks as well as some of the tools in the Autonomous Database cloud to do this. If you are a DBA and want to learn something about machine learning and use the tools to perform your tasks more efficiently and automaticall
Machine Learning in Autonomous Data WarehouseSandesh Rao
Machine Learning in Autonomous Data Warehouse: One can use Oracle Autonomous Data Warehouse for machine learning. There are several ways to do this. This presentation explores these different but related options for performing machine learning. Each of these options enables people with different backgrounds to engage with building machine learning solutions on their data. At the end of the session, you will know which option will work best for you
This is from the Bay area Cloud Computing event https://www.meetup.com/All-Things-Cloud-Computing-Bay-Area/events/271017950/
Keynote presentation from ECBS conference. The talk is about how to use machine learning and AI in improving software engineering. Experiences from our project in Software Center (www.software-center.se).
Introduction to Machine Learning and Data Science using the Autonomous databa...Sandesh Rao
This session will focus on basics of what Machine Learning is, different types of Machine Learning and Neural Networks, supervised and unsupervised machine learning, AutoML for training models and this ends with an example of how to predict workloads using Average Active sessions and different algorithms as an example and also how to predict maintenance windows for your databases. We will also use many examples from the ADW Oracle Autonomous Database offering, Oracle Machine Learning library to make this a session with lots of code examples in addition to the theory of Machine Learning and you will walk out having a definitive path to being a data scientist
Law firms & lawyers - rid the manual review of text documents, correspondence, etc. Text Analytics of unstructured documents signals potential knowledge that brings relevance & helps win cases. Moreover, use of text analytics helps offer small firms the same advantage that big firms have. As the information can be used to strengthen solutions and provide advice to attorneys, courtrooms will also benefit from more informed, better prepared legal teams and swift action, keeping long years of litigation away!
Automated EDW Assessment and Actionable Recommendations - Impetus WebinarImpetus Technologies
Assessing analytical workloads is the first step towards successful cloud migration. However, an assessment typically provides a non-actionable list of inventories.
An intelligent automation-based workload assessment offered by Impetus’ Workload Transformation Solution can help you get actionable insights. It profiles workloads and maps their compatibility with your target cloud environment. As a result, you are prepared to avoid common pitfalls and ensure a successful cloud transition of your ETL and analytics workloads.
In this session, our experts will share insights on how this solution can help you:
Identify workload complexities, patterns, and technical debt
Map existing workloads to your target cloud stack
Create a blueprint for future-state architecture based on an automation-based intelligent assessment
Implement best practices to de-risk your cloud transition
We will also share success stories of how Impetus has helped Fortune 500 enterprises make the right decisions for a seamless EDW transformation.
To learn more view our webinar here - https://bit.ly/37zSwML
Introduction to Machine learning - DBA's to data scientists - Oct 2020 - OGBEmeaSandesh Rao
This session will focus on basics of what Machine Learning is , different types of Machine Learning and Neural Networks , supervised and unsupervised machine learning with examples, AutoML for training models and this ends with an example of how to predict fraud , to determining shopping patterns to Wine picking and different algorithms as an example and also how to predict workload for your databases. We will also use OML in the Autonomous Database cloud to do this. If you are a DBA and want to learn something about machine learning and use the tools to perform your tasks more efficiently and automatically
Introduction to Machine Learning - From DBA's to Data Scientists - OGBEMEASandesh Rao
This session will focus on basics of what Machine Learning is , different types of Machine Learning and Neural Networks , supervised and unsupervised machine learning with examples, AutoML for training models and this ends with an example of how to predict fraud , to determining shopping patterns to Wine picking and different algorithms as an example and also how to predict workload for your databases. We will also use OML in the Autonomous Database cloud to do this. If you are a DBA and want to learn something about machine learning and use the tools to perform your tasks more efficiently and automatically
Guiding through a typical Machine Learning PipelineMichael Gerke
Many People are talking about AI and Machine Learning. Here's a quick guideline how to manage ML Projects and what to consider in order to implement machine learning use cases.
Effective Software Effort Estimation Leveraging Machine Learning for Digital ...Shakas Technologies
Effective Software Effort Estimation Leveraging Machine Learning for Digital Transformation.
Shakas Technologies ( Galaxy of Knowledge)
#11/A 2nd East Main Road,
Gandhi Nagar,
Vellore - 632006.
Mobile : +91-9500218218 / 8220150373| land line- 0416- 3552723
Shakas Training & Development | Shakas Sales & Services | Shakas Educational Trust|IEEE projects | Research & Development | Journal Publication |
Email : info@shakastech.com | shakastech@gmail.com |
website: www.shakastech.com
Facebook: https://www.facebook.com/pages/Shakas-Technologies
Similar to #Interactive Session by Vivek Patle and Jahnavi Umarji, "Empowering Functional Testing with Support Vector Machines: An Experimental Journey" at #ATAGTR2023. (20)
#Interactive Session by Anindita Rath and Mahathee Dandibhotla, "From Good to...Agile Testing Alliance
#Interactive Session by Anindita Rath and Mahathee Dandibhotla, "From Good to Great: Enhancing Testability in Software Testing " at ATAGTR2023
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
#Interactive Session by Ajay Balamurugadas, "Where Are The Real Testers In T...Agile Testing Alliance
#Interactive Session by Ajay Balamurugadas, "Where Are The Real Testers In The Age of AI? " at ATAGTR2023
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
#Interactive Session by Jishnu Nambiar and Mayur Ovhal, "Monitoring Web Per...Agile Testing Alliance
#Interactive Session by Jishnu Nambiar and Mayur Ovhal, "Monitoring Web Performance: Leveraging Grafana and Selenium for Real-Time Issue Alerts" at ATAGTR2023
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
#Interactive Session by Pradipta Biswas and Sucheta Saurabh Chitale, "Navigat...Agile Testing Alliance
#Interactive Session by Pradipta Biswas and Sucheta Saurabh Chitale, "Navigating the IoT Performance Testing Landscape" at #ATAGTR2023.
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
#Interactive Session by Apoorva Ram, "The Art of Storytelling for Testers" at...Agile Testing Alliance
#Interactive Session by Apoorva Ram, "The Art of Storytelling for Testers" at #ATAGTR2023.
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
#Interactive Session by Nikhil Jain, "Catch All Mail With Graph" at #ATAGTR2023.Agile Testing Alliance
#Interactive Session by Nikhil Jain, "Catch All Mail With Graph" at #ATAGTR2023.
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
#Interactive Session by Ashok Kumar S, "Test Data the key to robust test cove...Agile Testing Alliance
#Interactive Session by Ashok Kumar S, "Test Data the key to robust test coverage" at #ATAGTR2023.
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
#Interactive Session by Seema Kohli, "Test Leadership in the Era of Artificia...Agile Testing Alliance
#Interactive Session by Seema Kohli, "Test Leadership in the Era of Artificial Intelligence" at #ATAGTR2023.
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
#Interactive Session by Ashwini Lalit, RRR of Test Automation Maintenance" at...Agile Testing Alliance
#Interactive Session by Ashwini Lalit, RRR of Test Automation Maintenance" at #ATAGTR2023.
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
#Interactive Session by Srithanga Aishvarya T, "Machine Learning Model to aut...Agile Testing Alliance
#Interactive Session by Srithanga Aishvarya T, "Machine Learning Model to automate performance test script development using Jmeter" at #ATAGTR2023.
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
#Interactive Session by Kirti Ranjan Satapathy and Nandini K, "Elements of Qu...Agile Testing Alliance
#Interactive Session by Kirti Ranjan Satapathy and Nandini K, "Elements of Quality Engineering in Remote IoT System" at #ATAGTR2023.
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
#Interactive Session by Sudhir Upadhyay and Ashish Kumar, "Strengthening Test...Agile Testing Alliance
#Interactive Session by Sudhir Upadhyay and Ashish Kumar, "Strengthening Testing Oversight Using Environment Automation" at #ATAGTR2023.
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
#Interactive Session by Sayan Deb Kundu, "Testing Gen AI Applications" at #AT...Agile Testing Alliance
#Interactive Session by Sayan Deb Kundu, "Testing Gen AI Applications" at #ATAGTR2023.
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
#Interactive Session by Dinesh Boravke, "Zero Defects – Myth or Reality" at #...Agile Testing Alliance
#Interactive Session by Dinesh Boravke, "Zero Defects – Myth or Reality" at #ATAGTR2023.
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
#Interactive Session by Saby Saurabh Bhardwaj, "Redefine Quality Assurance – Journey from Centralized to Decentralized, Distributed Blockchain/Web3 testing" at #ATAGTR2023.
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
#Keynote Session by Sanjay Kumar, "Innovation Inspired Testing!!" at #ATAGTR2...Agile Testing Alliance
#Keynote Session by Sanjay Kumar, "Innovation Inspired Testing!!" at #ATAGTR2023.
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
#Keynote Session by Schalk Cronje, "Don’t Containerize me" at #ATAGTR2023.Agile Testing Alliance
#Keynote Session by Schalk Cronje, "Don’t Containerize me" at #ATAGTR2023.
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
#Interactive Session by Chidambaram Vetrivel and Venkatesh Belde, "Revolution...Agile Testing Alliance
#Interactive Session by Chidambaram Vetrivel and Venkatesh Belde, "Revolutionizing Security Testing with AI" at #ATAGTR2023.
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
#Interactive Session by Aniket Diwakar Kadukar and Padimiti Vaidik Eswar Dat...Agile Testing Alliance
#Interactive Session by Aniket Diwakar Kadukar and Padimiti Vaidik Eswar Datta, "A Holistic Testing Methodology for Immersive Experience in AR, VR, and the Metaverse" at #ATAGTR2023.
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
#Interactive Session by Sumit Mundhada, "Observability in Software Testing" a...Agile Testing Alliance
#Interactive Session by Sumit Mundhada, "Observability in Software Testing"at #ATAGTR2023.
#ATAGTR2023 was the 8th Edition of Global Testing Retreat.
To know more about #ATAGTR2023, please visit: https://gtr.agiletestingalliance.org/
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...Ramesh Iyer
In today's fast-changing business world, Companies that adapt and embrace new ideas often need help to keep up with the competition. However, fostering a culture of innovation takes much work. It takes vision, leadership and willingness to take risks in the right proportion. Sachin Dev Duggal, co-founder of Builder.ai, has perfected the art of this balance, creating a company culture where creativity and growth are nurtured at each stage.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
Generating a custom Ruby SDK for your web service or Rails API using Smithyg2nightmarescribd
Have you ever wanted a Ruby client API to communicate with your web service? Smithy is a protocol-agnostic language for defining services and SDKs. Smithy Ruby is an implementation of Smithy that generates a Ruby SDK using a Smithy model. In this talk, we will explore Smithy and Smithy Ruby to learn how to generate custom feature-rich SDKs that can communicate with any web service, such as a Rails JSON API.
Generating a custom Ruby SDK for your web service or Rails API using Smithy
#Interactive Session by Vivek Patle and Jahnavi Umarji, "Empowering Functional Testing with Support Vector Machines: An Experimental Journey" at #ATAGTR2023.
2. • Less than and Equal to 1
• Between 2 to 5
• Between 5-10
• More than 10
3. Background
Functional Testing Challenges
Test Case Prioritization: Limited resources and time for executing all test cases.
Defect Prediction: Identifying potential defects early in the development lifecycle.
Code Review Assistance: Ensuring code quality during development.
Regression Testing Optimization: Efficiently identifying impacted areas for regression
testing after changes.
Automation Test Script Generation: Writing and maintaining automated test scripts for
evolving applications
Test Case Prioritization and Defect Prediction Solutions
SVMs can assist in prioritizing test cases by predicting which test cases are more likely
to uncover defects. Test case attributes like code coverage, historical defect data, and
code complexity can be used as features for SVM-based prioritization.
SVMs can analyze historical data to predict the likelihood of defects in specific modules
or components. This can aid in focusing testing efforts on critical areas.
4. Agenda
• Understanding the Terminology
• Machine Learning Overview
• Machine Learning Workflow
• Data Preparation
• Challenges in Data Preparation
• Data Quality Issues and Model Impact
• Training, Validation and Test Datasets
• Support Vector Machine – Introduction
• Why SVM for Functional Testing
• Use Case 1 –Test Case Prioritization
• Use Case 2 – Defect Prediction
5. Understanding Terminologies
Generative AI, a branch of artificial intelligence and a subset of Deep Learning, focuses on creating
models capable of generating new content that resemble existing data. These models aim to generate
content that is indistinguishable from what might be created by humans. Generative Adversarial Networks
(GANs) are popular examples of generative AI models that use deep neural networks to generate realistic
content such as images, text, or even music
Example : Image Generation, Video Synthesis, Social Media Content Generation
AI is broadly defined as the ability of machines to mimic human behavior. It encompasses a broad range
of techniques and approaches aimed at enabling machines to perceive, reason, learn, and make
decisions. AI can be rule-based, statistical, or involve machine learning algorithms.
Example : Virtual Assistants , Healthcare Diagnosis and Imaging , Virtual Reality and Augmented
Reality
The term “ML” focuses on machines learning from data without the need for explicit programming.
Machine Learning algorithms leverage statistical techniques to automatically detect patterns and make
predictions or decisions based on historical data that they are trained on. While ML is a subset of AI, the
term was coined to emphasize the importance of data-driven learning and the ability of machines to
improve their performance through exposure to relevant data
Example: Predictive text, recommendation systems , Time Series Forecasting
6. 01
02
03
04
Definition of machine learning
Supervised learning vs
unsupervised learning
Common algorithms used in
machine learning
Real-world applications of
machine learning
Introduction to Machine Learning
8. Machine Learning
Models
• Supervised Learning
• Classification
• Spam email classification example
• Regression
• Predicting software development
effort
• Unsupervised Learning
• Clustering
• Grouping similar software modules
• Association
• Discovering relationships
between software features
9. 01
03
02
04
Explanation of supervised
learning
Types of supervised learning:
classification and regression
Examples of supervised
learning algorithms: linear
regression, decision trees,
support vector machines
Training and testing process in
supervised learning
Supervised Learning
11. Machine
Learning
Workflow
Data Collection Emphasis on high-quality
data for effective ML
Data Preparation Transition to the next
slide
Model Selection
Importance of choosing
appropriate ML
algorithms
Training Models learn patterns
from the training data
Evaluation Assessment of model
performance
12. Data Preparation in
ML Workflow
• Importance of Data Preparation
• Success of ML models depends on quality
data
• Steps Involved
• Data Cleaning
• Remove or handle missing values,
outliers
• Data Transformation
• Normalize or scale features
• Feature Engineering
• Create or modify features
13. Challenges in Data
Preparation
• Common Challenges
• Missing Data
• Example: Handling missing values in software
defect prediction
• Outliers
• Example: Impact of outliers in user behavior
data
• Feature Scaling
• Example: Unscaled features in code complexity
metrics
• Categorical Data Handling
• Example: Challenges of non-numeric data in
customer feedback analysis
14. Data Quality Issues
and Model Impact
• Impact of Poor Data Quality
• Low-quality data leads to inaccurate models
• Examples of Data Quality Issues
• Inconsistent Data
• Example: Inconsistencies in user
feedback data
• Biased Data
• Example: Addressing biases in
demographic-specific data
• Noisy Data
• Example: Random variations in
performance metrics
15. Training, Validation,
and Test Datasets
• Need for Splitting Data
• Importance of dividing data for fair
model assessment
• Overview of Datasets
• Training Dataset
• Used for teaching the model
• Validation Dataset
• Fine-tuning hyperparameters
• Test Dataset
• Unbiased evaluation of the model
17. Support Vector Machine
SVMs are a type of machine learning algorithm that have gained
popularity in recent years due to their effectiveness in solving complex
problems
SVMs can be used for both classification and regression tasks
The basic idea behind SVMs is to find the hyperplane that best
separates the data into different classes or predicts the value of
the target variable in regression. The hyperplane is chosen so
that it maximizes the margin between the two classes or the
predicted values and the actual values in regression
To find the optimal hyperplane, SVMs use a technique
called kernel trick, which transforms the input data into a
higher-dimensional space where it becomes easier to find a
separating hyperplane. Common kernel functions include
linear, polynomial, radial basis function (RBF), and sigmoid.
18. SVM for classification
• In classification, SVMs are used to separate data into different classes based on their attributes . SVMs can
be applied to various classification tasks in software testing due to their ability to handle complex decision
boundaries and non-linear relationships. Some applications are:
Software Defect Prediction:
SVMs can predict whether a piece of code is likely to contain defects based on features
such as code metrics, historical defect data, and complexity measures.
Anomaly Detection:
SVMs can be used to classify abnormal patterns in system behavior, identifying potential
software anomalies or performance issues.
Requirement Traceability:
SVMs can classify whether a given piece of code corresponds to a specific requirement,
facilitating traceability and impact analysis.
19. SVM for regression
• In regression, SVMs are used to predict numerical values based on input data SVMs can be applied to
various regression tasks in software testing, providing solutions to predict and optimize continuous
outcomes. Here are some applications:
Test Suite Prioritization:
SVM regression can predict the priority or importance of test cases within a test suite, optimizing the order in
which test cases are executed based on predicted outcomes.
Resource Consumption Estimation:
SVM regression can be used to estimate resource consumption during the execution of software tests,
helping in resource allocation and planning.
Code Complexity Prediction:
SVM regression models can predict code complexity metrics, aiding in identifying potentially challenging or
error-prone code sections.
20. Defect Prediction
• Definition and Importance
• Predicts areas likely to contain defects
• Proactively identifies and addresses
high-risk areas
• Complements Test Case Prioritization
• Prioritization focuses on predicted
defect locations
• Defect prediction informs prioritization
22. Data Preparation
GHPR Dataset
• GHPR is a public dataset to identify bug fixing based on Pull
Requests(PRs) in Github
• It has a total of 6052 instances which contain 3026 defective
instances and 3026 non-defective instances
• It uses 21 static metrics for the total 6052 instances which is the
data used for the baseline approaches
Reference:
@ARTICLE{***,
author={Jiaxi Xu;Fei Wang;Jun Ai},
journal={IEEE Transactions on Reliability},
title={Defect Prediction With Semantics and Context Features of
Codes Based on Graph Representation Learning},
year={2021}, }
Metrics name Known as
Coupling between objects CBO
Weight method class WMC
Depth inheritance tree DIT
Response for a class RFC
Lack of cohesion of methods LCOM
Counts the number of methods totalMethods
Counts the number of fields totalFields
Lines of code LOC
Quantity of returns returnQty
Quantity of loops loopQty
Quantity of comparisons comparisonsQty
Quantity of try/catches tryCatchQty
Quantity of parenthesized
expressions
parenthesizedExpsQty
String literals stringLiteralsQty
Quantity of number numbersQty
Quantity of variables assignmentsQty
Quantity of math operations mathOperationsQty
Quantity of variables variablesQty
Max nested blocks maxNestedBlocks
Number of unique words uniqueWordsQty
21 static metrics
23. Feature Selection and Data Pre-
Processing
• Feature Selection:
Identify and select features that are crucial for defect prediction
For our dataset, all 21 metrics are selected as features
• Data pre-processing:
Before training SVM model, the data is Pre-processed.
The selected features are normalized to ensure that they have a consistent scale
• Data Splitting:
The dataset were split between train and validation sets using the 80/20 split
values.
24. Training the SVM Model
• SVM Model Selection:
An appropriate SVM variant (linear, polynomial, or radial basis function kernel)is chosen based
on the characteristics of the dataset.
Since the relationship between input features and the target variable is non-linear and we want
our SVM model to capture complex patterns and decision boundaries in the data, we chose
Radial Basis (RBF) Kernel.
• Model Training:
SVM model is trained on the training dataset.
The SVM algorithm aims to find the hyperplane that best separates instances of different
classes.
• Hyperparameter Tuning:
Fine-tune the model's hyperparameters, such as the regularization parameter (C) and kernel
parameters, to optimize the model's performance
To find the optimal values of hyperparameters, Grid Search technique is used which
25. Model Evaluation
• Our SVM model achieved an overall accuracy of approximately
72.9% on the evaluation dataset
• In predicting class 0 (Non-defective), the model demonstrated a
precision of 82%, recall of 60%, and an F1-score of 69%
• For class 1 (Defective), the precision was 68%, recall was 86%, and
the F1-score reached 76%.
Class Precision Recall F1-
score
Non-defective (0) 82% 60% 69%
Defective (1) 68% 86% 76%
Overall Accuracy 72.9%
26. Model Evaluation
• Area Under the Curve (AUC) is
0.8, indicating a relatively high
discriminatory power of the
model
• It also indicates that the model
has a good ability to distinguish
between positive and negative
instances
27. Use Case - Test
Case
Prioritization
Application of
SVMs
SVMs applied to
prioritize test cases
Results and
Insights
Showcase of key
findings, possibly
through charts or
graphs
Performance
Improvements
How SVMs
contributed to
enhancing test case
prioritization
Lessons Learned
and Challenges
Faced
Insights gained
during the test case
prioritization
experiment
29. Data Preparation
Industrial Datasets:
• The Cisco dataset is a test suite used for testing video-conferencing systems, provided by Cisco
Systems.
• The other two industrial datasets are used for testing industrial robotics applications provided by
ABB Robotics (Paint Control and Input/Output Control, noted as IOF/ROL)
Synthetic Datasets (Data Augmentation):
• The ratio of failed test executions is extremely low in industrial dataset
• To address the problem of insufficient representation of relevant test cases in the industrial
datasets, [4] performed data augmentation. Specifically, uses SMOGN which is a technique for
tackling imbalanced regression datasets by generating diverse new data points for the given data.
• The synthetic data generated was concatenated to the industrial datasets
• We have used this concatenated dataset from [4] for our use case
30. Feature Selection and Pre-processing
• Data preprocessing was necessary to format all the three test suites in the same way, for example, to make
the number and type of features constant across all the files
• Features selected were:
• Target Variable: Priority Values for the testcases
• Our target is to minimize the loss between the actual and predicted priority values for prioritizing test cases in
the correct descending order
Column Name Content
DurationFeatur
e
Average test case execution time computed across all its previous executions
E1, E2, E3 Execution Status (0,1)
LastRunFeatur
e
Previous last execution of the test case as date-time-string (Format: YYYY-MM-
DD HH:ii)
Distance
Absolute difference of a test case execution status between the least recent and
most recent CI cycle
ChangeInStatu
s
Number of times a test case execution status has changed from pass to fail in all
its previous executions
31. Training SVM Regression Model
• Data Split: The dataset were split between train and validation sets using
the 80/20 split values.
• Choose SVM Regression:
SVM regression variant, SVR (Support Vector Regressor) is chosen which is
specifically designed for regression tasks
Radial Basis Kernel was chosen for the dataset
• Train SVM Model: SVM regression model is trained on the training set
• Hyperparameter tuning: To find the optimal values of hyperparameters
like C and gamma, Grid Search technique is used which exhaustively
searches through a manually specified subset of hyperparameter values
32. Model Evaluation
• Our SVR model’s performance is evaluated using regression-specific metrics like:
1. Mean Absolute Error (MAE): Measures the average absolute errors between predicted and actual
values. This serves as the loss function for regression
2. R-squared (R2): Represents the proportion of the variance in the dependent variable that is predictable
from the independent variables.
MAE and R2-score of our model came out to be:
• An MAE of 0.050 suggests that, on average, the absolute difference between your predicted and actual
'PRIORITY_VALUE' is very small which indicates a high level of accuracy in the predictions
• An R2 value of 0.985 means that approximately 98.5% of the variability in 'PRIORITY_VALUE' is captured
by the SVR model. This is an excellent result, indicating that our model is effectively capturing the
underlying patterns in the data.
33. Model Evaluation
• To assess the regression analysis visually,
a scatter plot is used:
Blue Points (Actual Values): Each
blue point represents the actual
'PRIORITY_VALUE' from the test set.
This line represents the ideal scenario
where the predicted values perfectly
match the actual values.
Red Cross Markers (Predicted
Values): Each red cross marker
represents the corresponding
predicted 'PRIORITY_VALUE' for a
data point in the test set.
• Since the red markers follow the diagonal
line closely, it suggests that the model is
making accurate predictions.
34. Q&A
• Engage Audience
• Encourage questions and discussions
• Clarify any queries related to the experimental
journey
35. Conclusion
• Main Takeaways
• The value of SVMs in enhancing functional
testing
• Contribution of prioritization and defect
prediction to efficient testing
38. Feature
Selection
Chosen Features
For prioritization:
code coverage,
historical defect data,
complexity metrics
For defect prediction:
code metrics,
historical bug data,
code changes
Importance of
Feature
Engineering
Enhances SVM
model performance
39. Data Preparation
• Data Collection and Preparation
• Gather historical data, extract features,
create labeled dataset
• Ensure dataset represents the software
characteristics
40. Training the SVM
Model
• SVM Training Process
• Split dataset into training and testing sets
• Choose kernel, tune parameters for optimal
performance
41. Evaluation Metrics
• Metrics for Evaluation
• Prioritization: accuracy, precision, recall, F1-
score
• Defect Prediction: similar metrics for identifying
defect-prone areas
45. Lessons Learned
• Key Lessons
• Importance of well-defined features and
representative datasets
• Adaptability of SVMs in addressing
testing challenges
46. Future Directions
• Future Directions
• Explore new features or data sources
• Consider improvements to SVM model or
alternative techniques
47. References
1. https://www.analyticsvidhya.com/blog/2021/10/support-vector-machinessvm-a-complete-guide-for-
beginners/
2. Khatibsyarbini M, Isa MA, Jawawi DN, Tumeng R. Test case prioritization approaches in regression
testing: A systematic literature review. Information and Software Technology 2017; 93: 74–93. doi:
10.1016/j.infsof.2017.08.014
3. Radial Basis Function (RBF) Kernel: The Go-To Kernel | by Sushanth Sreenivasa | Towards Data Science
4. A. Sharif, D. Marijan and M. Liaaen, "DeepOrder: Deep Learning for Test Case Prioritization in Continuous
Integration Testing," 2021 IEEE International Conference on Software Maintenance and Evolution
(ICSME), Luxembourg, 2021, pp. 525-534, doi: 10.1109/ICSME52107.2021.00053.
5. http://promise.site.uottawa.ca/SERepository - PROMISE Dataset
6. GHPR_dataset/README.md at master · feiwww/GHPR_dataset · GitHub – GHPR Dataset
7. https://synoptek.com/insights/it-blogs/data-insights/ai-ml-dl-and-generative-ai-face-off-a-comparative-
analysis/#:~:text=AI%20serves%20as%20the%20broad,Generative%20AI%20creates%20new%20conten
t.
48. Appendix
• Optimal hyperparameter values chosen:
• Classification-specific evaluation metrics:
Use case C Gamma Kernel
Software Defect
Prediction
1000 0.01 RBF
Test Suite Prioritization 100 0.01 RBF
Metric Description Formula
Accuracy Number of correctly classified data instances
over the total number of instances Accuracy =
𝑇𝑁+𝑇𝑃
𝑇𝑁+𝐹𝑃+𝑇𝑃+𝐹𝑁
Precision Positive Predictive value, should ideally be 1. Precision =
𝑇𝑃
𝐹𝑃+𝑇𝑃
Recall Sensitivity or true positive rate, should ideally be
1
Precision =
𝑇𝑃
𝐹𝑁+𝑇𝑃
F1-score Harmonic mean of precision and recall ,
becomes 1 only when precision and recall are
both 1
F1-score = 2 *
𝑃𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛 ∗𝑅𝑒𝑐𝑎𝑙𝑙
𝑃𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛+𝑅𝑒𝑐𝑎𝑙𝑙
49. Appendix
• Regression-specific Evaluation metrics:
Metric Description Formula
Mean
Absolute
Error (MAE)
Average of the absolute differences
between the predicted and actual values.
Lower MAE indicates better model
performance
MAE=
1
𝑛 𝑖=1
𝑛
∣yi−y^i∣
n: number of data points
yi: actual value
y^i: predicted value
R-squared
(R2) score
Measures the proportion of the variance
in the dependent variable that is
predictable from the independent
variables. A higher R2 score suggests a
better fit of the model to the data
𝑅2
= 1 -
𝑖=1
𝑛
(𝑦𝑡𝑟𝑢𝑒,𝑖 − 𝑦𝑝𝑟𝑒𝑑,𝑖)2
𝑖=1
𝑛
(𝑦𝑡𝑟𝑢𝑒,𝑖 − 𝑦𝑡𝑟𝑢𝑒
− )2
𝑦𝑡𝑟𝑢𝑒
−
∶ mean of actual target values
50. Kernel Functions and Hyperparameters
Hyperparameters: Hyperparameters are external configurations or settings for an ML model that are not learned from the training data but
rather set by the user. Some key hyperparameters for SVM:
o C (Regularization Parameter): The regularization parameter C trades off correct classification of training examples against maximizing the
decision function's margin. A smaller C encourages a larger margin but may misclassify some training points. A larger C penalizes
misclassifications more heavily but results in a smaller margin.
o Kernel: SVMs use a kernel function to transform the input features into a higher-dimensional space. The choice of the kernel determines
the shape of the decision boundary. Common kernel functions include:
Linear Kernel (kernel='linear'): Suitable for linearly separable data.
Polynomial Kernel (kernel='poly'): Introduces polynomial features to handle non-linear decision boundaries.
Radial Basis Function (RBF) Kernel (kernel='rbf' or kernel='sigmoid'): Useful for capturing complex non-linear relationships.
o Gamma (Kernel Coefficient): Relevant for RBF and polynomial kernels. It defines the influence of a single training example, with low
values meaning 'far' and high values meaning 'close.' A small gamma leads to a smooth decision boundary, while a large gamma can
result in a more complex, wiggly boundary.
o Degree (Degree of the Polynomial Kernel): Relevant for polynomial kernels. It specifies the degree of the polynomial kernel function.
Higher degrees can capture more complex relationships but may also lead to overfitting.
o Class Weights: For imbalanced datasets, you can assign different weights to different classes to influence the optimization process.
51. Appendix
ROC (Receiver Operating Characteristic) :
• It is a graphical representation used in binary classification to assess the performance of a classification model at
different threshold settings
• The ROC curve is a useful tool for visualizing the trade-off between sensitivity and specificity
• Area Under the ROC Curve (AUC-ROC):
• AUC-ROC represents the area under the ROC curve. A higher AUC-ROC value indicates better model
performance.
• AUC-ROC = 0.5 corresponds to random guessing, while AUC-ROC = 1 indicates a perfect classifier.
• Threshold Setting:
• ROC curves are created by plotting the TPR against the FPR at various threshold settings for the model. Each point
on the curve corresponds to a different threshold.
• Diagonal Line (Random Classifier):
• The diagonal line (from (0,0) to (1,1)) represents the performance of a random classifier that makes predictions by
chance.
• Optimal Point: