- Machine learning models are negatively impacted by noisy or inconsistent labels in training data. This is a challenge for tasks like bug severity classification where labels can be subjective.
- A new evaluation metric called Krippendorff's alpha is proposed to measure agreement between labels while accounting for inconsistencies. It is shown to better reflect performance than accuracy when labels are inconsistent.
- Making "big data thick" by improving quality is an important future direction, but challenging at scale. Lightweight methods are needed to reduce noise without extensive manual labelling. Performance measures also need to account for noise inherent in some real-world problems.
what is user support system???
This file will provide detailed overview about the user support system and how it will works in human computer interaction and why we need it .....
This document provides an overview of deep recommender systems and some of their shortcomings. It discusses neural network architectures like NeuMF, Wide&Deep, Neural FM, DeepFM, and DSCF that have been applied to recommendation. It also covers sequential recommendation methods, optimization techniques, and challenges like short-term rewards, manually designed architectures, isolated data, and security issues like poisoning attacks.
Applying Deep Learning with Weak and Noisy labelsDarian Frajberg
Scientific seminar at Politecnico di Milano
Como, Italy
September 2018
In recent years, Deep Learning has achieved outstanding results outperforming previous techniques and even humans, thus becoming the state-of-the-art in a wide range of tasks, among which Computer Vision has been one of the most benefited areas.
Nonetheless, most of this success is tightly coupled to strongly supervised learning tasks, which require highly accurate, expensive and labor-intensive defined ground truth labels.
In this presentation, we will introduce diverse alternatives to deal with this problem and support the training of Deep Learning models for Computer Vision tasks by simplifying the process of data labelling or exploiting the unlimited supply of publicly available data in Internet (such as user-tagged images from Flickr). Such alternatives rely on data comprising noisy and weak labels, which are much easier to collect but require special care to be used.
Botnet detection using Wgans for securityssuser3f5a831
This document proposes a GAN-based technique to improve botnet detection by addressing class imbalance issues in datasets. It involves using a WGAN to generate synthetic samples for the underrepresented botnet class, which are then combined with the original dataset and used to train and evaluate several machine learning classifiers. The proposed "Bot-WGANs" method achieves over 99.9% accuracy, 100% precision, 99.8% recall, and 99.9% F1 score on the CICIDS2017 dataset, outperforming existing methods that use oversampling techniques like SMOTE. The approach shows potential for enhancing cybersecurity by simplifying and strengthening botnet detection.
Dear Sir/Ma’am
I am interested to work as a data specialist in your organization. I believe my experience, skills and work attitude will aid your organization in a great way. Please accept my enclosed resume with this letter.
I worked at Accenture for the last four years. My key responsibilities here were to collect, analyse, store and create data. I made sure that these data were accurate and not damaged. As far as my educational background is concerned, I have a bachelor's degree in EXTC. I am excellent at solving problems and have great analytical skills. I am capable of working well with network administration and can explain the technical problems.
I would appreciate if we could meet up for an interview wherein we can discuss more on this. I can be contacted at +919493377607 or you can email me at imtiaz.khan.sw39@gmail.com
Thank You.
Yours sincerely,
Imtiaz Khan
This document discusses responsible artificial intelligence. It begins by showing survey results that most consumers want organizations to be accountable for misusing AI and want privacy protected. It then defines responsible AI as evaluating, developing, and implementing AI safely, reliably, and ethically. The main principles discussed are privacy using differential privacy, fairness by mitigating unfair impacts on groups, and transparency through explainable AI tools. General recommendations are given such as clarifying a system's purpose, considering social biases, and encouraging feedback. Benefits of responsible AI include minimizing unintentional bias and ensuring transparency.
Review of Algorithms for Crime Analysis & PredictionIRJET Journal
This document reviews algorithms that can be used for crime analysis and prediction. It discusses various data mining and machine learning techniques including classification algorithms like decision trees, k-nearest neighbors, and random forests as well as clustering algorithms like k-means clustering. Deep learning techniques are also examined for identifying relationships between different types of crimes and predicting where and when crimes may occur. The document evaluates these different algorithmic approaches and concludes that major developments in data science and machine learning now allow for effective crime analysis and prediction by discovering patterns in criminal data.
SEMI SUPERVISED BASED SPATIAL EM FRAMEWORK FOR MICROARRAY ANALYSISIRJET Journal
This document presents a semi-supervised spatial EM framework for microarray analysis to efficiently classify and predict diseases based on gene expression data. It uses a spatial EM algorithm to cluster gene expression data, followed by an SVM classifier to predict diseases and their severity levels. The proposed approach is evaluated based on classification accuracy, computation time, and ability to identify biologically significant genes. Experimental results on disease datasets show improved accuracy compared to other supervised and unsupervised methods. The authors conclude that using the same classifier for gene selection and classification enhances predictive performance, and future work will focus on partitioning genes into clusters correlated with sample categories to further improve accuracy.
what is user support system???
This file will provide detailed overview about the user support system and how it will works in human computer interaction and why we need it .....
This document provides an overview of deep recommender systems and some of their shortcomings. It discusses neural network architectures like NeuMF, Wide&Deep, Neural FM, DeepFM, and DSCF that have been applied to recommendation. It also covers sequential recommendation methods, optimization techniques, and challenges like short-term rewards, manually designed architectures, isolated data, and security issues like poisoning attacks.
Applying Deep Learning with Weak and Noisy labelsDarian Frajberg
Scientific seminar at Politecnico di Milano
Como, Italy
September 2018
In recent years, Deep Learning has achieved outstanding results outperforming previous techniques and even humans, thus becoming the state-of-the-art in a wide range of tasks, among which Computer Vision has been one of the most benefited areas.
Nonetheless, most of this success is tightly coupled to strongly supervised learning tasks, which require highly accurate, expensive and labor-intensive defined ground truth labels.
In this presentation, we will introduce diverse alternatives to deal with this problem and support the training of Deep Learning models for Computer Vision tasks by simplifying the process of data labelling or exploiting the unlimited supply of publicly available data in Internet (such as user-tagged images from Flickr). Such alternatives rely on data comprising noisy and weak labels, which are much easier to collect but require special care to be used.
Botnet detection using Wgans for securityssuser3f5a831
This document proposes a GAN-based technique to improve botnet detection by addressing class imbalance issues in datasets. It involves using a WGAN to generate synthetic samples for the underrepresented botnet class, which are then combined with the original dataset and used to train and evaluate several machine learning classifiers. The proposed "Bot-WGANs" method achieves over 99.9% accuracy, 100% precision, 99.8% recall, and 99.9% F1 score on the CICIDS2017 dataset, outperforming existing methods that use oversampling techniques like SMOTE. The approach shows potential for enhancing cybersecurity by simplifying and strengthening botnet detection.
Dear Sir/Ma’am
I am interested to work as a data specialist in your organization. I believe my experience, skills and work attitude will aid your organization in a great way. Please accept my enclosed resume with this letter.
I worked at Accenture for the last four years. My key responsibilities here were to collect, analyse, store and create data. I made sure that these data were accurate and not damaged. As far as my educational background is concerned, I have a bachelor's degree in EXTC. I am excellent at solving problems and have great analytical skills. I am capable of working well with network administration and can explain the technical problems.
I would appreciate if we could meet up for an interview wherein we can discuss more on this. I can be contacted at +919493377607 or you can email me at imtiaz.khan.sw39@gmail.com
Thank You.
Yours sincerely,
Imtiaz Khan
This document discusses responsible artificial intelligence. It begins by showing survey results that most consumers want organizations to be accountable for misusing AI and want privacy protected. It then defines responsible AI as evaluating, developing, and implementing AI safely, reliably, and ethically. The main principles discussed are privacy using differential privacy, fairness by mitigating unfair impacts on groups, and transparency through explainable AI tools. General recommendations are given such as clarifying a system's purpose, considering social biases, and encouraging feedback. Benefits of responsible AI include minimizing unintentional bias and ensuring transparency.
Review of Algorithms for Crime Analysis & PredictionIRJET Journal
This document reviews algorithms that can be used for crime analysis and prediction. It discusses various data mining and machine learning techniques including classification algorithms like decision trees, k-nearest neighbors, and random forests as well as clustering algorithms like k-means clustering. Deep learning techniques are also examined for identifying relationships between different types of crimes and predicting where and when crimes may occur. The document evaluates these different algorithmic approaches and concludes that major developments in data science and machine learning now allow for effective crime analysis and prediction by discovering patterns in criminal data.
SEMI SUPERVISED BASED SPATIAL EM FRAMEWORK FOR MICROARRAY ANALYSISIRJET Journal
This document presents a semi-supervised spatial EM framework for microarray analysis to efficiently classify and predict diseases based on gene expression data. It uses a spatial EM algorithm to cluster gene expression data, followed by an SVM classifier to predict diseases and their severity levels. The proposed approach is evaluated based on classification accuracy, computation time, and ability to identify biologically significant genes. Experimental results on disease datasets show improved accuracy compared to other supervised and unsupervised methods. The authors conclude that using the same classifier for gene selection and classification enhances predictive performance, and future work will focus on partitioning genes into clusters correlated with sample categories to further improve accuracy.
NSL KDD Cup 99 dataset Anomaly Detection using Machine Learning Technique Sujeet Suryawanshi
This document summarizes a presentation given on using decision trees and machine learning techniques for anomaly detection on the NSL KDD Cup 99 dataset. It discusses anomaly detection, machine learning, different machine learning algorithms like decision trees, SVM, Naive Bayes etc. and their application for intrusion detection. It then describes an experiment conducted using the decision tree algorithm on the NSL KDD Cup 99 dataset to classify network traffic as normal or anomalous. The results showed the decision tree model achieved over 98% accuracy on both the full dataset and a reduced feature set.
This document summarizes a lecture on data preprocessing for machine learning. It discusses the importance of data preprocessing, including handling missing or noisy data. The major tasks covered are data cleaning, integration, transformation, and reduction. Specific techniques discussed include data normalization, discretization, concept hierarchy generation, and dimensionality reduction. The goal of preprocessing is to transform raw data into a cleaner format suitable for analysis by machine learning algorithms.
AI TESTING: ENSURING A GOOD DATA SPLIT BETWEEN DATA SETS (TRAINING AND TEST) ...ijsc
This document discusses challenges in effectively splitting a dataset into training and test sets for machine learning models. It proposes using k-means clustering followed by decision tree analysis to improve the split. K-means clustering groups the data points into clusters to ensure each cluster is well-represented in both the training and test sets. Then a decision tree is used to split the clustered data, aiming to maximize purity within each subset and minimize overlap between training and test data. This approach aims to capture the full domain of the dataset and avoid underrepresenting any parts of the data in either the training or test sets.
An Overview of Advesarial-attack-in-Recommender-system.pptxvudinhphuong96
The document discusses adversarial attacks against recommender systems. It begins with an introduction to recommender systems and then covers adversarial attacks, including how they can target training or inference, be targeted or non-targeted, and utilize black-box or white-box knowledge. Specifically, poisoning attacks contaminate training data while evasion attacks try to mislead trained models. The document also categorizes recommender system types like content-based, collaborative, and hybrid filtering.
Data Quality Analytics: Understanding what is in your data, before using itDomino Data Lab
Analytics and data science are ever growing fields, as business decision makers continue to use data to drive decisions. The pinnacle of these fields are the models and their accuracy/fit,; what about the data? Is your data clean, and how do you know that? Our discussion will focus on best practices for data preprocessing for analytic uses. Beginning with essential distributional checks of a dataset to a propose method for automated data validation process during ETL for transactional data.
1. The document discusses responsible AI and outlines several principles for developing AI systems responsibly, including privacy, fairness, transparency, reliability, inclusiveness, and accountability.
2. It provides examples of techniques like differential privacy and model constraints that can help mitigate privacy and fairness issues in AI systems.
3. The document also discusses the importance of transparency in AI through explainability, highlighting packages and methods for interpreting models.
Real-world Strategies for Debugging Machine Learning SystemsDatabricks
You used cross-validation, early stopping, grid search, monotonicity constraints, and regularization to train a generalizable, interpretable, and stable machine learning (ML) model. Its fit statistics look just fine on out-of-time test data, and better than the linear model it’s replacing. You selected your probability cutoff based on business goals and you even containerized your model to create a real-time scoring engine for your pals in information technology (IT). Time to deploy?
Not so fast. Current best practices for ML model training and assessment can be insufficient for high-stakes, real-world systems. Much like other complex IT systems, ML models must be debugged for logical or run-time errors and security vulnerabilities. Recent, high-profile failures have made it clear that ML models must also be debugged for disparate impact and other types of discrimination.
This presentation introduces model debugging, an emergent discipline focused on finding and fixing errors in the internal mechanisms and outputs of ML models. Model debugging attempts to test ML models like code (because they are code). It enhances trust in ML directly by increasing accuracy in new or holdout data, by decreasing or identifying hackable attack surfaces, or by decreasing discrimination. As a side-effect, model debugging should also increase the understanding and interpretability of model mechanisms and predictions.
29 SETTEMBRE 2021 – Aula Magna – Corso Duca degli Abruzzi, 24 – Politecnico di Torino
Ricerca, trasferimento tecnologico e supporto alle aziende sui temi fondamentali dei Big Data, Intelligenza Artificiale, la robotica e la rivoluzione digitale
Part of the ongoing effort with Skater for enabling better Model Interpretation for Deep Neural Network models presented at the AI Conference.
https://conferences.oreilly.com/artificial-intelligence/ai-ny/public/schedule/detail/65118
Identifying and classifying unknown Network Disruptionjagan477830
This document discusses identifying and classifying unknown network disruptions using machine learning algorithms. It begins by introducing the problem and importance of identifying network disruptions. Then it discusses related work on classifying network protocols. The document outlines the dataset and problem statement of predicting fault severity. It describes the machine learning workflow and various algorithms like random forest, decision tree and gradient boosting that are evaluated on the dataset. Finally, it concludes with achieving the objective of classifying disruptions and discusses future work like optimizing features and using neural networks.
A Novel Feature Selection with Annealing For Computer Vision And Big Data Lea...theijes
Numerous PC vision and medical imaging issues a confronted with gaining from expansive scale datasets, with a huge number of perceptions furthermore, highlights.A novel productive learning plan that fixes a sparsity imperative by continuously expelling variables taking into account a measure and a timetable. The alluring actuality that the issue size continues dropping all through the cycles makes it especially reasonable for enormous information learning. Methodology applies nonexclusively to the advancement of any differentiable misfortune capacity, and discovers applications in relapse, order and positioning. The resultant calculations assemble variable screening into estimation and are amazingly easy to execute. It gives hypothetical assurances of joining and determination consistency. Investigates genuine and engineered information demonstrate that the proposed strategy contrasts exceptionally well and other cutting edge strategies in relapse, order and positioning while being computationally exceptionally effective and adaptable.
A survey of random forest based methods forNikhil Sharma
This document summarizes a research paper that surveys random forest based methods for intrusion detection systems. It begins with an introduction describing the increasing threats to information security with growing network and data usage. It then reviews 35 papers applying random forest techniques to intrusion detection and compares their approaches. These include using random forest for classification, feature selection, and clustering. The document concludes that while random forest methods generally perform well on imbalanced data like intrusion detection, open challenges remain around high data throughput, unlabeled data, and limited benchmark datasets.
Fuzzy Rule Base System for Software Classificationijcsit
This document describes a fuzzy rule-based system for classifying Java applications using object-oriented metrics. Key features of the system include automatically extracting OO metrics from source code, a configurable set of fuzzy rules, and classifying software at both the application and class level. The system is designed to address limitations of existing OO metric tools by providing an automated, unified analysis and classification without requiring complex post-processing methods. The document outlines the system design, including subsystems for the fuzzy rules engine and extracting OO metrics, and defines membership functions and fuzzy rules for classification.
The document provides an overview of software testing methodology and discusses key topics including:
1) The purpose of testing is to catch bugs and improve productivity by reducing rework costs. Testing aims to prevent and discover bugs.
2) There are dichotomies in testing such as the differences between testing and debugging, functional vs structural testing, and designer vs tester roles. Balancing these dichotomies is an art.
3) A model for testing includes considering the environment, program, bugs, tests, and different testing levels from unit to integration. Models help design effective tests and identify unexpected results requiring changes to tests or the program.
A Survey of Image Classification with Deep Learning in the Presence of Noisy ...MonicaDommaraju
This document summarizes techniques for image classification in the presence of noisy labels. It discusses two types of noise: feature noise which corrupts data features, and label noise which incorrectly changes a data point's label. Label noise is more harmful to performance. Label noise can occur due to automated labeling systems, multiple experts with varying abilities, or intentionally injected noise. The document then categorizes techniques to address label noise into noise-model based methods, which estimate noise structure, and noise-model free methods, which aim to make classifiers robust without a noise model. Key noise-model based techniques are noisy channel, label cleansing, and sample weighting. Key noise-model free techniques use robust losses, meta-learning, regularizers,
الموعد الإثنين 03 يناير 2022
143
مبادرة
#تواصل_تطوير
المحاضرة ال 143 من المبادرة
المهندس / محمد الرافعي طرباي
نقيب المبرمجين بالدقهلية
بعنوان
"IT INDUSTRY"
How To Getting Into IT With Zero Experience
وذلك يوم الإثنين 03 يناير2022
السابعة مساء توقيت القاهرة
الثامنة مساء توقيت مكة المكرمة
و الحضور من تطبيق زووم
https://us02web.zoom.us/meeting/register/tZUpf-GsrD4jH9N9AxO39J013c1D4bqJNTcu
علما ان هناك بث مباشر للمحاضرة على القنوات الخاصة بجمعية المهندسين المصريين
ونأمل أن نوفق في تقديم ما ينفع المهندس ومهمة الهندسة في عالمنا العربي
والله الموفق
للتواصل مع إدارة المبادرة عبر قناة التليجرام
https://t.me/EEAKSA
ومتابعة المبادرة والبث المباشر عبر نوافذنا المختلفة
رابط اللينكدان والمكتبة الالكترونية
https://www.linkedin.com/company/eeaksa-egyptian-engineers-association/
رابط قناة التويتر
https://twitter.com/eeaksa
رابط قناة الفيسبوك
https://www.facebook.com/EEAKSA
رابط قناة اليوتيوب
https://www.youtube.com/user/EEAchannal
رابط التسجيل العام للمحاضرات
https://forms.gle/vVmw7L187tiATRPw9
ملحوظة : توجد شهادات حضور مجانية لمن يسجل فى رابط التقيم اخر المحاضرة
Presentation at Advanced Intelligent Systems for Sustainable Development (AISSD 2021) 20-22 August 2021 organized by the scientific research group in Egypt with Collaboration with Faculty of Computers and AI, Cairo University and the Chinese University in Egypt
IRJET- Prediction of Crime Rate Analysis using Supervised Classification Mach...IRJET Journal
This document presents a study that uses machine learning techniques to predict crime rates. Specifically, it aims to analyze crime data using supervised machine learning classification algorithms like decision trees, support vector machines, logistic regression, k-nearest neighbors, and random forests. The document outlines collecting and preprocessing crime data, selecting relevant features, training models on a portion of the data and testing them on the remaining data. It finds that random forest achieved the best prediction accuracy compared to other algorithms tested. The goal is to help law enforcement agencies better predict and reduce crime rates by analyzing historical crime data patterns.
The document presents a proposal for using community detection methods to generate hybrid partitions for multi-label classification. It introduces the limitations of global and local multi-label approaches and proposes a hybrid approach called HPML. HPML uses community detection on label co-occurrence graphs to identify correlated groups of labels and generate partitions for classification. Experiments applying HPML to 20 datasets show its partitions perform competitively with local and better than global partitions on average, demonstrating the value of exploring label correlations through community detection for multi-label classification. However, room for improvement remains as classifiers still struggle with some datasets, suggesting further research is needed on multi-label methods and evaluation.
Studying the Integration Practices and the Evolution of Ad Libraries in the G...SAIL_QU
In-app advertisements have become a major revenue for app developers in the mobile app economy. Ad libraries play an integral part in this ecosystem as app
developers integrate these libraries into their apps to display ads. However, little is known about how app developers integrate these libraries with their apps and how these libraries have evolved over time.
In this thesis, we study the ad library integration practices and the evolution of such libraries. To understand the integration practices of ad libraries, we manually study apps and derive a set of rules to automatically identify four strategies for integrating
multiple ad libraries. We observe that integrating multiple ad libraries commonly occurs in apps with a large number of downloads and ones in categories with a high percentage of apps that display ads. We also observe that app developers prefer to manage their own integrations instead of using off the shelf features of ad libraries for integrating multiple ad libraries.
To study the evolution of ad libraries, we conduct a longitudinal study of the 8 most popular ad libraries. In particular, we look at their evolution in terms of size, the main drivers for releasing a new ad library version, and their architecture. We observe that ad libraries are continuously evolving with a median release interval of 34 days. Some ad libraries have grown exponentially in size (e.g., Facebook Audience Network ad library), while other libraries have worked to reduce their size. To study the main drivers for releasing an ad library version, we manually study the release notes of the eight studied ad libraries. We observe that ad library developers continuously update their ad libraries to support a wider range of Android versions (i.e., to ensure that more devices can use the libraries without errors). Finally, we derive a reference architecture for ad libraries and study how the studied ad libraries diverged from this architecture during our study period.
Our findings can assist ad library developers to understand the challenges for developing ad libraries and the desired features of these libraries.
NSL KDD Cup 99 dataset Anomaly Detection using Machine Learning Technique Sujeet Suryawanshi
This document summarizes a presentation given on using decision trees and machine learning techniques for anomaly detection on the NSL KDD Cup 99 dataset. It discusses anomaly detection, machine learning, different machine learning algorithms like decision trees, SVM, Naive Bayes etc. and their application for intrusion detection. It then describes an experiment conducted using the decision tree algorithm on the NSL KDD Cup 99 dataset to classify network traffic as normal or anomalous. The results showed the decision tree model achieved over 98% accuracy on both the full dataset and a reduced feature set.
This document summarizes a lecture on data preprocessing for machine learning. It discusses the importance of data preprocessing, including handling missing or noisy data. The major tasks covered are data cleaning, integration, transformation, and reduction. Specific techniques discussed include data normalization, discretization, concept hierarchy generation, and dimensionality reduction. The goal of preprocessing is to transform raw data into a cleaner format suitable for analysis by machine learning algorithms.
AI TESTING: ENSURING A GOOD DATA SPLIT BETWEEN DATA SETS (TRAINING AND TEST) ...ijsc
This document discusses challenges in effectively splitting a dataset into training and test sets for machine learning models. It proposes using k-means clustering followed by decision tree analysis to improve the split. K-means clustering groups the data points into clusters to ensure each cluster is well-represented in both the training and test sets. Then a decision tree is used to split the clustered data, aiming to maximize purity within each subset and minimize overlap between training and test data. This approach aims to capture the full domain of the dataset and avoid underrepresenting any parts of the data in either the training or test sets.
An Overview of Advesarial-attack-in-Recommender-system.pptxvudinhphuong96
The document discusses adversarial attacks against recommender systems. It begins with an introduction to recommender systems and then covers adversarial attacks, including how they can target training or inference, be targeted or non-targeted, and utilize black-box or white-box knowledge. Specifically, poisoning attacks contaminate training data while evasion attacks try to mislead trained models. The document also categorizes recommender system types like content-based, collaborative, and hybrid filtering.
Data Quality Analytics: Understanding what is in your data, before using itDomino Data Lab
Analytics and data science are ever growing fields, as business decision makers continue to use data to drive decisions. The pinnacle of these fields are the models and their accuracy/fit,; what about the data? Is your data clean, and how do you know that? Our discussion will focus on best practices for data preprocessing for analytic uses. Beginning with essential distributional checks of a dataset to a propose method for automated data validation process during ETL for transactional data.
1. The document discusses responsible AI and outlines several principles for developing AI systems responsibly, including privacy, fairness, transparency, reliability, inclusiveness, and accountability.
2. It provides examples of techniques like differential privacy and model constraints that can help mitigate privacy and fairness issues in AI systems.
3. The document also discusses the importance of transparency in AI through explainability, highlighting packages and methods for interpreting models.
Real-world Strategies for Debugging Machine Learning SystemsDatabricks
You used cross-validation, early stopping, grid search, monotonicity constraints, and regularization to train a generalizable, interpretable, and stable machine learning (ML) model. Its fit statistics look just fine on out-of-time test data, and better than the linear model it’s replacing. You selected your probability cutoff based on business goals and you even containerized your model to create a real-time scoring engine for your pals in information technology (IT). Time to deploy?
Not so fast. Current best practices for ML model training and assessment can be insufficient for high-stakes, real-world systems. Much like other complex IT systems, ML models must be debugged for logical or run-time errors and security vulnerabilities. Recent, high-profile failures have made it clear that ML models must also be debugged for disparate impact and other types of discrimination.
This presentation introduces model debugging, an emergent discipline focused on finding and fixing errors in the internal mechanisms and outputs of ML models. Model debugging attempts to test ML models like code (because they are code). It enhances trust in ML directly by increasing accuracy in new or holdout data, by decreasing or identifying hackable attack surfaces, or by decreasing discrimination. As a side-effect, model debugging should also increase the understanding and interpretability of model mechanisms and predictions.
29 SETTEMBRE 2021 – Aula Magna – Corso Duca degli Abruzzi, 24 – Politecnico di Torino
Ricerca, trasferimento tecnologico e supporto alle aziende sui temi fondamentali dei Big Data, Intelligenza Artificiale, la robotica e la rivoluzione digitale
Part of the ongoing effort with Skater for enabling better Model Interpretation for Deep Neural Network models presented at the AI Conference.
https://conferences.oreilly.com/artificial-intelligence/ai-ny/public/schedule/detail/65118
Identifying and classifying unknown Network Disruptionjagan477830
This document discusses identifying and classifying unknown network disruptions using machine learning algorithms. It begins by introducing the problem and importance of identifying network disruptions. Then it discusses related work on classifying network protocols. The document outlines the dataset and problem statement of predicting fault severity. It describes the machine learning workflow and various algorithms like random forest, decision tree and gradient boosting that are evaluated on the dataset. Finally, it concludes with achieving the objective of classifying disruptions and discusses future work like optimizing features and using neural networks.
A Novel Feature Selection with Annealing For Computer Vision And Big Data Lea...theijes
Numerous PC vision and medical imaging issues a confronted with gaining from expansive scale datasets, with a huge number of perceptions furthermore, highlights.A novel productive learning plan that fixes a sparsity imperative by continuously expelling variables taking into account a measure and a timetable. The alluring actuality that the issue size continues dropping all through the cycles makes it especially reasonable for enormous information learning. Methodology applies nonexclusively to the advancement of any differentiable misfortune capacity, and discovers applications in relapse, order and positioning. The resultant calculations assemble variable screening into estimation and are amazingly easy to execute. It gives hypothetical assurances of joining and determination consistency. Investigates genuine and engineered information demonstrate that the proposed strategy contrasts exceptionally well and other cutting edge strategies in relapse, order and positioning while being computationally exceptionally effective and adaptable.
A survey of random forest based methods forNikhil Sharma
This document summarizes a research paper that surveys random forest based methods for intrusion detection systems. It begins with an introduction describing the increasing threats to information security with growing network and data usage. It then reviews 35 papers applying random forest techniques to intrusion detection and compares their approaches. These include using random forest for classification, feature selection, and clustering. The document concludes that while random forest methods generally perform well on imbalanced data like intrusion detection, open challenges remain around high data throughput, unlabeled data, and limited benchmark datasets.
Fuzzy Rule Base System for Software Classificationijcsit
This document describes a fuzzy rule-based system for classifying Java applications using object-oriented metrics. Key features of the system include automatically extracting OO metrics from source code, a configurable set of fuzzy rules, and classifying software at both the application and class level. The system is designed to address limitations of existing OO metric tools by providing an automated, unified analysis and classification without requiring complex post-processing methods. The document outlines the system design, including subsystems for the fuzzy rules engine and extracting OO metrics, and defines membership functions and fuzzy rules for classification.
The document provides an overview of software testing methodology and discusses key topics including:
1) The purpose of testing is to catch bugs and improve productivity by reducing rework costs. Testing aims to prevent and discover bugs.
2) There are dichotomies in testing such as the differences between testing and debugging, functional vs structural testing, and designer vs tester roles. Balancing these dichotomies is an art.
3) A model for testing includes considering the environment, program, bugs, tests, and different testing levels from unit to integration. Models help design effective tests and identify unexpected results requiring changes to tests or the program.
A Survey of Image Classification with Deep Learning in the Presence of Noisy ...MonicaDommaraju
This document summarizes techniques for image classification in the presence of noisy labels. It discusses two types of noise: feature noise which corrupts data features, and label noise which incorrectly changes a data point's label. Label noise is more harmful to performance. Label noise can occur due to automated labeling systems, multiple experts with varying abilities, or intentionally injected noise. The document then categorizes techniques to address label noise into noise-model based methods, which estimate noise structure, and noise-model free methods, which aim to make classifiers robust without a noise model. Key noise-model based techniques are noisy channel, label cleansing, and sample weighting. Key noise-model free techniques use robust losses, meta-learning, regularizers,
الموعد الإثنين 03 يناير 2022
143
مبادرة
#تواصل_تطوير
المحاضرة ال 143 من المبادرة
المهندس / محمد الرافعي طرباي
نقيب المبرمجين بالدقهلية
بعنوان
"IT INDUSTRY"
How To Getting Into IT With Zero Experience
وذلك يوم الإثنين 03 يناير2022
السابعة مساء توقيت القاهرة
الثامنة مساء توقيت مكة المكرمة
و الحضور من تطبيق زووم
https://us02web.zoom.us/meeting/register/tZUpf-GsrD4jH9N9AxO39J013c1D4bqJNTcu
علما ان هناك بث مباشر للمحاضرة على القنوات الخاصة بجمعية المهندسين المصريين
ونأمل أن نوفق في تقديم ما ينفع المهندس ومهمة الهندسة في عالمنا العربي
والله الموفق
للتواصل مع إدارة المبادرة عبر قناة التليجرام
https://t.me/EEAKSA
ومتابعة المبادرة والبث المباشر عبر نوافذنا المختلفة
رابط اللينكدان والمكتبة الالكترونية
https://www.linkedin.com/company/eeaksa-egyptian-engineers-association/
رابط قناة التويتر
https://twitter.com/eeaksa
رابط قناة الفيسبوك
https://www.facebook.com/EEAKSA
رابط قناة اليوتيوب
https://www.youtube.com/user/EEAchannal
رابط التسجيل العام للمحاضرات
https://forms.gle/vVmw7L187tiATRPw9
ملحوظة : توجد شهادات حضور مجانية لمن يسجل فى رابط التقيم اخر المحاضرة
Presentation at Advanced Intelligent Systems for Sustainable Development (AISSD 2021) 20-22 August 2021 organized by the scientific research group in Egypt with Collaboration with Faculty of Computers and AI, Cairo University and the Chinese University in Egypt
IRJET- Prediction of Crime Rate Analysis using Supervised Classification Mach...IRJET Journal
This document presents a study that uses machine learning techniques to predict crime rates. Specifically, it aims to analyze crime data using supervised machine learning classification algorithms like decision trees, support vector machines, logistic regression, k-nearest neighbors, and random forests. The document outlines collecting and preprocessing crime data, selecting relevant features, training models on a portion of the data and testing them on the remaining data. It finds that random forest achieved the best prediction accuracy compared to other algorithms tested. The goal is to help law enforcement agencies better predict and reduce crime rates by analyzing historical crime data patterns.
The document presents a proposal for using community detection methods to generate hybrid partitions for multi-label classification. It introduces the limitations of global and local multi-label approaches and proposes a hybrid approach called HPML. HPML uses community detection on label co-occurrence graphs to identify correlated groups of labels and generate partitions for classification. Experiments applying HPML to 20 datasets show its partitions perform competitively with local and better than global partitions on average, demonstrating the value of exploring label correlations through community detection for multi-label classification. However, room for improvement remains as classifiers still struggle with some datasets, suggesting further research is needed on multi-label methods and evaluation.
Similar to On the Unreliability of Bug Severity Data (20)
Studying the Integration Practices and the Evolution of Ad Libraries in the G...SAIL_QU
In-app advertisements have become a major revenue for app developers in the mobile app economy. Ad libraries play an integral part in this ecosystem as app
developers integrate these libraries into their apps to display ads. However, little is known about how app developers integrate these libraries with their apps and how these libraries have evolved over time.
In this thesis, we study the ad library integration practices and the evolution of such libraries. To understand the integration practices of ad libraries, we manually study apps and derive a set of rules to automatically identify four strategies for integrating
multiple ad libraries. We observe that integrating multiple ad libraries commonly occurs in apps with a large number of downloads and ones in categories with a high percentage of apps that display ads. We also observe that app developers prefer to manage their own integrations instead of using off the shelf features of ad libraries for integrating multiple ad libraries.
To study the evolution of ad libraries, we conduct a longitudinal study of the 8 most popular ad libraries. In particular, we look at their evolution in terms of size, the main drivers for releasing a new ad library version, and their architecture. We observe that ad libraries are continuously evolving with a median release interval of 34 days. Some ad libraries have grown exponentially in size (e.g., Facebook Audience Network ad library), while other libraries have worked to reduce their size. To study the main drivers for releasing an ad library version, we manually study the release notes of the eight studied ad libraries. We observe that ad library developers continuously update their ad libraries to support a wider range of Android versions (i.e., to ensure that more devices can use the libraries without errors). Finally, we derive a reference architecture for ad libraries and study how the studied ad libraries diverged from this architecture during our study period.
Our findings can assist ad library developers to understand the challenges for developing ad libraries and the desired features of these libraries.
Improving the testing efficiency of selenium-based load testsSAIL_QU
Slides for a paper published at AST 2019:
Shahnaz M. Shariff, Heng Li, Cor-Paul Bezemer, Ahmed E. Hassan, Thanh H. D. Nguyen, and Parminder Flora. 2019. Improving the testing efficiency of selenium-based load tests. In Proceedings of the 14th International Workshop on Automation of Software Test (AST '19). IEEE Press, Piscataway, NJ, USA, 14-20. DOI: https://doi.org/10.1109/AST.2019.00008
Studying User-Developer Interactions Through the Distribution and Reviewing M...SAIL_QU
This document discusses studying user-developer interactions through the distribution and reviewing mechanisms of the Google Play Store. It analyzes emergency updates made by developers to fix issues, the dialogue between users and developers through reviews and responses, and how the reviewing mechanism can help identify good and bad updates. The study found that responding to reviews is six times more likely to increase an app's rating, with 84% of rating increases going to four or five stars. Three common patterns of developer responses were identified: responding to negative or long reviews, only negative reviews, and reviews shortly after an update.
Studying online distribution platforms for games through the mining of data f...SAIL_QU
Our studies of Steam platform data provided insights into online game distribution:
1) Urgent game updates were used to fix crashes, balance issues, and functionality; frequent updaters released more 0-day patches.
2) The Early Access model attracted indie developers and increased game participation; reviews were more positive during Early Access.
3) Game reviews were typically short and in English; sales increased review volume more than new updates; negative reviews came after longer play.
Understanding the Factors for Fast Answers in Technical Q&A Websites: An Empi...SAIL_QU
This study analyzed factors that impact the speed of questions receiving accepted answers on four popular Stack Exchange websites: Stack Overflow, Mathematics, Ask Ubuntu, and Super User. The researchers examined question, answerer, asker, and answer factors from over 150,000 questions. They built classification models and found that key factors for fast answers included the past speed of answerers, length of the question, and past speed of answers for the question's tags. The models achieved AUCs of 0.85-0.95. Fast answers relied heavily on answerers, especially frequent answerers. The study suggests improving incentives for non-frequent and more difficult questions to attract diverse answerers.
Investigating the Challenges in Selenium Usage and Improving the Testing Effi...SAIL_QU
Selenium is a popular tool for browser-based automation testing. The author analyzes challenges in using Selenium by mining Selenium questions on Stack Overflow. Programming language-related questions, especially for Java and Python, are most common and growing fastest. Less than half of questions receive accepted answers, and questions about browsers and components take longest. In the second part, the author develops an approach to improve efficiency of Selenium-based load testing by sharing browsers among user instances. This increases the number of error-free users by 20-22% while reducing memory usage.
Mining Development Knowledge to Understand and Support Software Logging Pract...SAIL_QU
This document summarizes Heng Li's PhD thesis on mining development knowledge to understand and support software logging practices. It discusses how logging code is used to record runtime information but can be difficult for developers to maintain. The thesis aims to understand current logging practices and develop tools by mining change history, source code, issue reports, and other development knowledge. It presents research that analyzes logging-related issues to identify developers' logging concerns, uses code topics and structure to predict where logging statements should be added, leverages code changes to suggest when logging code needs updating, and applies machine learning models to recommend appropriate log levels.
Which Log Level Should Developers Choose For a New Logging Statement?SAIL_QU
The document discusses choosing an appropriate log level when adding a new logging statement. It finds that an ordinal regression model can effectively model log levels, achieving an AUC of 0.76-0.81 in within-project evaluation and 0.71-0.8 in cross-project evaluation. The most influential factors for determining log levels vary between projects and include metrics related to the logging statement, containing code block, and file as well as code change and historical change metrics.
Towards Just-in-Time Suggestions for Log ChangesSAIL_QU
The document presents a study on providing just-in-time suggestions for log changes when developers make code changes. The researchers analyzed over 32,000 log changes from 4 systems. They found 20 reasons for log changes that fall into 4 categories: block changes, log improvements, dependence-driven changes, and logging issues. A random forest classifier using 25 software metrics related to code changes, history, and complexity achieved 0.84-0.91 AUC in predicting whether a log change is needed. Change metrics and product metrics were the most influential factors. The study aims to help developers make better logging decisions for failure diagnosis.
The Impact of Task Granularity on Co-evolution AnalysesSAIL_QU
The document discusses how task granularity at different levels (e.g. commits, pull requests, work items) can impact analyses of co-evolution in software projects. It finds that analyzing at the commit-level can overlook relationships between tasks that span multiple commits. Work item level analysis is recommended to provide a more complete view of co-evolution, as median of 29% of work items consist of multiple commits, and analyzing at the commit level would miss 24% of co-changed files and inability to group 83% of related commits.
How are Discussions Associated with Bug Reworking? An Empirical Study on Open...SAIL_QU
1) Initial bug fix discussions with more comments and more developers participating are more likely to experience later bug reworking through re-opening or re-patching of the bug.
2) Manual analysis found that defective initial fixes and failure to reach consensus in discussions contributed to later reworking.
3) For re-opened bugs, initial discussions focused on addressing a particular problem through a burst of comments, while re-patched bugs lacked thorough code review and testing during the initial fix period.
A Study of the Relation of Mobile Device Attributes with the User-Perceived Q...SAIL_QU
This study examined the relationship between mobile device attributes and user-perceived quality of Android apps. The researchers analyzed 150,373 star ratings from Google Play across 30 devices and 280 apps. They found that the perceived quality of apps varies across devices, and having better characteristics of an attribute does not necessarily correlate with higher quality. Device OS version, resolution, and CPU showed significant relationships with ratings, as did some app attributes like lines of code and number of inputs. However, some device attributes had stronger relationships than app attributes.
A Large-Scale Study of the Impact of Feature Selection Techniques on Defect C...SAIL_QU
This document presents the results of a large-scale study on the impact of feature selection techniques on defect classification models. The study used expanded scopes including multiple datasets from NASA and PROMISE with different feature types, more classification techniques from different paradigms, and additional feature selection techniques. The results show that correlation-based feature subset selection techniques like FS1 and FS2 consistently appear in the top ranks across most of the datasets, projects within the datasets, and classification techniques. The document concludes that future defect classification studies should consider applying correlation-based feature selection techniques.
Studying the Dialogue Between Users and Developers of Free Apps in the Google...SAIL_QU
The study analyzes user-developer interactions through reviews and responses on the Google Play Store. It finds that responding to reviews has a significant positive impact, with 84% of rating increases due to the developer addressing the issue or providing guidance. Three common response patterns were identified: only negative reviews, negative or longer reviews, and reviews shortly after an update. Developers most often thank the user, ask for details, provide guidance, or ask for an endorsement. Guidance responses can address common issues through FAQs. The analysis considered over 2,000 apps, 355,000 review changes, 128,000 responses, and 4 million reviews.
What Do Programmers Know about Software Energy Consumption?SAIL_QU
This document summarizes the results of a survey of 122 programmers about their knowledge of software energy consumption. The survey found that programmers have limited awareness of energy consumption and how to reduce it. They were unaware of the main causes of high energy usage. Programmers lacked knowledge about how to properly rank the energy consumption of different hardware components and were unfamiliar with strategies to improve efficiency, such as minimizing I/O and avoiding polling. The study concludes that programmers would benefit from more education on software energy usage and its causes.
Revisiting the Experimental Design Choices for Approaches for the Automated R...SAIL_QU
Prior research on automated duplicate issue report retrieval focused on improving performance metrics like recall rate. The author revisits experimental design choices from four perspectives: needed effort, data changes, data filtration, and evaluation process.
The thesis contributions are: 1) Showing the importance of considering needed effort in performance measurement. 2) Proposing a "realistic evaluation" approach and analyzing prior findings with it. 3) Developing a genetic algorithm to filter old issue reports and improve performance. 4) Highlighting the impact of "just-in-time" features on evaluation. The findings help better understand benefits and limitations of prior work in this area.
Measuring Program Comprehension: A Large-Scale Field Study with ProfessionalsSAIL_QU
The document summarizes a large-scale field study that tracked the program comprehension activities of 78 professional developers over 3,148 hours. The study found that:
1) Program comprehension accounted for approximately 58% of developers' time on average, with navigation and editing making up the remaining portions.
2) Developers frequently used web browsers and document editors to aid comprehension beyond just IDEs.
3) Interviews and observations revealed that insufficient documentation, unclear code, and complex inheritance hierarchies contributed to long comprehension sessions.
WWDC 2024 Keynote Review: For CocoaCoders AustinPatrick Weigel
Overview of WWDC 2024 Keynote Address.
Covers: Apple Intelligence, iOS18, macOS Sequoia, iPadOS, watchOS, visionOS, and Apple TV+.
Understandable dialogue on Apple TV+
On-device app controlling AI.
Access to ChatGPT with a guest appearance by Chief Data Thief Sam Altman!
App Locking! iPhone Mirroring! And a Calculator!!
E-Invoicing Implementation: A Step-by-Step Guide for Saudi Arabian CompaniesQuickdice ERP
Explore the seamless transition to e-invoicing with this comprehensive guide tailored for Saudi Arabian businesses. Navigate the process effortlessly with step-by-step instructions designed to streamline implementation and enhance efficiency.
The Rising Future of CPaaS in the Middle East 2024Yara Milbes
Explore "The Rising Future of CPaaS in the Middle East in 2024" with this comprehensive PPT presentation. Discover how Communication Platforms as a Service (CPaaS) is transforming communication across various sectors in the Middle East.
The Key to Digital Success_ A Comprehensive Guide to Continuous Testing Integ...kalichargn70th171
In today's business landscape, digital integration is ubiquitous, demanding swift innovation as a necessity rather than a luxury. In a fiercely competitive market with heightened customer expectations, the timely launch of flawless digital products is crucial for both acquisition and retention—any delay risks ceding market share to competitors.
Everything You Need to Know About X-Sign: The eSign Functionality of XfilesPr...XfilesPro
Wondering how X-Sign gained popularity in a quick time span? This eSign functionality of XfilesPro DocuPrime has many advancements to offer for Salesforce users. Explore them now!
14 th Edition of International conference on computer visionShulagnaSarkar2
About the event
14th Edition of International conference on computer vision
Computer conferences organized by ScienceFather group. ScienceFather takes the privilege to invite speakers participants students delegates and exhibitors from across the globe to its International Conference on computer conferences to be held in the Various Beautiful cites of the world. computer conferences are a discussion of common Inventions-related issues and additionally trade information share proof thoughts and insight into advanced developments in the science inventions service system. New technology may create many materials and devices with a vast range of applications such as in Science medicine electronics biomaterials energy production and consumer products.
Nomination are Open!! Don't Miss it
Visit: computer.scifat.com
Award Nomination: https://x-i.me/ishnom
Conference Submission: https://x-i.me/anicon
For Enquiry: Computer@scifat.com
UI5con 2024 - Bring Your Own Design SystemPeter Muessig
How do you combine the OpenUI5/SAPUI5 programming model with a design system that makes its controls available as Web Components? Since OpenUI5/SAPUI5 1.120, the framework supports the integration of any Web Components. This makes it possible, for example, to natively embed own Web Components of your design system which are created with Stencil. The integration embeds the Web Components in a way that they can be used naturally in XMLViews, like with standard UI5 controls, and can be bound with data binding. Learn how you can also make use of the Web Components base class in OpenUI5/SAPUI5 to also integrate your Web Components and get inspired by the solution to generate a custom UI5 library providing the Web Components control wrappers for the native ones.
Odoo releases a new update every year. The latest version, Odoo 17, came out in October 2023. It brought many improvements to the user interface and user experience, along with new features in modules like accounting, marketing, manufacturing, websites, and more.
The Odoo 17 update has been a hot topic among startups, mid-sized businesses, large enterprises, and Odoo developers aiming to grow their businesses. Since it is now already the first quarter of 2024, you must have a clear idea of what Odoo 17 entails and what it can offer your business if you are still not aware of it.
This blog covers the features and functionalities. Explore the entire blog and get in touch with expert Odoo ERP consultants to leverage Odoo 17 and its features for your business too.
An Overview of Odoo ERP
Odoo ERP was first released as OpenERP software in February 2005. It is a suite of business applications used for ERP, CRM, eCommerce, websites, and project management. Ten years ago, the Odoo Enterprise edition was launched to help fund the Odoo Community version.
When you compare Odoo Community and Enterprise, the Enterprise edition offers exclusive features like mobile app access, Odoo Studio customisation, Odoo hosting, and unlimited functional support.
Today, Odoo is a well-known name used by companies of all sizes across various industries, including manufacturing, retail, accounting, marketing, healthcare, IT consulting, and R&D.
The latest version, Odoo 17, has been available since October 2023. Key highlights of this update include:
Enhanced user experience with improvements to the command bar, faster backend page loading, and multiple dashboard views.
Instant report generation, credit limit alerts for sales and invoices, separate OCR settings for invoice creation, and an auto-complete feature for forms in the accounting module.
Improved image handling and global attribute changes for mailing lists in email marketing.
A default auto-signature option and a refuse-to-sign option in HR modules.
Options to divide and merge manufacturing orders, track the status of manufacturing orders, and more in the MRP module.
Dark mode in Odoo 17.
Now that the Odoo 17 announcement is official, let’s look at what’s new in Odoo 17!
What is Odoo ERP 17?
Odoo 17 is the latest version of one of the world’s leading open-source enterprise ERPs. This version has come up with significant improvements explained here in this blog. Also, this new version aims to introduce features that enhance time-saving, efficiency, and productivity for users across various organisations.
Odoo 17, released at the Odoo Experience 2023, brought notable improvements to the user interface and added new functionalities with enhancements in performance, accessibility, data analysis, and management, further expanding its reach in the market.
Most important New features of Oracle 23c for DBAs and Developers. You can get more idea from my youtube channel video from https://youtu.be/XvL5WtaC20A
8 Best Automated Android App Testing Tool and Framework in 2024.pdfkalichargn70th171
Regarding mobile operating systems, two major players dominate our thoughts: Android and iPhone. With Android leading the market, software development companies are focused on delivering apps compatible with this OS. Ensuring an app's functionality across various Android devices, OS versions, and hardware specifications is critical, making Android app testing essential.
INTRODUCTION TO AI CLASSICAL THEORY TARGETED EXAMPLESanfaltahir1010
Image: Include an image that represents the concept of precision, such as a AI helix or a futuristic healthcare
setting.
Objective: Provide a foundational understanding of precision medicine and its departure from traditional
approaches
Role of theory: Discuss how genomics, the study of an organism's complete set of AI ,
plays a crucial role in precision medicine.
Customizing treatment plans: Highlight how genetic information is used to customize
treatment plans based on an individual's genetic makeup.
Examples: Provide real-world examples of successful application of AI such as genetic
therapies or targeted treatments.
Importance of molecular diagnostics: Explain the role of molecular diagnostics in identifying
molecular and genetic markers associated with diseases.
Biomarker testing: Showcase how biomarker testing aids in creating personalized treatment plans.
Content:
• Ethical issues: Examine ethical concerns related to precision medicine, such as privacy, consent, and
potential misuse of genetic information.
• Regulations and guidelines: Present examples of ethical guidelines and regulations in place to safeguard
patient rights.
• Visuals: Include images or icons representing ethical considerations.
Content:
• Ethical issues: Examine ethical concerns related to precision medicine, such as privacy, consent, and
potential misuse of genetic information.
• Regulations and guidelines: Present examples of ethical guidelines and regulations in place to safeguard
patient rights.
• Visuals: Include images or icons representing ethical considerations.
Content:
• Ethical issues: Examine ethical concerns related to precision medicine, such as privacy, consent, and
potential misuse of genetic information.
• Regulations and guidelines: Present examples of ethical guidelines and regulations in place to safeguard
patient rights.
• Visuals: Include images or icons representing ethical considerations.
Real-world case study: Present a detailed case study showcasing the success of precision
medicine in a specific medical scenario.
Patient's journey: Discuss the patient's journey, treatment plan, and outcomes.
Impact: Emphasize the transformative effect of precision medicine on the individual's
health.
Objective: Ground the presentation in a real-world example, highlighting the practical
application and success of precision medicine.
Data challenges: Address the challenges associated with managing large sets of patient data in precision
medicine.
Technological solutions: Discuss technological innovations and solutions for handling and analyzing vast
datasets.
Visuals: Include graphics representing data management challenges and technological solutions.
Objective: Acknowledge the data-related challenges in precision medicine and highlight innovative solutions.
Data challenges: Address the challenges associated with managing large sets of patient data in precision
medicine.
Technological solutions: Discuss technological innovations and solutions
Measures in SQL (SIGMOD 2024, Santiago, Chile)Julian Hyde
SQL has attained widespread adoption, but Business Intelligence tools still use their own higher level languages based upon a multidimensional paradigm. Composable calculations are what is missing from SQL, and we propose a new kind of column, called a measure, that attaches a calculation to a table. Like regular tables, tables with measures are composable and closed when used in queries.
SQL-with-measures has the power, conciseness and reusability of multidimensional languages but retains SQL semantics. Measure invocations can be expanded in place to simple, clear SQL.
To define the evaluation semantics for measures, we introduce context-sensitive expressions (a way to evaluate multidimensional expressions that is consistent with existing SQL semantics), a concept called evaluation context, and several operations for setting and modifying the evaluation context.
A talk at SIGMOD, June 9–15, 2024, Santiago, Chile
Authors: Julian Hyde (Google) and John Fremlin (Google)
https://doi.org/10.1145/3626246.3653374
1. SMU Classification: Restricted
Strategic Partner:
On the Unreliability of Bug Severity Data
Yuan TIAN
Data Scientist at Living Analytics Research Centre,
Singapore Management University
ytian@smu.edu.sg
April 18th, 2018 @ Queen’s University, Canada
2. SMU Classification: Restricted
2
Supervised Machine Learning models rely heavily on labels
Labels
Predictive
Model
Feature
Extraction
Learning
Expected Label
Training
Data
New
Data
Cat
Not Cat
3. SMU Classification: Restricted
3
Traditional machine learning models suffer from noisy labels
in training set
Noisy labels due to:
• Human mistakes.
• Non-expert generated labels.
• Machine generated labels.
• Communication or encoding
problems.
4. SMU Classification: Restricted
4
Noisy labels due to:
• Human mistakes.
• Non-expert generated labels.
• Machine generated labels.
• Communication or encoding
problems.
Decreased
Performance
Traditional machine learning models suffer from noisy labels
in training set
Noise Level=3%
F1=0.83
F1=0.4
5. SMU Classification: Restricted
5
Noisy labels due to:
• Human mistakes.
• Non-expert generated labels.
• Machine generated labels.
• Communication or encoding
problems.
Decreased
Performance
More Complex
Model
Unreliable
Performance
Measures
Incorrect
Influential
Features
Traditional machine learning models suffer from noisy labels
in training set
6. SMU Classification: Restricted
6
Decreased
Performance
More Complex
Model
Unreliable
Performance
Measures
Incorrect
Influential
Features
Traditional machine learning models suffer from noisy labels
in training set
Assume that absolute ground truth
for labels exists although the labels
may be noisy for some reasons.
7. SMU Classification: Restricted
7
Medical Diagnosis Malware Detection
Classification is in some cases subjective, which results in
inter-labeller variability
Image Tagging
“inconsistent
labels” noise
8. SMU Classification: Restricted
8
“Inconsistent Labels” Noise
Classification is in some cases subjective, which results in
inter-labeller variability (cont.)
How interesting is this book ?
How informative is this tweet ?
Is it a high quality product ?
How severe is the problem ?
User created tags for images, content, etc.
Job titles created by different companies
Numeric Label
Categorical Label
9. SMU Classification: Restricted
9
Before collecting labels
Readily available data
• Shared labelling criteria
• Multiple labelers
• Repeated labelling
• Labels are agreed by
all/majority
• Pairwise comparison
• Averaging
• Majority voting
• Consensus voting
• Remove outliers
• Learning with
uncertain labels
?
How to measure
inconsistency?
Multiple labelers
Single labeler
How people cope with label noise (“inconsistent labels”)
caused by the subjective labelling, etc.
How to cope with
inconsistency?
10. SMU Classification: Restricted
10
Community Intelligence
How people cope with label noise (“inconsistent labels”)
caused by the subjective labelling, etc.
How to measure
inconsistency?
How to cope with
inconsistency?
11. SMU Classification: Restricted
11
Part 1:
│Noisy labels negatively impact learning
│Overview of approaches for coping with inconsistent labels
│Sample : Bug severity levels
Part 2:
│Future research direction: Big Data to Thick (high quality) Data
Talk Outline
14. SMU Classification: Restricted
14
Prior work in automated bug severity labelling
Existing approaches assume that all assigned
severity levels are consistent.
Classifiers are evaluated using Accuracy &
F-measure, AUC.
15. SMU Classification: Restricted
15
How to measure label
inconsistency?
How to evaluate machine
learning models with
inconsistent labels?
human-machine
_____________
human-human
Challenges and our solutions
16. SMU Classification: Restricted
16
Bug Detection &
Reporting
Bug Triaging
#Validity Check
# Duplicate Bug Detection
# Bug Prioritization
# Bug Assignment
Debugging &
Bug Fixing
Duplicate bugs should
have the same severity
levels, if “severity level”
labels are consistent.
How to measure the inconsistency?
17. SMU Classification: Restricted
17
Blocker Critical
Major
Minor
Inconsistent Duplicate Buckets
1
2
3
Clean Duplicate Buckets
Manual verify 1,394 bug reports
(statistically representative
sample)
95% of the inconsistent
buckets are reporting the
same bug.
Up to 51% of human-assigned bug severity labels are
inconsistent !
18. SMU Classification: Restricted
18
human-machine
_____________
human-human
Krippendorff's alpha
A new evaluation measure for machine learning tasks with
inconsistent labels:
𝛼 = 1 −
𝐷𝑜
𝐷𝑒
Observed disagreements
Expected disagreements
when the bug severity
levels are randomly
assigned.
𝛼 = 1 regarded as perfect agreement.
Benefit of Alpha:
• Allow multiple labellers
• Good for ordinal labels
• Factoring class distributions
• Less biased to number of
labels and the number of
coders
19. SMU Classification: Restricted
19
Bug Report 1 2 3 4 5 6 7 8 9 10
Human 1 2 3 3 4 3 3 4 3 5
Machine A 2 3 3 3 3 3 3 5 3 2
Machine B 3 4 3 3 3 3 3 5 3 2
Machine C 3 3 3 3 3 3 3 3 3 3
Krippendorff’s Alpha Vs Accuracy
Accuracy: Machine A = Machine B = Machine C
Alpha: Machine A > Machine B > Machine C
21. SMU Classification: Restricted
21
Community intelligence can be used to
identify and quantify the inconsistency
of subjective labels.
Performance of machine learning
models should be measured within
context, e.g., relative to human inter-
agreement.
human-machine
_____________
human-human
Take away messages
22. SMU Classification: Restricted
22
During 2017, every minute of the day:
4M tweets
36 M google
searches
600 page edits
0.4 M trips
120 new
professionals $258,751.90 in sales
Big Data is everywhere, however…
Bad data is costing organizations some $3.1 trillion a year in
the US alone.
83% of companies said their revenue is affected by inaccurate
and incomplete customer or prospect data.
25. SMU Classification: Restricted
25
Big
Data
Thick Data
Size of
Data
Quality
Big
Data
Size of
Data
Quality
Challenges:
1. Impossible to specify all the data semantics beforehand.
2. Manual labelling of noise is expensive and time
consuming, impossible to scale.
3. Lack of quality metrics for big data.
4. Lack of performance measures for noisy data.
Make big data “ thick ”
26. SMU Classification: Restricted
26
• Integrate existing internal/external resources collectively
created by relevant communities.
• Utilize knowledge in unstructured data.
How to make big data thick in a lightweight cost-effective way?
27. SMU Classification: Restricted
27
Assessing Quality of
Big Data
Lightweight Scalable Noise
Reduction/Correction
Techniques
New Noise Tolerant Learning
Algorithm
for Big Data
New Performance Measures
for Noisy Data
Future Roadmap: Big Data to Thick Data
28. SMU Classification: Restricted
28
human-machine
_____________
human-human
• New measure for machine learning
performance with inconsistent labels
Challenges: Big Data to Thick Data
• Need metrics for quality assessment and
model evaluation
• Need lightweight noise reduction methods
Conclusion ytian@smu.edu.sg
• Classification is in some cases subjective,
which results in inter-labeler variability.
30. SMU Classification: Restricted
30
Bucket/
Bug Report
1 2 3 4 5 6 7 8 9 10
1 2 3 3 4 3 3 4 3 5
2 3 3 3 3 3 3 5 3 2
#Raters 2 2 2 2 2 2 2 2 2 2
Count Matrix
Derived from
Ratings
Predefined
Distance Matrix
This is why alpha is
good for ordinal labels!
(c,k) represents a
pair of ratings
Computation of Krippendorff's alpha
𝛼 = 0.27
31. SMU Classification: Restricted
31
How do inconsistent labels affect varies machine learning
models?
Duplicate Bug
Reports
Clean Bug Reports Inconsistent Bug Reports
Test Bug
Reports
Train (Inconsistent Data Ratio:0%)
Inconsistent Data Ratio:20%
32. SMU Classification: Restricted
32
Noise injection leads to drop in alpha for all datasets
(Mozilla)
(OpenOffice)
Decision Tree
Naïve Bayes Multinomial
Support Vector Machine
REP + k-Nearest Neighbor
Ratio of Inconsistent Bug Reports in Training Set
(Eclipse)
Editor's Notes
Thanks prof. z for the introduction, thanks all for taking time to attend this talk, I am honored to be invited here.
As you can see from the title, I’d like to share some of my experience in dealing with inconsistent labels, which is important for making your later analysis on the data reliable and effective.
Please feel free to interrupt if you find any difficulties regarding the slide.
To begin with, I would like to introduce supervised machine learning models, which are probability the most common machine learning models used nowadays. And they are also the ones that are affected the most by the quality of labels.
In the slide, we see a general flow of supervised machine learning, it takes labelled training data as input, gleaning information from it, and eventually learn a model that can label new data.
Let’s me give you a simple example here, at the top of the slide, you can see 7 images, each of them is associated with a label 1/0 indicating whether there is a cat in the image. Supervised machine learning takes these data-label pairs as training input, and then extracts features from the images. In traditional machine learning flow, features are defined manually, while in the latest deep learning techniques, features are learned from the data. After the feature extraction process, a model is learned for a mapping between feature values and labels, so that given the new image shown at the right bottom of the slide, we hope that the learned model is able to identify that there is a cat in the image.
Since supervised machine learning models are popular, and rely heavily on labels, intuitively, we machine learning practitioners want clean labels in our training data.
However, things don’t always go as we wish. Have you noticed that just among the 7 images we saw in the previous slide, the image cycled in red contains a dog, instead of a cat, in this case, we say we encounter an incorrect label.
In fact, real-world data often contain noisy labels due to various reasons. For example, we human make mistakes, including experts. secondly, as collecting reliable labels is a expensive and time costly task, many studies leverage crowdsourcing platforms to collect non-expert labels in a cheap and fast way, and some regard machine generated labels as the ground truth. Last, noisy labels might simply due to communication or encoding problems.
In fact, people have studied the impact of noisy labels for a long time in the machine learning areas, theoretically or empirically demonstrate that noisy labels can bring negative consequences for learning. The most important two includes descried performance and unreliable performance measures. The image on the right side shows an study regarding performance of traditional classifiers on a particular task, we could see that the performance of the classifiers all drop dramatically after noise level is greater than 3%.
Noisy labels often lead to more complex model, and incorrect influential features.
http://www.stat.purdue.edu/~jianzhan/papers/sigir03zhang.pdf
In fact, people have studied the impact of noisy labels for a long time in the machine learning areas, theoretically or empirically demonstrate that noisy labels can bring negative consequences for learning. The most important two includes descried performance and unreliable performance measures. The image on the right side shows an study regarding performance of traditional classifiers on a particular task, we could see that the performance of the classifiers all drop dramatically after noise level is greater than 3%.
Noisy labels often lead to more complex model, and incorrect influential features.
So the message I want to deliver here is, we should take care of noisy label when we design machine learning models.
In most of the research on analysing noisy labels, we have an important assumption that absolute ground truth for label exist, like we could easily tell that the cycled image is wrongly labelled.
However, classification can be subjective, where the ground truth is hard to tell.
For instance, two doctors may give different diagnosis regarding the same patient based on their experiences, especially when the information are not fully collected.
In the field of computer security, different companies have their own standard in determining whether a software is malware or not. Thus when people combine different malware benchmarks together, there might be conflict labels on the same software.
In the image tagging process, users are allowed to create tags by themselves, thus when we may find different tags regarding the same object.
Other data such as movie ratings and application ratings also encounter inconsistent caused by the subjective classification process.
To summarize the inconsistent labels noise we have seen so far, we can divide them into two groups depend on whether the label can be represented using a numeric variable or categorical variable.
For questions regarding opinions of people,like interestingness of a book, a score between 0-10, or 1-5, is usually assigned to measure levels of agreement on a statement. So we can use a numeric value to represent each category.
For scenario like image tagging, each tag is a categorical label. Similarity, job titles created by different companies may be different for the same person, or the same job. There are no standard terminology regarding the same data.
Since subjective classification tasks often happens, especially when we want to model user behaviours and preference, and if we just ignore the inconsistent label noise introduced by the process, the learning will suffer from performance drop, and many other negative consequences as we have talked in the beginning of this talk.
So here comes the question, how people cope with inconsistent label noise?
Well, to answer this question, we need to first figure out when the inconsistent labels are encountered. Sometimes, we are the ones who can control the label collection process, sometime, we start with labelled data.
In the label collection process, some people do not aware of potential inconsistency introduced by the subjective nature of the classification task. While some people do care about the labeling process, especially when they are creating benchmark dataset. Several strategies are adopted, which I believe all researchers should consider before collecting human annotated labels.
For example, involving multiple labelers and making sure that one instance has been labelled more than one labelers. If disagreement appear regarding one instance, we should consider either throw the data, or resolve the disagreement among labelers. Recently, there is also a trend of adopting pairwise comparison rather than assigning an absolute score, but this method would require many pairs of comparisons.
If we do not have control on the process, but we have labels provided by multiple labelers, many studies go for different voting methods to merge multiple labelers’ into one label, but this method … Other studies will study the reliability of each labeler and filter outliers, or treat label as a distribution over all possible labels, which is called uncertain label.
But how about we only have one label per instance, especially when we know that nothing has been controlled in the label collection process. In the literature, rare work consider this case, but we keep seeing the danger of ignoring such inconsistent noise, which motivate my work on this area.
The key challenges for single labeler case, are:
How to measure inconsistency of labelers?
How to cope with such inconsistency?
And the key point of our solution is to leveraging collectively provided labels, which I call community intelligence on other tasks to transfer single labeler setting into multi labeler settings.
This work is part of my long-term research program which focuses on coping with data quality issues in big data settings.
bug severity level reflects the impact of bug on the system, it is assigned during bug reporting process, which is shown in the slide.