This document summarizes a hierarchical transformer model for early detection of Alzheimer's disease from speech transcripts. The model uses BERT, RoBERTa, and ALBERT transformers trained on individual speech tasks and then combined in a pipeline and joint learning manner. Performance is evaluated on the B-SHARP dataset using accuracy, sensitivity, and specificity metrics. Attention analysis is conducted to understand what language features the models focus on. An ensemble of the transformers achieves the best performance for early Alzheimer's detection.
In this talk we explore how to build Machine Learning Systems that can that can learn "continuously" from their mistakes (feedback loop) and adapt to an evolving data distribution.
The youtube link to video of the talk is here:
https://www.youtube.com/watch?v=VtBvmrmMJaI
FriendsQA: Open-domain Question Answering on TV Show TranscriptsJinho Choi
This thesis presents FriendsQA, a challenging question answering dataset that contains 1,222 dialogues and 10,610 open-domain questions, to tackle machine comprehension on everyday conversations. Each dialogue, involving multiple speakers, is annotated with six types of questions what, when, why, where, who, how regarding the dialogue contexts, and the answers are annotated with contiguous spans in the dialogue. A series of crowdsourcing tasks are conducted to ensure good annotation quality, resulting a high inter-annotator agreement of 81.82%. A comprehensive annotation analytics is provided for a deeper understanding in this dataset. Three state-of-the-art QA systems are experimented, R-Net, QANet, and BERT, and evaluated on this dataset. BERT in particular depicts promising results, an accuracy of 74.2% for answer utterance selection and an F1-score of 64.2% for answer span selection, suggesting that the FriendsQA task is hard yet has a great potential of elevating QA research on multiparty dialogue to another level.
The slide is based on my seventh-semester project. I have done this project with my classmates. The project mainly focuses on the Nepali Character Classification Using the Neural Network Technique.
Deep Learning for NLP (without Magic) - Richard Socher and Christopher ManningBigDataCloud
A tutorial given at NAACL HLT 2013.
Richard Socher and Christopher Manning
http://nlp.stanford.edu/courses/NAACL2013/
Machine learning is everywhere in today's NLP, but by and large machine learning amounts to numerical optimization of weights for human designed representations and features. The goal of deep learning is to explore how computers can take advantage of data to develop features and representations appropriate for complex interpretation tasks. This tutorial aims to cover the basic motivation, ideas, models and learning algorithms in deep learning for natural language processing. Recently, these methods have been shown to perform very well on various NLP tasks such as language modeling, POS tagging, named entity recognition, sentiment analysis and paraphrase detection, among others. The most attractive quality of these techniques is that they can perform well without any external hand-designed resources or time-intensive feature engineering. Despite these advantages, many researchers in NLP are not familiar with these methods. Our focus is on insight and understanding, using graphical illustrations and simple, intuitive derivations. The goal of the tutorial is to make the inner workings of these techniques transparent, intuitive and their results interpretable, rather than black boxes labeled "magic here". The first part of the tutorial presents the basics of neural networks, neural word vectors, several simple models based on local windows and the math and algorithms of training via backpropagation. In this section applications include language modeling and POS tagging. In the second section we present recursive neural networks which can learn structured tree outputs as well as vector representations for phrases and sentences. We cover both equations as well as applications. We show how training can be achieved by a modified version of the backpropagation algorithm introduced before. These modifications allow the algorithm to work on tree structures. Applications include sentiment analysis and paraphrase detection. We also draw connections to recent work in semantic compositionality in vector spaces. The principle goal, again, is to make these methods appear intuitive and interpretable rather than mathematically confusing. By this point in the tutorial, the audience members should have a clear understanding of how to build a deep learning system for word-, sentence- and document-level tasks. The last part of the tutorial gives a general overview of the different applications of deep learning in NLP, including bag of words models. We will provide a discussion of NLP-oriented issues in modeling, interpretation, representational power, and optimization.
Presentation in Vietnam Japan AI Community in 2019-05-26.
The presentation summarizes what I've learned about Regularization in Deep Learning.
Disclaimer: The presentation is given in a community event, so it wasn't thoroughly reviewed or revised.
In this talk we explore how to build Machine Learning Systems that can that can learn "continuously" from their mistakes (feedback loop) and adapt to an evolving data distribution.
The youtube link to video of the talk is here:
https://www.youtube.com/watch?v=VtBvmrmMJaI
FriendsQA: Open-domain Question Answering on TV Show TranscriptsJinho Choi
This thesis presents FriendsQA, a challenging question answering dataset that contains 1,222 dialogues and 10,610 open-domain questions, to tackle machine comprehension on everyday conversations. Each dialogue, involving multiple speakers, is annotated with six types of questions what, when, why, where, who, how regarding the dialogue contexts, and the answers are annotated with contiguous spans in the dialogue. A series of crowdsourcing tasks are conducted to ensure good annotation quality, resulting a high inter-annotator agreement of 81.82%. A comprehensive annotation analytics is provided for a deeper understanding in this dataset. Three state-of-the-art QA systems are experimented, R-Net, QANet, and BERT, and evaluated on this dataset. BERT in particular depicts promising results, an accuracy of 74.2% for answer utterance selection and an F1-score of 64.2% for answer span selection, suggesting that the FriendsQA task is hard yet has a great potential of elevating QA research on multiparty dialogue to another level.
The slide is based on my seventh-semester project. I have done this project with my classmates. The project mainly focuses on the Nepali Character Classification Using the Neural Network Technique.
Deep Learning for NLP (without Magic) - Richard Socher and Christopher ManningBigDataCloud
A tutorial given at NAACL HLT 2013.
Richard Socher and Christopher Manning
http://nlp.stanford.edu/courses/NAACL2013/
Machine learning is everywhere in today's NLP, but by and large machine learning amounts to numerical optimization of weights for human designed representations and features. The goal of deep learning is to explore how computers can take advantage of data to develop features and representations appropriate for complex interpretation tasks. This tutorial aims to cover the basic motivation, ideas, models and learning algorithms in deep learning for natural language processing. Recently, these methods have been shown to perform very well on various NLP tasks such as language modeling, POS tagging, named entity recognition, sentiment analysis and paraphrase detection, among others. The most attractive quality of these techniques is that they can perform well without any external hand-designed resources or time-intensive feature engineering. Despite these advantages, many researchers in NLP are not familiar with these methods. Our focus is on insight and understanding, using graphical illustrations and simple, intuitive derivations. The goal of the tutorial is to make the inner workings of these techniques transparent, intuitive and their results interpretable, rather than black boxes labeled "magic here". The first part of the tutorial presents the basics of neural networks, neural word vectors, several simple models based on local windows and the math and algorithms of training via backpropagation. In this section applications include language modeling and POS tagging. In the second section we present recursive neural networks which can learn structured tree outputs as well as vector representations for phrases and sentences. We cover both equations as well as applications. We show how training can be achieved by a modified version of the backpropagation algorithm introduced before. These modifications allow the algorithm to work on tree structures. Applications include sentiment analysis and paraphrase detection. We also draw connections to recent work in semantic compositionality in vector spaces. The principle goal, again, is to make these methods appear intuitive and interpretable rather than mathematically confusing. By this point in the tutorial, the audience members should have a clear understanding of how to build a deep learning system for word-, sentence- and document-level tasks. The last part of the tutorial gives a general overview of the different applications of deep learning in NLP, including bag of words models. We will provide a discussion of NLP-oriented issues in modeling, interpretation, representational power, and optimization.
Presentation in Vietnam Japan AI Community in 2019-05-26.
The presentation summarizes what I've learned about Regularization in Deep Learning.
Disclaimer: The presentation is given in a community event, so it wasn't thoroughly reviewed or revised.
Classifying Non-Referential It for Question Answer PairsJinho Choi
This paper introduces a new corpus, QA-It, for the classification of non-referential it. Our dataset is unique in a sense that it is annotated on question answer pairs collected from multiple genres, useful for developing advanced QA systems. Our annotation scheme makes clear distinctions between 4 types of it, providing guidelines for many erroneous cases. Several statistical models are built for the classification of it, showing encouraging results. To the best of our knowledge, this is the first time that such a corpus is created for question answering.
GPT-2: Language Models are Unsupervised Multitask LearnersYoung Seok Kim
Review of paper
Language Models are Unsupervised Multitask Learners
(GPT-2)
by Alec Radford et al.
Paper link: https://d4mucfpksywv.cloudfront.net/better-language-models/language_models_are_unsupervised_multitask_learners.pdf
YouTube presentation: https://youtu.be/f5zULULWUwM
(Slides are written in English, but the presentation is done in Korean)
This presentation nlp classifiers, the different types of models tfidf, word2vec & DL models such as feed forward NN , CNN & siamese networks. Details on important metrics such as precision, recall AUC are also given
The ionization state of a chemical, reflected in pKa values, affects lipophilicity, solubility, protein binding and the ability of a chemical to cross the plasma membrane. These properties govern the pharmacokinetic parameters such as absorption, distribution, metabolism, excretion and toxicity and thus pKa is a fundamental chemical property and is used in many models of chemical toxicity.
Experimentally determining pKa is not feasible for high-throughput assays. Predicting pKa is challenging and existing models have been developed only using restricted chemical space (e.g., anilines, phenols, benzoic acids, primary amines) and lack of a generalized model impedes ADME modeling.
No free and open source models exist for heterogeneous chemical classes, however, several proprietary programs exist. In this work, pKa open data bundled with DataWarrior (http://www.openmolecules.org/) were used to develop predictive models for pKa. After data cleaning, there were ~3100 and ~3900 monoprotic chemicals with an acidic or basic pKa, respectively. 1D and 2D chemical descriptors (AlogP, Topological polar surface area, etc) in addition to 12 fingerprints (presence or absence of a chemical group) were generated using PaDEL software. Three datasets were used: acidic, basic and acidic and basic combined.
13 datasets were examined, the 1D/2D descriptors and 12 fingerprints. Using the Extreme Gradient Boosting algorithm showed that the MACCS and Substructure Count fingerprints yielded the best results, with models showing an R-Squared of ~0.78 and a RMSE of 1.42.
Recently, Deep Learning models have showed remarkable progress in image recognition and natural language processing. To determine if the Deep Learning algorithms would increase model performance we examined the datasets and found that the Deep Learning models were somewhat superior than Extreme Gradient Boosting with an R-Squared of ~0.80 and an RMSE of ~1.38.
This work does not reflect U.S. EPA policy.
2022_03_28 EDUCON 2022 “Replication of an Evaluation of Teacher Training in t...eMadrid network
2022_03_28 EDUCON 2022 “Replication of an Evaluation of Teacher Training in the Classification of Programming Exercises Using Bloom’s Taxonomy” - Ángel Velázquez Iturbide
Discusses the concept of Language Models in Natural Language Processing. The n-gram models, markov chains are discussed. Smoothing techniques such as add-1 smoothing, interpolation and discounting methods are addressed.
Toxic Comment Classification using Neural Network and Machine LearningCammy Soh Hui Shan
This project focuses on building a multi-label classification model to detect different types of toxicity in Wikipedia comments. The dataset is from https://www.kaggle.com/c/jigsaw-toxic-comment-classification-challenge.
Many models built by Kaggle participants center around word embeddings and deep learning models. We would like to observe if complex approaches of word embeddings coupled with deep learning pose any significant improvement in accuracy over simpler approaches.
Hence, we explored traditional approaches for vectorization using Bag-of-Words (BoW) and Term Frequency – Inverse Document Frequency (TF-IDF), as well as newer approaches like word embeddings.
We have deployed a range of classification methods, from simpler classification methods like Naïve Bayes, Logistic Regression, Support Vector Machines, to more complex methods of Neural Networks and CNN.
Deep Learning Architectures for NLP (Hungarian NLP Meetup 2016-09-07)Márton Miháltz
A brief survey of current deep learning/neural network methods currently used in NLP: recurrent networks (LSTM, GRU), recursive networks, convolutional networks, hybrid architectures, attention models. We will look at specific papers in the literature, targeting sentiment analysis, text classification and other tasks.
Decision Trees - The Machine Learning Magic UnveiledLuca Zavarella
Often a Machine Learning algorithm is seen as one of those magical weapons capable of revealing possible future scenarios to whoever holds it. In truth, it's a direct application of mathematical and statistical concepts, which sometimes generate complex models to be interpreted as output. However, there are predictive models based on decision trees that are really simple to understand. In this slide deck I'll explain what is behind a predictive model of this type.
Here the demo files: https://goo.gl/K6dgWC
Machine Learning in q/kdb+ - Teaching KDB to Read JapaneseMark Lefevre, CQF
Briefly introduce machine learning, supervised learning and neural networks. Implement neural network algorithms in q/kdb+ to recognize handwritten Japanese characters.
Classifying Non-Referential It for Question Answer PairsJinho Choi
This paper introduces a new corpus, QA-It, for the classification of non-referential it. Our dataset is unique in a sense that it is annotated on question answer pairs collected from multiple genres, useful for developing advanced QA systems. Our annotation scheme makes clear distinctions between 4 types of it, providing guidelines for many erroneous cases. Several statistical models are built for the classification of it, showing encouraging results. To the best of our knowledge, this is the first time that such a corpus is created for question answering.
GPT-2: Language Models are Unsupervised Multitask LearnersYoung Seok Kim
Review of paper
Language Models are Unsupervised Multitask Learners
(GPT-2)
by Alec Radford et al.
Paper link: https://d4mucfpksywv.cloudfront.net/better-language-models/language_models_are_unsupervised_multitask_learners.pdf
YouTube presentation: https://youtu.be/f5zULULWUwM
(Slides are written in English, but the presentation is done in Korean)
This presentation nlp classifiers, the different types of models tfidf, word2vec & DL models such as feed forward NN , CNN & siamese networks. Details on important metrics such as precision, recall AUC are also given
The ionization state of a chemical, reflected in pKa values, affects lipophilicity, solubility, protein binding and the ability of a chemical to cross the plasma membrane. These properties govern the pharmacokinetic parameters such as absorption, distribution, metabolism, excretion and toxicity and thus pKa is a fundamental chemical property and is used in many models of chemical toxicity.
Experimentally determining pKa is not feasible for high-throughput assays. Predicting pKa is challenging and existing models have been developed only using restricted chemical space (e.g., anilines, phenols, benzoic acids, primary amines) and lack of a generalized model impedes ADME modeling.
No free and open source models exist for heterogeneous chemical classes, however, several proprietary programs exist. In this work, pKa open data bundled with DataWarrior (http://www.openmolecules.org/) were used to develop predictive models for pKa. After data cleaning, there were ~3100 and ~3900 monoprotic chemicals with an acidic or basic pKa, respectively. 1D and 2D chemical descriptors (AlogP, Topological polar surface area, etc) in addition to 12 fingerprints (presence or absence of a chemical group) were generated using PaDEL software. Three datasets were used: acidic, basic and acidic and basic combined.
13 datasets were examined, the 1D/2D descriptors and 12 fingerprints. Using the Extreme Gradient Boosting algorithm showed that the MACCS and Substructure Count fingerprints yielded the best results, with models showing an R-Squared of ~0.78 and a RMSE of 1.42.
Recently, Deep Learning models have showed remarkable progress in image recognition and natural language processing. To determine if the Deep Learning algorithms would increase model performance we examined the datasets and found that the Deep Learning models were somewhat superior than Extreme Gradient Boosting with an R-Squared of ~0.80 and an RMSE of ~1.38.
This work does not reflect U.S. EPA policy.
2022_03_28 EDUCON 2022 “Replication of an Evaluation of Teacher Training in t...eMadrid network
2022_03_28 EDUCON 2022 “Replication of an Evaluation of Teacher Training in the Classification of Programming Exercises Using Bloom’s Taxonomy” - Ángel Velázquez Iturbide
Discusses the concept of Language Models in Natural Language Processing. The n-gram models, markov chains are discussed. Smoothing techniques such as add-1 smoothing, interpolation and discounting methods are addressed.
Toxic Comment Classification using Neural Network and Machine LearningCammy Soh Hui Shan
This project focuses on building a multi-label classification model to detect different types of toxicity in Wikipedia comments. The dataset is from https://www.kaggle.com/c/jigsaw-toxic-comment-classification-challenge.
Many models built by Kaggle participants center around word embeddings and deep learning models. We would like to observe if complex approaches of word embeddings coupled with deep learning pose any significant improvement in accuracy over simpler approaches.
Hence, we explored traditional approaches for vectorization using Bag-of-Words (BoW) and Term Frequency – Inverse Document Frequency (TF-IDF), as well as newer approaches like word embeddings.
We have deployed a range of classification methods, from simpler classification methods like Naïve Bayes, Logistic Regression, Support Vector Machines, to more complex methods of Neural Networks and CNN.
Deep Learning Architectures for NLP (Hungarian NLP Meetup 2016-09-07)Márton Miháltz
A brief survey of current deep learning/neural network methods currently used in NLP: recurrent networks (LSTM, GRU), recursive networks, convolutional networks, hybrid architectures, attention models. We will look at specific papers in the literature, targeting sentiment analysis, text classification and other tasks.
Decision Trees - The Machine Learning Magic UnveiledLuca Zavarella
Often a Machine Learning algorithm is seen as one of those magical weapons capable of revealing possible future scenarios to whoever holds it. In truth, it's a direct application of mathematical and statistical concepts, which sometimes generate complex models to be interpreted as output. However, there are predictive models based on decision trees that are really simple to understand. In this slide deck I'll explain what is behind a predictive model of this type.
Here the demo files: https://goo.gl/K6dgWC
Machine Learning in q/kdb+ - Teaching KDB to Read JapaneseMark Lefevre, CQF
Briefly introduce machine learning, supervised learning and neural networks. Implement neural network algorithms in q/kdb+ to recognize handwritten Japanese characters.
Welocme to ViralQR, your best QR code generator.ViralQR
Welcome to ViralQR, your best QR code generator available on the market!
At ViralQR, we design static and dynamic QR codes. Our mission is to make business operations easier and customer engagement more powerful through the use of QR technology. Be it a small-scale business or a huge enterprise, our easy-to-use platform provides multiple choices that can be tailored according to your company's branding and marketing strategies.
Our Vision
We are here to make the process of creating QR codes easy and smooth, thus enhancing customer interaction and making business more fluid. We very strongly believe in the ability of QR codes to change the world for businesses in their interaction with customers and are set on making that technology accessible and usable far and wide.
Our Achievements
Ever since its inception, we have successfully served many clients by offering QR codes in their marketing, service delivery, and collection of feedback across various industries. Our platform has been recognized for its ease of use and amazing features, which helped a business to make QR codes.
Our Services
At ViralQR, here is a comprehensive suite of services that caters to your very needs:
Static QR Codes: Create free static QR codes. These QR codes are able to store significant information such as URLs, vCards, plain text, emails and SMS, Wi-Fi credentials, and Bitcoin addresses.
Dynamic QR codes: These also have all the advanced features but are subscription-based. They can directly link to PDF files, images, micro-landing pages, social accounts, review forms, business pages, and applications. In addition, they can be branded with CTAs, frames, patterns, colors, and logos to enhance your branding.
Pricing and Packages
Additionally, there is a 14-day free offer to ViralQR, which is an exceptional opportunity for new users to take a feel of this platform. One can easily subscribe from there and experience the full dynamic of using QR codes. The subscription plans are not only meant for business; they are priced very flexibly so that literally every business could afford to benefit from our service.
Why choose us?
ViralQR will provide services for marketing, advertising, catering, retail, and the like. The QR codes can be posted on fliers, packaging, merchandise, and banners, as well as to substitute for cash and cards in a restaurant or coffee shop. With QR codes integrated into your business, improve customer engagement and streamline operations.
Comprehensive Analytics
Subscribers of ViralQR receive detailed analytics and tracking tools in light of having a view of the core values of QR code performance. Our analytics dashboard shows aggregate views and unique views, as well as detailed information about each impression, including time, device, browser, and estimated location by city and country.
So, thank you for choosing ViralQR; we have an offer of nothing but the best in terms of QR code services to meet business diversity!
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
SAP Sapphire 2024 - ASUG301 building better apps with SAP Fiori.pdfPeter Spielvogel
Building better applications for business users with SAP Fiori.
• What is SAP Fiori and why it matters to you
• How a better user experience drives measurable business benefits
• How to get started with SAP Fiori today
• How SAP Fiori elements accelerates application development
• How SAP Build Code includes SAP Fiori tools and other generative artificial intelligence capabilities
• How SAP Fiori paves the way for using AI in SAP apps
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Le nuove frontiere dell'AI nell'RPA con UiPath Autopilot™UiPathCommunity
In questo evento online gratuito, organizzato dalla Community Italiana di UiPath, potrai esplorare le nuove funzionalità di Autopilot, il tool che integra l'Intelligenza Artificiale nei processi di sviluppo e utilizzo delle Automazioni.
📕 Vedremo insieme alcuni esempi dell'utilizzo di Autopilot in diversi tool della Suite UiPath:
Autopilot per Studio Web
Autopilot per Studio
Autopilot per Apps
Clipboard AI
GenAI applicata alla Document Understanding
👨🏫👨💻 Speakers:
Stefano Negro, UiPath MVPx3, RPA Tech Lead @ BSP Consultant
Flavio Martinelli, UiPath MVP 2023, Technical Account Manager @UiPath
Andrei Tasca, RPA Solutions Team Lead @NTT Data
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...
Hierarchical Transformer for Early Detection of Alzheimer’s Disease
1. Hierarchical Transformer
for Early Detection of
Alzheimer’s Disease
Renxuan Li
Advisor: Dr. Jinho Choi
Emory University, Department of Computer Science
2. Content
• Classification task: MCI detection
• Background and Previous Work
• Dataset: B-SHARP
• Models for Individual Tasks
• Hierarchical Transformer Model
• Attention Analysis
• Ensemble Model Analysis
3. Introduction and motivation
• In a short word, we are trying to help the diagnosis of the early stage of
Alzheimer’s Disease(AD) with Natural Language Processing (NLP) methods.
• This study is essentially an MCI/normal classification task
• Early stage of AD is also known as Mild Cognitive Impairment (MCI)
• AD is not reversible or curable, but there are treatments to slow down the
development of AD, therefore, MCI diagnosis is very crucial.
• Our study focus on predicting subjects’ brain-health condition by processing their
transcribed speech.
5. Classification layers
• All classification layers in this study are Multi-Layer Perceptron with
different input size.
• MLP in formula: 𝑜𝑢𝑡 = 𝑖𝑛𝑝𝑢𝑡 ∗ 𝑊 + 𝑏
6. Related works: DementiaBank
• Tasks: Cookie Theft Picture Description
• 99 control subjects, 194 dementia subjects;
243 control recordings and 309 dementia
recordings; 552 in total.
• Yearly visits from subjects.
• Include grammatical features
• Cleaned up transcripts
7. Previous works: CNN
• Convolutional Neural Network (CNN)
• Captures local features well
• May use filters of different sizes
https://arxiv.org/pdf/1408.5882.pdf
8. Previous work: C-LSTM
• LSTM: Long Short Term Memory
Network, contains information of
sequence over time
• C-LSTM: Add LSTM layer over CNN
https://www.aclweb.org/anthology/N18-2110.pdf
9. B-SHARP dataset
• Brain Stress Hypertension Aging
Research Program
• Focus on MCI
• 185 normal subjects, 141 MCI
patients
• 650 transcripts in total: 385 Normal
and 265 MCI
• Transcribed using online tool
without manual correction
10. Speech Protocol
• In this study, we use only three of out five tasks of the speech protocol to do
classification.
• Question 1: Describe subjects’ events of the day
• Question 2: Describe the examination room
• Question 3: Describe the picture: Circus Procession
13. Language features
• We count 7 major average measures: # tokens, # sentences, # nouns, #verb,
# Conjunctions, #Complex structures, # Discourse.
• We see in all categories except # Discourse, the control subjects have
higher values, which means longer, and more complex speech.
14. Challenge
• 1. Transcription error: our text is not corrected manually.
• 2. Long sequence: Control entire transcript has on average 578 tokens and
MCI transcript has on average 548 tokens. Too long to fit into transformer
model directly
15. Approach
• CNN Baseline
• On individual Tasks
• BERT
• RoBERTa
• ALBERT
• Pipeline Hierarchical Transformer Model
• Hierarchical Transformer Joint Learning
16. CNN with larger filters
• We add filters of size 3,5,7,9,11,13,15, meaning that we can capture
relations 15 words apart.
• Static FastText Embedding with embedding size 300
• Channel size 600
17. self attention and
transformer encoder
• Transformer encoder:
• Based entirely on Self-
Attention Mechanism
• Multi-head attention of 12
heads
https://www.kdnuggets.com/2019/03/deconstructing-bert-part-2-visualizing-inner-workings-
attention.html
18. BERT for sequence classification
• Base model: 12 layers of transformer encoder, 12 attention heads each
• Pretrain tasks: Mask Language Model, Next Sentence Prediction
• Trained on BookCurpus (800M words) English Wiki(2500M words)
• Normally sequence over 512 cause memory overflow (on our machine)
• Only trained on individual tasks in speech protocol
20. RoBERTa for sequence classification
• Same architecture as BERT
• Dynamic Masked Language Model
• More careful parameter choice
• Longer training and bigger batch
• Normally sequence over length 512 cause memory overflow (on our
machine)
• Only trained on individual tasks in speech protocol
21. ALBERT for sequence classification
• Similar Architecture with BERT and RoBERTa
• Compressed embedding size.
• Share parameters across layers
• 9 times smaller than BERT-base
• Can fit three models onto our machine
22. Pipeline Hierarchical Transformer Model
• Each layer of model are kept
independent during training.
• Take [CLS] tokens from each
individual transformer models and
feed into an MLP after
concatenation
• Transformer trained on their own.
24. Data split
• Dataset is too small for conventional
8:1:1 train:dev:test set split,
therefore we use Five-Fold Cross
Validation.
• Each subject only appears in one set
Set0 Set1 Set2 Set3 Set4
# MCI 53 53 53 53 53
# Control 77 77 77 77 77
avg
MoCA
MCI
21.8 20.0 21.4 22.3 22.5
avg
MoCA
Control
26.1 26.5 25.7 26.9 26.7
MoCA: Montreal Cognitive Assessment.
Higher scores means better brain-condition
26. Experiment set up
• Every model is trained three times, average and standard derivation are recorded.
• Three metric:
• Accuracy:
!"#$%&'()(*$+,"#$-$./)(*$
011
• Sensitivity:
!"#$%&'()(*$
!"#$%&'()(*$+2/3'$-$./)(*$
• Specificity:
!"#$-$./)(*$
!"#$-$./)(*$+2/3'$%&'()(*$
27. Performance on Individual tasks
• All three models performs good on Q2
• Overall, RoBERTa does the best.
• BERT does better on Q3
28. Performance on entire transcript
• We see gradual and substantial improvement on classification
performance
• RoBERTa ensemble model does the best among single model ensemble
• Sensitivity Improvement
• Specificity remains level
• Joint Learning not as good as Pipeline style
30. Attention analysis:
measures and methods
• We rank the tokens from a
transcript from highest ACR to
the lowest ACR.
• If a token has high ACR, it’s
getting more attended during
training, and more related to
how our models predict.
Attention Change Ratio:
31. On Roberta q1 model
MCI
• 1 Yeah 142.50%
• 2 ina 129.14%
• 3 tests 124.17%
• 4 Okay 122.25%
• 5 And 112.76%
• 6 ud 80.30%
• 7 now 75.60%
• 8 no 70.89%
• 9 body 66.21%
• 10 me 60.03%
Normal
• 1 friend 182.31%
• 2 arrived 169.92%
• 3 From 145.92%
• 4 Tiffany 140.62%
• 5 visit 137.53%
• 6 taken 122.23%
• 7 tests 114.66%
• 8 downstairs 109.35%
• 9 snack 107.38%
• 10 how 93.17%
43. Task Comparison: Measures
• Correct Vote Ratio for question i: among 466 transcripts that the
ensemble model correctly classifies, in how much percentage that
the vote of question i is correct.
• Normalized Correct Vote Ratio: for each transcript, if the ensemble
model predicts correctly, then there will be 1 “point” given evenly to
the questions whose votes are correct. NCV is the average “point” a
question get for each transcript.
• % of Dominant Cases: percentage that the vote for question i is the
only correct one.
45. B+R+A ensemble
86.6% ARE
MAJORITY VOTES
6-CORRECT
SITUATION:34.50%
5-CORRECT
SITUATION 28.51%
RARELY ALL 9
INPUTS AGREE
(0.21%)
NO SITUATIONS
THAT LESS THAN 4
VOTES ARE
CORRECT.
46. B+R+A Ensemble: which input is more
important?
• The top three inputs form a set of transcripts
• Again, individual model accuracy (IMA) does not always match
importance in ensemble model
48. Bibliography
• [1] James T. Becker, Francois Boller, Oscar L. Lopez, Judith Saxton, andKaren L. McGonigle. The
natural history of alzheimer’s disease: descrip-tion of study cohort and accuracy of
diagnosis.Archives of Neurology,51(6):585–594, 1994.
• [2] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova.BERT: Pre-training of Deep
Bidirectional Transformers for Lan-guage Understanding. InProceedings of the Conference of the
NorthAmerican Chapter of the Association for Computational Linguis-tics: Human Language
Technologies, pages 4171–4186, 2019. URLhttps://www.aclweb.org/anthology/N19-1423.
• [3] Serge Gauthier, Michel Panisset, Josephine Nalban-toglu, and JudesPoirie. Alzheimer’s dis-
ease: current knowledge, management and re-search.Canadian Medical Association Journal,
157(8):1047–1052, 1997.54
• [4] George G Glenner.Alzheimers disease.Springer, 1990.
49. Bibliography
• [5] Sweta Karleka, Tong Niu, and Mohit Bansal. Detecting linguistic charac-teristics of alzheimer’s dementia by interpreting
neural models. InPro-ceedings of the 2018 Conference of the North American Chapter of the As-sociation for
Computational Linguistics: Human Language Technologies,Volume 2 (Short Papers), pages 701–707, New Orleans,
Louisiana, June2018. Association for Computational Linguistics. doi: 10.18653/v1/N18-2110.
URLhttps://www.aclweb.org/anthology/N18-2110.
• [6] Yoon Kim.Convolutional Neural Networks for Sentence Clas-sification.InProceedings of the 2014 Conference on
EmpiricalMethods in Natural Language Processing (EMNLP), 2014.URLhttps://www.aclweb.org/anthology/D14-1181.
• [7] Taku Kudo and John Richardson. SentencePiece: A simple and lan-guage independent subword tokenizer and
detokenizer for Neural TextProcessing. InProceedings of the 2018 Conference on Empirical Methodsin Natural Language
Processing: System Demonstrations, pages 66–71,2018. URLhttps://www.aclweb.org/anthology/D18-2012.[8]Zhenzhong
Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel,
• [8] Piyush Sharma, and Radu Soricut. ALBERT: A Lite BERT for Self-supervised Learning of Language Representations.arXiv,
11942(1909),2019.
50. Bibliography
• [9]Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, DanqiChen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and
Veselin Stoyanov.RoBERTa: A Robustly Optimized BERT Pretraining Approach. InProceedings of the International
Conference on Learning Representations,2020. URLhttps://openreview.net/forum?id=SyxS0T4tvS.
• [10]Frank Rudzicz, Rosalie Wang, Momotaz Begum, and Alex Mihailidis.Speech recognition in alzheimer’s disease with
personal assistive robots.InProceedings of the 5th Workshop on Speech and Language Processingfor Assistive
Technologies), page 20–28, Baltimore, Maryland, USA, June2014. Association for Computational Linguistics. doi:
10.3115/v1/W14-1904. URLhttps://www.aclweb.org/anthology/W14-1904.
• [11] Richard Suzman and John Beard. Global health and aging., 2011.
• [12]Yonghui Wu, Mike Schuster, Zhifeng Chen, Quoc V. Le, MohammadNorouzi, Wolfgang Macherey, Maxim Krikun, Yuan
Cao, Qin Gao,Klaus Macherey, Jeff Klingner, Apurva Shah, Melvin Johnson, Xiaob]ing Liu, Lukasz Kaiser, Stephan Gouws,
Yoshikiyo Kato, Taku Kudo,Hideto Kazawa, Keith Stevens, George Kurian, Nishant Patil, WeiWang, Cliff Young, Jason Smith,
Jason Riesa, Alex Rudnick, OriolVinyals, Greg Corrado, Macduff Hughes, and Jeffrey Dean. Google’sNeural Machine
Translation System: Bridging the Gap between Hu-man and Machine Translation.arXiv, 1609(08144),
2016.URLhttp://arxiv.org/abs/1609.08144.