This document presents a new method for extracting class diagrams from textual requirements using natural language processing (NLP) techniques. It proposes the Requirements Analysis and Class diagram Extraction (RACE) system, which uses tools like the OpenNLP parser, a stemming algorithm, and WordNet to extract concepts and identify classes, attributes and relationships. The RACE system applies heuristic rules and a domain ontology to the output of the NLP tools to refine and finalize the extracted class diagram. The paper concludes that the RACE system demonstrates the effective use of NLP techniques to automate the extraction of class diagrams from informal natural language requirements specifications.
Imran Sarwar Bajwa, [2010], "Context Based Meaning Extraction by Means of Markov Logic", in International Journal of Computer Theory and Engineering - (IJCTE) 2(1) pp:35-38, February 2010
A New Concept Extraction Method for Ontology Construction From Arabic TextCSCJournals
Ontology is one of the most popular representation model used for knowledge representation, sharing and reusing. The Arabic language has complex morphological, grammatical, and semantic aspects. Due to complexity of Arabic language, automatic Arabic terminology extraction is difficult. In addition, concept extraction from Arabic documents has been challenging research area, because, as opposed to term extraction, concept extraction are more domain related and more selective. In this paper, we present a new concept extraction method for Arabic ontology construction, which is the part of our ontology construction framework. A new method to extract domain relevant single and multi-word concepts in the domain has been proposed, implemented and evaluated. Our method combines linguistic, statistical information and domain knowledge. It first uses linguistic patterns based on POS tags to extract concept candidates, and then stop words filter is implemented to filter unwanted strings. To determine relevance of these candidates within the domain, different statistical measures and new domain relevance measure are implemented for first time for Arabic language. To enhance the performance of concept extraction, a domain knowledge will be integrated into the module. The concepts scores are calculated according to their statistical values and domain knowledge values. In order to evaluate the performance of the method, precision scores were calculated. The results show the high effectiveness of the proposed approach to extract concepts for Arabic ontology construction.
Chunking means splitting the sentences into tokens and then grouping them in a meaningful way. When it comes to high-performance chunking systems, transformer models have proved to be the state of the art benchmarks. To perform chunking as a task it requires a large-scale high quality annotated corpus where each token is attached with a particular tag similar as that of Named Entity Recognition Tasks. Later these tags are used in conjunction with pointer frameworks to find the final chunk. To solve this for a specific domain problem, it becomes a highly costly affair in terms of time and resources to manually annotate and produce a large-high-quality training set. When the domain is specific and diverse, then cold starting becomes even more difficult because of the expected large number of manually annotated queries to cover all aspects. To overcome the problem, we applied a grammar-based text generation mechanism where instead of annotating a sentence we annotate using grammar templates. We defined various templates corresponding to different grammar rules. To create a sentence we used these templates along with the rules where symbol or terminal values were chosen from the domain data catalog. It helped us to create a large number of annotated queries. These annotated queries were used for training the machine learning model using an ensemble transformer-based deep neural network model [24.] We found that grammar-based annotation was useful to solve domain-based chunks in input query sentences without any manual annotation where it was found to achieve a classification F1 score of 96.97% in classifying the tokens for the out of template queries.
Sentiment Analysis In Myanmar Language Using Convolutional Lstm Neural Networkkevig
In recent years, there has been an increasing use of social media among people in Myanmar and writing
review on social media pages about the product, movie, and trip are also popular among people. Moreover,
most of the people are going to find the review pages about the product they want to buy before deciding
whether they should buy it or not. Extracting and receiving useful reviews over interesting products is very
important and time consuming for people. Sentiment analysis is one of the important processes for extracting
useful reviews of the products. In this paper, the Convolutional LSTM neural network architecture is
proposed to analyse the sentiment classification of cosmetic reviews written in Myanmar Language. The
paper also intends to build the cosmetic reviews dataset for deep learning and sentiment lexicon in Myanmar
Language.
Introduction to C++ : Object Oriented Technology, Advantages of OOP, Input- output in
C++, Tokens, Keywords, Identifiers, Data Types C++, Derives data types. The void data
type, Type Modifiers, Typecasting, Constant
Arabic morphology complex is a core challenge of
Arabic information retrieval. This limitation makes
Arabic language is so complex environment of
information retrieval specialists due to the high
inflected in Arabic language. This paper explains the
importance of the stemming in information retrieval
through developed a method to search about information needs in Arabic text. The new developed
method based on light stemmer to increase matching
the keywords between the query, with the related
document in the test collection.
Imran Sarwar Bajwa, [2010], "Context Based Meaning Extraction by Means of Markov Logic", in International Journal of Computer Theory and Engineering - (IJCTE) 2(1) pp:35-38, February 2010
A New Concept Extraction Method for Ontology Construction From Arabic TextCSCJournals
Ontology is one of the most popular representation model used for knowledge representation, sharing and reusing. The Arabic language has complex morphological, grammatical, and semantic aspects. Due to complexity of Arabic language, automatic Arabic terminology extraction is difficult. In addition, concept extraction from Arabic documents has been challenging research area, because, as opposed to term extraction, concept extraction are more domain related and more selective. In this paper, we present a new concept extraction method for Arabic ontology construction, which is the part of our ontology construction framework. A new method to extract domain relevant single and multi-word concepts in the domain has been proposed, implemented and evaluated. Our method combines linguistic, statistical information and domain knowledge. It first uses linguistic patterns based on POS tags to extract concept candidates, and then stop words filter is implemented to filter unwanted strings. To determine relevance of these candidates within the domain, different statistical measures and new domain relevance measure are implemented for first time for Arabic language. To enhance the performance of concept extraction, a domain knowledge will be integrated into the module. The concepts scores are calculated according to their statistical values and domain knowledge values. In order to evaluate the performance of the method, precision scores were calculated. The results show the high effectiveness of the proposed approach to extract concepts for Arabic ontology construction.
Chunking means splitting the sentences into tokens and then grouping them in a meaningful way. When it comes to high-performance chunking systems, transformer models have proved to be the state of the art benchmarks. To perform chunking as a task it requires a large-scale high quality annotated corpus where each token is attached with a particular tag similar as that of Named Entity Recognition Tasks. Later these tags are used in conjunction with pointer frameworks to find the final chunk. To solve this for a specific domain problem, it becomes a highly costly affair in terms of time and resources to manually annotate and produce a large-high-quality training set. When the domain is specific and diverse, then cold starting becomes even more difficult because of the expected large number of manually annotated queries to cover all aspects. To overcome the problem, we applied a grammar-based text generation mechanism where instead of annotating a sentence we annotate using grammar templates. We defined various templates corresponding to different grammar rules. To create a sentence we used these templates along with the rules where symbol or terminal values were chosen from the domain data catalog. It helped us to create a large number of annotated queries. These annotated queries were used for training the machine learning model using an ensemble transformer-based deep neural network model [24.] We found that grammar-based annotation was useful to solve domain-based chunks in input query sentences without any manual annotation where it was found to achieve a classification F1 score of 96.97% in classifying the tokens for the out of template queries.
Sentiment Analysis In Myanmar Language Using Convolutional Lstm Neural Networkkevig
In recent years, there has been an increasing use of social media among people in Myanmar and writing
review on social media pages about the product, movie, and trip are also popular among people. Moreover,
most of the people are going to find the review pages about the product they want to buy before deciding
whether they should buy it or not. Extracting and receiving useful reviews over interesting products is very
important and time consuming for people. Sentiment analysis is one of the important processes for extracting
useful reviews of the products. In this paper, the Convolutional LSTM neural network architecture is
proposed to analyse the sentiment classification of cosmetic reviews written in Myanmar Language. The
paper also intends to build the cosmetic reviews dataset for deep learning and sentiment lexicon in Myanmar
Language.
Introduction to C++ : Object Oriented Technology, Advantages of OOP, Input- output in
C++, Tokens, Keywords, Identifiers, Data Types C++, Derives data types. The void data
type, Type Modifiers, Typecasting, Constant
Arabic morphology complex is a core challenge of
Arabic information retrieval. This limitation makes
Arabic language is so complex environment of
information retrieval specialists due to the high
inflected in Arabic language. This paper explains the
importance of the stemming in information retrieval
through developed a method to search about information needs in Arabic text. The new developed
method based on light stemmer to increase matching
the keywords between the query, with the related
document in the test collection.
BIDIRECTIONAL LONG SHORT-TERM MEMORY (BILSTM)WITH CONDITIONAL RANDOM FIELDS (...ijnlc
This study investigates the effectiveness of Knowledge Named Entity Recognition in Online Judges (OJs). OJs are lacking in the classification of topics and limited to the IDs only. Therefore a lot of time is consumed in finding programming problems more specifically in knowledge entities.A Bidirectional Long Short-Term Memory (BiLSTM) with Conditional Random Fields (CRF) model is applied for the recognition of knowledge named entities existing in the solution reports.For the test run, more than 2000 solution reports are crawled from the Online Judges and processed for the model output. The stability of the model is
also assessed with the higher F1 value. The results obtained through the proposed BiLSTM-CRF model are more effectual (F1: 98.96%) and efficient in lead-time.
Imran Sarwar Bajwa, [2010], "Markov Logics Based Automated Business Requirements Analysis", in International Journal of Computer and Electrical Engineering (IJCEE) 2(3) pp:481-485, June 2010
Object Oriented Programming For Engineering Students as well as for B.Tech -IT. Covers Almost All From The Basics.
For more:
Google Search:: Prabhaharan Ellaiyan
This paper presents a natural language processing based automated system called DrawPlus for generating UML diagrams, user scenarios and test cases after analyzing the given business requirement specification which is written in natural language. The DrawPlus is presented for analyzing the natural languages and extracting the relative and required information from the given business requirement Specification by the user. Basically user writes the requirements specifications in simple English and the designed system has conspicuous ability to analyze the given requirement specification by using some of the core natural language processing techniques with our own well defined algorithms. After compound analysis and extraction of associated information, the DrawPlus system draws use case diagram, User scenarios and system level high level test case description. The DrawPlus provides the more convenient and reliable way of generating use case, user scenarios and test cases in a way reducing the time and cost of software development process while accelerating the 70 of works in Software design and Testing phase Janani Tharmaseelan ""Cohesive Software Design"" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-3 | Issue-3 , April 2019, URL: https://www.ijtsrd.com/papers/ijtsrd22900.pdf
Paper URL: https://www.ijtsrd.com/computer-science/other/22900/cohesive-software-design/janani-tharmaseelan
NL based Object Oriented modeling - EJSR 35(1)IT Industry
Imran Sarwar Bajwa, Shahzad Mumtaz, Ali Samad [2009], "Object Oriented Software Modeling using NLP Based Knowledge Extraction", European Journal of Scientific Research, Aug 2009, Vol. 35 No. 01, pp:22-33
Design analysis and Commissioning Of High Mast Lighting PolesIOSR Journals
Along a major highway, luminaire pole structures may be seen every 101 of a mile.From documented
cases, it appears that these structures started to experience fatigue problems in the last three decades. The
general public might not be aware of the problem, because if such a failure occurs, the structure is replaced.
Those working in the fatigue area realize that this issue is a serious matter[15][16]. Clearly, the damage is
costly, costing up to thousands of dollars per occurrence. For this purpose, a high mast lighting poles are
fabricated using steel due to its high strength, ductilityproperty and wear resistance. The high mast structure
(HMS) has the characters of light weight and high cost efficiency. It possess large ratio of height (H) to least
horizontal dimension (D) that makes it more slender and wind-sensitive than any other structures[17].
Therefore, the purpose of this research is to design optimal high mast poles taking into account its specification,
environmental conditions for placement and economy. Initially, among various pole designs, the high mast pole
is considered to be in tapered section as it is more reliable and economical. Then, analysis is performed in solid
works by keeping the base section to be fixed and applying compressive load on the top section of the pole due
to heavy weight of cantilever mast arm and luminaire. This project illustrates the theoretical basis and the
analytical development of the high mast lighting poles
IOSR Journal of Electrical and Electronics Engineering(IOSR-JEEE) is an open access international journal that provides rapid publication (within a month) of articles in all areas of electrical and electronics engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in electrical and electronics engineering. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
BIDIRECTIONAL LONG SHORT-TERM MEMORY (BILSTM)WITH CONDITIONAL RANDOM FIELDS (...ijnlc
This study investigates the effectiveness of Knowledge Named Entity Recognition in Online Judges (OJs). OJs are lacking in the classification of topics and limited to the IDs only. Therefore a lot of time is consumed in finding programming problems more specifically in knowledge entities.A Bidirectional Long Short-Term Memory (BiLSTM) with Conditional Random Fields (CRF) model is applied for the recognition of knowledge named entities existing in the solution reports.For the test run, more than 2000 solution reports are crawled from the Online Judges and processed for the model output. The stability of the model is
also assessed with the higher F1 value. The results obtained through the proposed BiLSTM-CRF model are more effectual (F1: 98.96%) and efficient in lead-time.
Imran Sarwar Bajwa, [2010], "Markov Logics Based Automated Business Requirements Analysis", in International Journal of Computer and Electrical Engineering (IJCEE) 2(3) pp:481-485, June 2010
Object Oriented Programming For Engineering Students as well as for B.Tech -IT. Covers Almost All From The Basics.
For more:
Google Search:: Prabhaharan Ellaiyan
This paper presents a natural language processing based automated system called DrawPlus for generating UML diagrams, user scenarios and test cases after analyzing the given business requirement specification which is written in natural language. The DrawPlus is presented for analyzing the natural languages and extracting the relative and required information from the given business requirement Specification by the user. Basically user writes the requirements specifications in simple English and the designed system has conspicuous ability to analyze the given requirement specification by using some of the core natural language processing techniques with our own well defined algorithms. After compound analysis and extraction of associated information, the DrawPlus system draws use case diagram, User scenarios and system level high level test case description. The DrawPlus provides the more convenient and reliable way of generating use case, user scenarios and test cases in a way reducing the time and cost of software development process while accelerating the 70 of works in Software design and Testing phase Janani Tharmaseelan ""Cohesive Software Design"" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-3 | Issue-3 , April 2019, URL: https://www.ijtsrd.com/papers/ijtsrd22900.pdf
Paper URL: https://www.ijtsrd.com/computer-science/other/22900/cohesive-software-design/janani-tharmaseelan
NL based Object Oriented modeling - EJSR 35(1)IT Industry
Imran Sarwar Bajwa, Shahzad Mumtaz, Ali Samad [2009], "Object Oriented Software Modeling using NLP Based Knowledge Extraction", European Journal of Scientific Research, Aug 2009, Vol. 35 No. 01, pp:22-33
Design analysis and Commissioning Of High Mast Lighting PolesIOSR Journals
Along a major highway, luminaire pole structures may be seen every 101 of a mile.From documented
cases, it appears that these structures started to experience fatigue problems in the last three decades. The
general public might not be aware of the problem, because if such a failure occurs, the structure is replaced.
Those working in the fatigue area realize that this issue is a serious matter[15][16]. Clearly, the damage is
costly, costing up to thousands of dollars per occurrence. For this purpose, a high mast lighting poles are
fabricated using steel due to its high strength, ductilityproperty and wear resistance. The high mast structure
(HMS) has the characters of light weight and high cost efficiency. It possess large ratio of height (H) to least
horizontal dimension (D) that makes it more slender and wind-sensitive than any other structures[17].
Therefore, the purpose of this research is to design optimal high mast poles taking into account its specification,
environmental conditions for placement and economy. Initially, among various pole designs, the high mast pole
is considered to be in tapered section as it is more reliable and economical. Then, analysis is performed in solid
works by keeping the base section to be fixed and applying compressive load on the top section of the pole due
to heavy weight of cantilever mast arm and luminaire. This project illustrates the theoretical basis and the
analytical development of the high mast lighting poles
IOSR Journal of Electrical and Electronics Engineering(IOSR-JEEE) is an open access international journal that provides rapid publication (within a month) of articles in all areas of electrical and electronics engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in electrical and electronics engineering. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
A Novel Edge Detection Technique for Image Classification and AnalysisIOSR Journals
Abstract: The main aim of this project is to propose a new method for image segmentation. Image
Segmentation is concerned with splitting an image up into segments (also called regions or areas) that each
holds some property distinct from their neighbor. Simply, another word for the Object Detection is
“Segmentation “. Segmentation is divided into two types they are Supervised Segmentation and Unsupervised
Segmentation. Segmentation consists of three types of methods which are divided on the basis of threshold, edge
and region. Thresholding is a commonly used enhancement whose goal is to segment an image into object and
background. Edge-based segmentations rely on edges found in an image by edge detecting operators. Region
based segmentations basic idea is to divide an image into zones of maximum homogeneity, where homogeneity
is an important property of regions. Edge detection has been a field of fundamental importance in digital image
processing research. Edge can be defined as a pixels located at points where abrupt changes in gray level take
place in this paper one novel approach for edge detection in gray scale images, which is based on diagonal
pixels in 2*2 region of the image, is proposed. This method first uses a threshold value to segment the image
and binary image. And then the proposed edge detector. In order to validate the results, seven different
kinds of test images are considered to examine the versatility of the proposed edge detector. It has been
observed that the proposed edge detector works effectively for different gray scale digital images. The results of
this study are quite promising. In this project we proposed a new algorithm for edge Detection. The main
advantage of this algorithm is with running mask on the original image we can detect the edges in the images by
using the proposed scheme for edge detection.
Keywords: Edge detection, segmentation, thresholding.
Ontology Based Approach for Semantic Information Retrieval SystemIJTET Journal
Abstract—The Information retrieval system is taking an important role in current search engine which performs searching operation based on keywords which results in an enormous amount of data available to the user, from which user cannot figure out the essential and most important information. This limitation may be overcome by a new web architecture known as the semantic web which overcome the limitation of the keyword based search technique called the conceptual or the semantic search technique. Natural language processing technique is mostly implemented in a QA system for asking user’s questions and several steps are also followed for conversion of questions to the query form for retrieving an exact answer. In conceptual search, search engine interprets the meaning of the user’s query and the relation among the concepts that document contains with respect to a particular domain that produces specific answers instead of showing lists of answers. In this paper, we proposed the ontology based semantic information retrieval system and the Jena semantic web framework in which, the user enters an input query which is parsed by Standford Parser then the triplet extraction algorithm is used. For all input queries, the SPARQL query is formed and further, it is fired on the knowledge base (Ontology) which finds appropriate RDF triples in knowledge base and retrieve the relevant information using the Jena framework.
Finding Bad Code Smells with Neural Network Models IJECEIAES
Code smell refers to any symptom introduced in design or implementation phases in the source code of a program. Such a code smell can potentially cause deeper and serious problems during software maintenance. The existing approaches to detect bad smells use detection rules or standards using a combination of different object-oriented metrics. Although a variety of software detection tools have been developed, they still have limitations and constraints in their capabilities. In this paper, a code smell detection system is presented with the neural network model that delivers the relationship between bad smells and object-oriented metrics by taking a corpus of Java projects as experimental dataset. The most well-known objectoriented metrics are considered to identify the presence of bad smells. The code smell detection system uses the twenty Java projects which are shared by many users in the GitHub repositories. The dataset of these Java projects is partitioned into mutually exclusive training and test sets. The training dataset is used to learn the network model which will predict smelly classes in this study. The optimized network model will be chosen to be evaluated on the test dataset. The experimental results show when the modelis highly trained with more dataset, the prediction outcomes are improved more and more. In addition, the accuracy of the model increases when it performs with higher epochs and many hidden layers.
Novel Database-Centric Framework for Incremental Information Extractionijsrd.com
Information extraction (IE) has been an active research area that seeks techniques to uncover information from a large collection of text. IE is the task of automatically extracting structured information from unstructured and/or semi structured machine-readable documents. In most of the cases this activity concerns processing human language texts by means of natural language processing (NLP). Recent activities in document processing like automatic annotation and content extraction could be seen as information extraction. Many applications call for methods to enable automatic extraction of structured information from unstructured natural language text. Due to the inherent challenges of natural language processing, most of the existing methods for information extraction from text tend to be domain specific. In this project a new paradigm for information extraction. In this extraction framework, intermediate output of each text processing component is stored so that only the improved component has to be deployed to the entire corpus. Extraction is then performed on both the previously processed data from the unchanged components as well as the updated data generated by the improved component. Performing such kind of incremental extraction can result in a tremendous reduction of processing time and there is a mechanism to generate extraction queries from both labeled and unlabeled data. Query generation is critical so that casual users can specify their information needs without learning the query language.
COMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUEJournal For Research
Natural Language Processing (NLP) techniques are one of the most used techniques in the field of computer applications. It has become one of the vast and advanced techniques. Language is the means of communication or interaction among humans and in present scenario when everything is dependent on machine or everything is computerized, communication between computer and human has become a necessity. To fulfill this necessity NLP has been emerged as the means of interaction which narrows the gap between machines (computers) and humans. It was evolved from the study of linguistics which was passed through the Turing test to check the similarity between data but it was limited to small set of data. Later on various algorithms were developed along with the concept of AI (Artificial Intelligence) for the successful execution of NLP. In this paper, the main emphasis is on the different techniques of NLP which have been developed till now, their applications and the comparison of all those techniques on different parameters.
BIDIRECTIONAL LONG SHORT-TERM MEMORY (BILSTM)WITH CONDITIONAL RANDOM FIELDS (...kevig
This study investigates the effectiveness of Knowledge Named Entity Recognition in Online Judges (OJs). OJs are lacking in the classification of topics and limited to the IDs only. Therefore a lot of time is consumed in finding programming problems more specifically in knowledge entities.A Bidirectional Long Short-Term Memory (BiLSTM) with Conditional Random Fields (CRF) model is applied for the recognition of knowledge named entities existing in the solution reports.For the test run, more than 2000 solution reports are crawled from the Online Judges and processed for the model output. The stability of the model is also assessed with the higher F1 value. The results obtained through the proposed BiLSTM-CRF model are more effectual (F1: 98.96%) and efficient in lead-time.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Le nuove frontiere dell'AI nell'RPA con UiPath Autopilot™UiPathCommunity
In questo evento online gratuito, organizzato dalla Community Italiana di UiPath, potrai esplorare le nuove funzionalità di Autopilot, il tool che integra l'Intelligenza Artificiale nei processi di sviluppo e utilizzo delle Automazioni.
📕 Vedremo insieme alcuni esempi dell'utilizzo di Autopilot in diversi tool della Suite UiPath:
Autopilot per Studio Web
Autopilot per Studio
Autopilot per Apps
Clipboard AI
GenAI applicata alla Document Understanding
👨🏫👨💻 Speakers:
Stefano Negro, UiPath MVPx3, RPA Tech Lead @ BSP Consultant
Flavio Martinelli, UiPath MVP 2023, Technical Account Manager @UiPath
Andrei Tasca, RPA Solutions Team Lead @NTT Data
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
GenAISummit 2024 May 28 Sri Ambati Keynote: AGI Belongs to The Community in O...
D017232729
1. IOSR Journal of Computer Engineering (IOSR-JCE)
e-ISSN: 2278-0661,p-ISSN: 2278-8727, Volume 17, Issue 2, Ver. III (Mar – Apr. 2015), PP 27-29
www.iosrjournals.org
DOI: 10.9790/0661-17232729 www.iosrjournals.org 27 | Page
Class Diagram Extraction from Textual Requirements Using NLP
Techniques
Vrushali Adhav1
, Darshana Ahire2
, Aarti Jadhav3
, Dipali Lokhande4
1,2,3,4,
(Computer Engineering, MET’s BKC IOE/ Savitribai Phule Pune University, India)
Abstract: A new method for translating software requirements to object-oriented model is proposed. Software
requirements specified in a natural language using scenario like format will be transformed into class diagrams
of Unified Modeling Language using specification transformation rule. The transformation rule is based on a
grammatical analysis, more specifically syntactical analysis, of requirements specification. This paper proposes
a new method for translating software requirements specified using natural language to formal specification (in
this context is executable and translatable Unified Modeling Language Class Diagram). Requirements
specification written in a scenario like format will be transformed into class diagram’s components the
automation of class generation from natural language requirements is highly challenging. This paper proposes
a method and a tool to facilitate requirements analysis process and class diagram extraction from textual
requirements supporting natural language processing techniques. Requirements engineers analyze requirements
manually to come out with analysis artifacts such as class diagram.
Keywords: NL (Natural Language), NLP (Natural Language processing), POS (Part Of Speech), RACE
(Requirement Analysis and Class Diagram Extraction), UML (Unified Modeling Language)
I. Introduction
Requirements are often poorly understood and articulated, even by domain experts. It is common to
find requirements specifications that are incomplete, inconsistent, ambiguous, and unstable. This problem
emerges since software requirements are inherently complex and involving various stakeholders. The fact that
specifying requirements should involve various stakeholders, who normally have no technical knowledge on
software analysis and design, imposes developer (software analyst and designer) to use specification media
which can be easily and intuitively understood by novice. In most practices, therefore, practitioners choose to
specify software requirements using natural languages. The reason why natural language is preferred as the
specification medium is because most stakeholders are more familiar with natural language compared to using
other types of media (such as formal notation or modeling language). Requirements specified in a natural
language are, however, normally written in a free form style. They still therefore need to be further analyzed and
transformed into more formal specification especially for development and documentation purposes. In a
number of software development methodologies, such as translate approach, formal specification is important
since software (source) code is obtained by translating it.
II. Related Work
a. Need of system [4]
One of the most challenging and costly functions performed by IT staff today is the deployment of
various software‟s to new or existing client computers. Currently, organizations spend a great deal of time and
expense planning, designing, and rolling out the latest version of the operating system throughout the
organization. Often this process is done manually, requiring a help desk professional to physically visit each
computer. This paper will assist analyst by providing an efficient and fast way to produce the class diagram for
their natural language requirement.
b. Existing Tools [3]
There have been several efforts for the analysis of natural language requirements. However, few are
focused on class diagram extraction from natural language requirements. Thus, few tools exist to assist analysts in
the extraction of class diagram. In this section we survey the works that use NLP or domain ontology techniques
to analyze NL requirements, and the works that aim to extract class diagram based on NLP or domain ontology
techniques. Following are some tools:
i. CIRCE
Ambriola and Gervasi present a Web-based environment called Circe. Circe helps in the elicitation,
selection, and validation of the software requirements. It can build semi-formal models, extract information from
the NL requirements, and measure the consistency of these models. Circe gives the user a complete environment
2. Class Diagram Extraction from Textual Requirements Using NLP Techniques
DOI: 10.9790/0661-17232729 www.iosrjournals.org 28 | Page
that integrates a number of tools. Cico is the main tool that is considered as a front-end for the other components;
it recognizes the NL sentences and extracts some facts from them. These facts are handed to the remaining tools
for graphical representation and analysis. Natural language sentence of requirements is parsed converted into a
forest of parse trees. These parse trees closely correspond to the original requirements but many surface features
of requirement document are not included in abstract UML model.
ii. Zhou and Zhou
It proposes a methodology that uses NLP and domain ontology. It is based on that the core classes are
always semantically connected to each other‟s by one to one, one to many, or many to many relationships in the
domain. This methodology Class Diagram from textual requirements using NLP techniques finds candidate
classes using NLP through a part of speech (POS) tagger, a link grammar ,parser, linguistic patterns and parallel
structure, and then the domain ontology is used to refine the result.
iii. LOLITA
Mich L. proposes a NLP system, LOLITA to generate an object model automatically from natural
language. This approach considers nouns as objects and use links to find relationships amongst objects. LOLITA
system is built on a large scale Semantic Network (SN) that does not distinguish between classes, attributes, and
objects. This approach is limited to extract objects and cannot identify classes. Hence for to overcome this
problem we are trying developing our system that mean we think how to solve this problem and we find solution
on this problem and we can apply on this our paper.
c. RACE System [2]
RACE system is decomposed into internal and external components and sub-systems. Figure 1
illustrates the architecture model of RACE:
i. OPEN NLP Parser
OpenNLP is an open-source and re-usable algorithm. It provides our system with lexical and syntactic
parsers. OpenNLP POS tagger (lexical) takes the English text as input and outputs the corresponding POS tags
for each word; On the other hand, OpenNLP Chunker (syntactic) chunks the sentence into phrases (Noun
phrase, verb phrase, etc.) according to English language grammar. The high accuracy and speed in OpenNLP
encouraged us to choose it rather than other existing parsers. OpenNLP uses lexical and syntactic annotations to
denote to the part of speech of the terms; for example, NN denotes to Proper Noun, VB denotes to Verb, and NP
denotes to Noun Phrase. OpenNLP parser supports our system with an efficient way to find the terms‟ part of
speech (POS) which we need in order to accomplish the noun and verb analysis.
ii. RACE Stemming Algorithm
Stemming is a technique that abbreviates word by removing affixes and suffixes. In RACE system, it is
very important to return words back to its base form; this will reduce the redundancy and increase the efficiency
of the system. To perform the stemming, we implemented a new stemming algorithm. Based on the stemming
result, we find that our stemming algorithm is efficient and sufficient to be used in the morphological analysis of
requirements in RACE system. Our stemming algorithm is simple and re-usable.
iii. WORDNET
WordNet is used to validate the semantic correctness of the sentences generated at the syntactic analysis.
It also enables users to display all hypernyms for a selected noun. We used this feature to verify Generalization
relationship where a noun phrase is supposed to be „a kind of‟ another noun phrase. WordNet can used to find
semantically similar terms, and for the acquisition of synonyms. We used synonyms to extract words which are
semantically related to each other. We calculated the words frequency to keep the synonyms with high
frequency in the document.
iv. Concept Extraction Engine
The aim of this module is to extract concepts according to the requirements document. This module uses
OpenNLP parser in, RACE stemming algorithm, and WordNet into extract concepts related to the given
requirements.
i. Class Extraction Engine
This module uses the output of concept extraction engine module and applies different heuristic rules to
extract the class diagram; However, We use domain ontology in this module to refine the extracted class
diagram. We can summarize the heuristic rules.
3. Class Diagram Extraction from Textual Requirements Using NLP Techniques
DOI: 10.9790/0661-17232729 www.iosrjournals.org 29 | Page
i. Domain Ontology
As mentioned early in this paper, domain ontology is used to improve the performance of concepts
identification. In RACE system we use a Library system Ontology as sample ontology. We gathered our
ontology based on a survey we conducted on some library systems, and then we organize the ontology in such
way to be easily maintained and re-used. The information in the ontology includes concepts related to classes
for Library system, Attributes, and relationships. We used the XML to build the ontology.
III. Figures And Tables
Fig 1: race system
IV. Conclusion
In this paper, we propose an enhanced approach that is based on NLP techniques to support the
extraction of class diagram from NL requirements. We validate our approach by implementing a system called
RACE referred to as “Requirements Analysis and class diagram Extraction”. RACE system efficiently
demonstrates the using of NLP techniques in the extraction of class diagram from informal requirements.
Acknowledgements
We have taken efforts in this paper. However, it would not have been possible without the kind support
and help of many individual. Our thanks and appreciation also goes to our beloved parents for their blessings, all
the staff members, friends and colleagues for providing help and support in our work.
References
Journal Papers:
[1] Booch, G. Object-Oriented Analysis and Design with Applications, 2nd Ed., Benjamin Cummings.
[2] Ambriola, V. and Gervasi, V. “Processing natural language requirements”, Proc. 12th IEEE Intl. Conf. on Automated Software
Engineering, pp.
[3] Farid Meziane, Nikos Athanasakis, Sophia Ananiadou, 2007,Generating Natural Language specifications from UML class diagrams,
Springer-Verlag London Limited .
[4] F Elizabeth D. Liddy & Jennifer H. Liddy, “An NLP Approach for Improving Access to Statistical Information for the Masses”.