Data mining is valuable technology to facilitate the extraction of useful patterns and trends from large volume of data. When these patterns are to be shared in a collaborative environment, they must be protectively shared among the parties concerned in order to preserve the confidentiality of the sensitive data. Sharing of information may be in the form of datasets or in any of the structured patterns like trees, graphs, lattices, etc., This paper propose a sanitization algorithm for protecting sensitive data in a structured frequent pattern(tree).
Comparative study of ksvdd and fsvm for classification of mislabeled dataeSAT Journals
Abstract Outlier detection is the important concept in data mining. These outliers are the data that differ from the normal data. Noise in the
application may cause the misclassification of data. Data are more likely to be mislabeled in presence of noise leading to
performance degradation. The proposed work focuses on these issues. Data before classifying is given a value that represents its
willingness towards the class. This data with likelihood value is then given to classifier to predict the data. SVDD algorithm is
used for classification of data with likelihood values.
Keywords: Confusion Matrix, FSVM, Outlier, Outlier Detection, SVDD
Analysis of Classification Algorithm in Data Miningijdmtaiir
Data Mining is the extraction of hidden predictive
information from large database. Classification is the process
of finding a model that describes and distinguishes data classes
or concept. This paper performs the study of prediction of class
label using C4.5 and Naïve Bayesian algorithm.C4.5 generates
classifiers expressed as decision trees from a fixed set of
examples. The resulting tree is used to classify future samples
.The leaf nodes of the decision tree contain the class name
whereas a non-leaf node is a decision node. The decision node
is an attribute test with each branch (to another decision tree)
being a possible value of the attribute. C4.5 uses information
gain to help it decide which attribute goes into a decision node.
A Naïve Bayesian classifier is a simple probabilistic classifier
based on applying Baye’s theorem with strong (naive)
independence assumptions. Naive Bayesian classifier assumes
that the effect of an attribute value on a given class is
independent of the values of the other attribute. This
assumption is called class conditional independence. The
results indicate that Predicting of class label using Naïve
Bayesian classifier is very effective and simple compared to
C4.5 classifier
Dimensionality reduction by matrix factorization using concept lattice in dat...eSAT Journals
Abstract Concept lattices is the important technique that has become a standard in data analytics and knowledge presentation in many fields such as statistics, artificial intelligence, pattern recognition ,machine learning ,information theory ,social networks, information retrieval system and software engineering. Formal concepts are adopted as the primitive notion. A concept is jointly defined as a pair consisting of the intension and the extension. FCA can handle with huge amount of data it generates concepts and rules and data visualization. Matrix factorization methods have recently received greater exposure, mainly as an unsupervised learning method for latent variable decomposition. In this paper a novel method is proposed to decompose such concepts by using Boolean Matrix Factorization for dimensionality reduction. This paper focuses on finding all the concepts and the object intersections. Keywords: Data mining, formal concepts, lattice, matrix factorization dimensionality reduction.
Comparative study of ksvdd and fsvm for classification of mislabeled dataeSAT Journals
Abstract Outlier detection is the important concept in data mining. These outliers are the data that differ from the normal data. Noise in the
application may cause the misclassification of data. Data are more likely to be mislabeled in presence of noise leading to
performance degradation. The proposed work focuses on these issues. Data before classifying is given a value that represents its
willingness towards the class. This data with likelihood value is then given to classifier to predict the data. SVDD algorithm is
used for classification of data with likelihood values.
Keywords: Confusion Matrix, FSVM, Outlier, Outlier Detection, SVDD
Analysis of Classification Algorithm in Data Miningijdmtaiir
Data Mining is the extraction of hidden predictive
information from large database. Classification is the process
of finding a model that describes and distinguishes data classes
or concept. This paper performs the study of prediction of class
label using C4.5 and Naïve Bayesian algorithm.C4.5 generates
classifiers expressed as decision trees from a fixed set of
examples. The resulting tree is used to classify future samples
.The leaf nodes of the decision tree contain the class name
whereas a non-leaf node is a decision node. The decision node
is an attribute test with each branch (to another decision tree)
being a possible value of the attribute. C4.5 uses information
gain to help it decide which attribute goes into a decision node.
A Naïve Bayesian classifier is a simple probabilistic classifier
based on applying Baye’s theorem with strong (naive)
independence assumptions. Naive Bayesian classifier assumes
that the effect of an attribute value on a given class is
independent of the values of the other attribute. This
assumption is called class conditional independence. The
results indicate that Predicting of class label using Naïve
Bayesian classifier is very effective and simple compared to
C4.5 classifier
Dimensionality reduction by matrix factorization using concept lattice in dat...eSAT Journals
Abstract Concept lattices is the important technique that has become a standard in data analytics and knowledge presentation in many fields such as statistics, artificial intelligence, pattern recognition ,machine learning ,information theory ,social networks, information retrieval system and software engineering. Formal concepts are adopted as the primitive notion. A concept is jointly defined as a pair consisting of the intension and the extension. FCA can handle with huge amount of data it generates concepts and rules and data visualization. Matrix factorization methods have recently received greater exposure, mainly as an unsupervised learning method for latent variable decomposition. In this paper a novel method is proposed to decompose such concepts by using Boolean Matrix Factorization for dimensionality reduction. This paper focuses on finding all the concepts and the object intersections. Keywords: Data mining, formal concepts, lattice, matrix factorization dimensionality reduction.
Hypothesis on Different Data Mining AlgorithmsIJERA Editor
In this paper, different classification algorithms for data mining are discussed. Data Mining is about
explaining the past & predicting the future by means of data analysis. Classification is a task of data mining,
which categories data based on numerical or categorical variables. To classify the data many algorithms are
proposed, out of them five algorithms are comparatively studied for data mining through classification. There are
four different classification approaches namely Frequency Table, Covariance Matrix, Similarity Functions &
Others. As work for research on classification methods, algorithms like Naive Bayesian, K Nearest Neighbors,
Decision Tree, Artificial Neural Network & Support Vector Machine are studied & examined using benchmark
datasets like Iris & Lung Cancer.
Hypothesis on Different Data Mining AlgorithmsIJERA Editor
In this paper, different classification algorithms for data mining are discussed. Data Mining is about
explaining the past & predicting the future by means of data analysis. Classification is a task of data mining,
which categories data based on numerical or categorical variables. To classify the data many algorithms are
proposed, out of them five algorithms are comparatively studied for data mining through classification. There are
four different classification approaches namely Frequency Table, Covariance Matrix, Similarity Functions &
Others. As work for research on classification methods, algorithms like Naive Bayesian, K Nearest Neighbors,
Decision Tree, Artificial Neural Network & Support Vector Machine are studied & examined using benchmark
datasets like Iris & Lung Cancer.
Performance Assessment of Faculties of Management Discipline From Student Per...Waqas Tariq
This paper deals with Faculty Performance Assessment from student perspective using Data Analysis and Mining techniques .Performance of a faculty depends on a number of parameters (77 parameters as identified) and the performance assessment of a faculty/faculties are broadly carried out by the Management Body ,the Student Community ,Self and Peer faculties of the organization .The parameters act as performance indicators for an individual and group and subsequently can impact on the decision making of the stakeholders. The idea proposed in this paper is to perform an analysis of faculty performance considering student feedback which can directly or indirectly impact management’s decision, teaching standards and norms set by the educational institute, understand certain patterns of faculty motivation, satisfaction, growth and decline in future. The analysis depends on many factors, encompassing student’s feedback, organizational feedback, institutional support in terms of finance, administration, research activity etc. The data analysis and mining methodology used for extracting useful patterns from the institutional database has been used to extract certain trends in faculty performance when assessed on student feedback.
Sponsoring & Marketing Plan prodotto per l'AC Savoia, squadra professionista di calcio neo-promossa in Lega-Pro.
Materiale prodotto durante il periodo trascorso come ArtDirector per VitaminMarketing, che si ringrazia per la collaborazione.
FARPI-FRANCE vous présente ses convoyeurs standards VIRGINIOFARPI-FRANCE
Spécialisée dans la commercialisation d'équipements pour la transformation des matières plastiques et l’industrie pharmaceutique cosmétique, FARPI-FRANCE sélectionne pour vous les meilleurs commettants du marché afin de vous apporter des solutions techniques, innovantes et transformer vos projets en une réussite maîtrisée.
An Efficient Approach for Enhancing the Security of Amazigh Text using Binary...Editor IJCATR
Now a day’s Cryptography is one of the broad areas for researchers. Due to its importance, several cryptography techniques
are adopted by many authors to secure the data, but still there is a scope to improve the previous approaches. The main of our research
is to develop a novel Approach for enhancing the security of Amazigh Text using binary tree. The plaintext considered is the
combination of Unicode characters. This paper contributes in the area of elliptic curve cryptography by encrypting data using matrix
approach and using the concept of tree traversal method for enhancing the security of the encrypted points. The security goals were
enhanced by making it difficult for attacker to predicate a pattern as well as speed of the encryption/decryption scheme. The results
show strength of the algorithm.
This is the second lecture in the CS 6212 class. Covers asymptotic notation and data structures. Also outlines the coming lectures wherein we will study the various algorithm design techniques.
Discovering Novel Information with sentence Level clustering From Multi-docu...irjes
Specific objective to discover some novel information from a set of documents initially retrieved in response to some query. Clustering sentences level text, effective use and update is still an open research issue, especially in domain of text mining. Since most existing system uses pattern belong to a single cluster. But here we can use patterns belongs to all cluster with different degree of membership. Since sentences of those documents we would expect at least one of the clusters to be closely related to the concepts described by the query term. This paper presents a Novel Fuzzy Clustering Algorithm that operates on relational input data (i.e. data in the form of square matrix of pair wise similarities between data objects).
This work is proposed the feed forward neural network with symmetric table addition method to design the
neuron synapses algorithm of the sine function approximations, and according to the Taylor series
expansion. Matlab code and LabVIEW are used to build and create the neural network, which has been
designed and trained database set to improve its performance, and gets the best a global convergence with
small value of MSE errors and 97.22% accuracy.
Reduct generation for the incremental data using rough set theorycsandit
n today’s changing world huge amount of data is ge
nerated and transferred frequently.
Although the data is sometimes static but most comm
only it is dynamic and transactional. New
data that is being generated is getting constantly
added to the old/existing data. To discover the
knowledge from this incremental data, one approach
is to run the algorithm repeatedly for the
modified data sets which is time consuming. The pap
er proposes a dimension reduction
algorithm that can be applied in dynamic environmen
t for generation of reduced attribute set as
dynamic reduct.
The method analyzes the new dataset, when it become
s available, and modifies
the reduct accordingly to fit the entire dataset. T
he concepts of discernibility relation, attribute
dependency and attribute significance of Rough Set
Theory are integrated for the generation of
dynamic reduct set, which not only reduces the comp
lexity but also helps to achieve higher
accuracy of the decision system. The proposed metho
d has been applied on few benchmark
dataset collected from the UCI repository and a dyn
amic reduct is computed. Experimental
result shows the efficiency of the proposed method
Paper Explained: Understanding the wiring evolution in differentiable neural ...Devansh16
Read my Explanation of the Paper here: https://medium.com/@devanshverma425/why-and-how-is-neural-architecture-search-is-biased-778763d03f38?sk=e16a3e54d6c26090a6b28f7420d3f6f7
Abstract: Controversy exists on whether differentiable neural architecture search methods discover wiring topology effectively. To understand how wiring topology evolves, we study the underlying mechanism of several existing differentiable NAS frameworks. Our investigation is motivated by three observed searching patterns of differentiable NAS: 1) they search by growing instead of pruning; 2) wider networks are more preferred than deeper ones; 3) no edges are selected in bi-level optimization. To anatomize these phenomena, we propose a unified view on searching algorithms of existing frameworks, transferring the global optimization to local cost minimization. Based on this reformulation, we conduct empirical and theoretical analyses, revealing implicit inductive biases in the cost's assignment mechanism and evolution dynamics that cause the observed phenomena. These biases indicate strong discrimination towards certain topologies. To this end, we pose questions that future differentiable methods for neural wiring discovery need to confront, hoping to evoke a discussion and rethinking on how much bias has been enforced implicitly in existing NAS methods.
X-TREPAN: A MULTI CLASS REGRESSION AND ADAPTED EXTRACTION OF COMPREHENSIBLE D...cscpconf
In this work, the TREPAN algorithm is enhanced and extended for extracting decision trees from neural networks. We empirically evaluated the performance of the algorithm on a set of databases from real world events. This benchmark enhancement was achieved by adapting Single-test TREPAN and C4.5 decision tree induction algorithms to analyze the datasets. The models are then compared with X-TREPAN for comprehensibility and classification accuracy. Furthermore, we validate the experimentations by applying statistical methods. Finally, the modified algorithm is extended to work with multi-class regression problems and the ability to comprehend generalized feed forward networks is achieved.
X-TREPAN : A Multi Class Regression and Adapted Extraction of Comprehensible ...csandit
In this work, the TREPAN algorithm is enhanced and extended for extracting decision trees
from neural networks. We empirically evaluated the performance of the algorithm on a set of
databases from real world events. This benchmark enhancement was achieved by adapting
Single-test TREPAN and C4.5 decision tree induction algorithms to analyze the datasets. The
models are then compared with X-TREPAN for comprehensibility and classification accuracy.
Furthermore, we validate the experimentations by applying statistical methods. Finally, the
modified algorithm is extended to work with multi-class regression problems and the ability to
comprehend generalized feed forward networks is achieved.
Here are my slides for my preparation class for possible Master students in Electrical Engineering and Computer Science (Specialization in Computer Science)... for the entrance examination here at Cinvestav GDL
Similar to An Optimal Approach For Knowledge Protection In Structured Frequent Patterns (20)
The Use of Java Swing’s Components to Develop a WidgetWaqas Tariq
Widget is a kind of application provides a single service such as a map, news feed, simple clock, battery-life indicators, etc. This kind of interactive software object has been developed to facilitate user interface (UI) design. A user interface (UI) function may be implemented using different widgets with the same function. In this article, we present the widget as a platform that is generally used in various applications, such as in desktop, web browser, and mobile phone. We also describe a visual menu of Java Swing’s components that will be used to establish widget. It will assume that we have successfully compiled and run a program that uses Swing components.
3D Human Hand Posture Reconstruction Using a Single 2D ImageWaqas Tariq
Passive sensing of the 3D geometric posture of the human hand has been studied extensively over the past decade. However, these research efforts have been hampered by the computational complexity caused by inverse kinematics and 3D reconstruction. In this paper, our objective focuses on 3D hand posture estimation based on a single 2D image with aim of robotic applications. We introduce the human hand model with 27 degrees of freedom (DOFs) and analyze some of its constraints to reduce the DOFs without any significant degradation of performance. A novel algorithm to estimate the 3D hand posture from eight 2D projected feature points is proposed. Experimental results using real images confirm that our algorithm gives good estimates of the 3D hand pose. Keywords: 3D hand posture estimation; Model-based approach; Gesture recognition; human- computer interface; machine vision.
Camera as Mouse and Keyboard for Handicap Person with Troubleshooting Ability...Waqas Tariq
Camera mouse has been widely used for handicap person to interact with computer. The utmost important of the use of camera mouse is must be able to replace all roles of typical mouse and keyboard. It must be able to provide all mouse click events and keyboard functions (include all shortcut keys) when it is used by handicap person. Also, the use of camera mouse must allow users troubleshooting by themselves. Moreover, it must be able to eliminate neck fatigue effect when it is used during long period. In this paper, we propose camera mouse system with timer as left click event and blinking as right click event. Also, we modify original screen keyboard layout by add two additional buttons (button “drag/ drop” is used to do drag and drop of mouse events and another button is used to call task manager (for troubleshooting)) and change behavior of CTRL, ALT, SHIFT, and CAPS LOCK keys in order to provide shortcut keys of keyboard. Also, we develop recovery method which allows users go from camera and then come back again in order to eliminate neck fatigue effect. The experiments which involve several users have been done in our laboratory. The results show that the use of our camera mouse able to allow users do typing, left and right click events, drag and drop events, and troubleshooting without hand. By implement this system, handicap person can use computer more comfortable and reduce the dryness of eyes.
A Proposed Web Accessibility Framework for the Arab DisabledWaqas Tariq
The Web is providing unprecedented access to information and interaction for people with disabilities. This paper presents a Web accessibility framework which offers the ease of the Web accessing for the disabled Arab users and facilitates their lifelong learning as well. The proposed framework system provides the disabled Arab user with an easy means of access using their mother language so they don’t have to overcome the barrier of learning the target-spoken language. This framework is based on analyzing the web page meta-language, extracting its content and reformulating it in a suitable format for the disabled users. The basic objective of this framework is supporting the equal rights of the Arab disabled people for their access to the education and training with non disabled people. Key Words : Arabic Moon code, Arabic Sign Language, Deaf, Deaf-blind, E-learning Interactivity, Moon code, Web accessibility , Web framework , Web System, WWW.
Real Time Blinking Detection Based on Gabor FilterWaqas Tariq
New method of blinking detection is proposed. The utmost important of blinking detections method is robust against different users, noise, and also change of eye shape. In this paper, we propose blinking detections method by measuring of distance between two arcs of eye (upper part and lower part). We detect eye arcs by apply Gabor filter onto eye image. As we know that Gabor filter has advantage on image processing application since it able to extract spatial localized spectral features, such line, arch, and other shape are more easily detected. After two of eye arcs are detected, we measure the distance between both by using connected labeling method. The open eye is marked by the distance between two arcs is more than threshold and otherwise, the closed eye is marked by the distance less than threshold. The experiment result shows that our proposed method robust enough against different users, noise, and eye shape changes with perfectly accuracy.
Computer Input with Human Eyes-Only Using Two Purkinje Images Which Works in ...Waqas Tariq
A method for computer input with human eyes-only using two Purkinje images which works in a real time basis without calibration is proposed. Experimental results shows that cornea curvature can be estimated by using two light sources derived Purkinje images so that no calibration for reducing person-to-person difference of cornea curvature. It is found that the proposed system allows usersf movements of 30 degrees in roll direction and 15 degrees in pitch direction utilizing detected face attitude which is derived from the face plane consisting three feature points on the face, two eyes and nose or mouth. Also it is found that the proposed system does work in a real time basis.
Toward a More Robust Usability concept with Perceived Enjoyment in the contex...Waqas Tariq
Mobile multimedia service is relatively new but has quickly dominated people¡¯s lives, especially among young people. To explain this popularity, this study applies and modifies the Technology Acceptance Model (TAM) to propose a research model and conduct an empirical study. The goal of study is to examine the role of Perceived Enjoyment (PE) and what determinants can contribute to PE in the context of using mobile multimedia service. The result indicates that PE is influencing on Perceived Usefulness (PU) and Perceived Ease of Use (PEOU) and directly Behavior Intention (BI). Aesthetics and flow are key determinants to explain Perceived Enjoyment (PE) in mobile multimedia usage.
Collaborative Learning of Organisational KnolwedgeWaqas Tariq
This paper presents recent research into methods used in Australian Indigenous Knowledge sharing and looks at how these can support the creation of suitable collaborative envi- ronments for timely organisational learning. The protocols and practices as used today and in the past by Indigenous communities are presented and discussed in relation to their relevance to a personalised system of knowledge sharing in modern organisational cultures. This research focuses on user models, knowledge acquisition and integration of data for constructivist learning in a networked repository of or- ganisational knowledge. The data collected in the repository is searched to provide collections of up-to-date and relevant material for training in a work environment. The aim is to improve knowledge collection and sharing in a team envi- ronment. This knowledge can then be collated into a story or workflow that represents the present knowledge in the organisation.
Our research aims to propose a global approach for specification, design and verification of context awareness Human Computer Interface (HCI). This is a Model Based Design approach (MBD). This methodology describes the ubiquitous environment by ontologies. OWL is the standard used for this purpose. The specification and modeling of Human-Computer Interaction are based on Petri nets (PN). This raises the question of representation of Petri nets with XML. We use for this purpose, the standard of modeling PNML. In this paper, we propose an extension of this standard for specification, generation and verification of HCI. This extension is a methodological approach for the construction of PNML with Petri nets. The design principle uses the concept of composition of elementary structures of Petri nets as PNML Modular. The objective is to obtain a valid interface through verification of properties of elementary Petri nets represented with PNML.
Development of Sign Signal Translation System Based on Altera’s FPGA DE2 BoardWaqas Tariq
The main aim of this paper is to build a system that is capable of detecting and recognizing the hand gesture in an image captured by using a camera. The system is built based on Altera’s FPGA DE2 board, which contains a Nios II soft core processor. Image processing techniques and a simple but effective algorithm are implemented to achieve this purpose. Image processing techniques are used to smooth the image in order to ease the subsequent processes in translating the hand sign signal. The algorithm is built for translating the numerical hand sign signal and the result are displayed on the seven segment display. Altera’s Quartus II, SOPC Builder and Nios II EDS software are used to construct the system. By using SOPC Builder, the related components on the DE2 board can be interconnected easily and orderly compared to traditional method that requires lengthy source code and time consuming. Quartus II is used to compile and download the design to the DE2 board. Then, under Nios II EDS, C programming language is used to code the hand sign translation algorithm. Being able to recognize the hand sign signal from images can helps human in controlling a robot and other applications which require only a simple set of instructions provided a CMOS sensor is included in the system.
An overview on Advanced Research Works on Brain-Computer InterfaceWaqas Tariq
A brain–computer interface (BCI) is a proficient result in the research field of human- computer synergy, where direct articulation between brain and an external device occurs resulting in augmenting, assisting and repairing human cognitive. Advanced works like generating brain-computer interface switch technologies for intermittent (or asynchronous) control in natural environments or developing brain-computer interface by Fuzzy logic Systems or by implementing wavelet theory to drive its efficacies are still going on and some useful results has also been found out. The requirements to develop this brain machine interface is also growing day by day i.e. like neuropsychological rehabilitation, emotion control, etc. An overview on the control theory and some advanced works on the field of brain machine interface are shown in this paper.
Exploring the Relationship Between Mobile Phone and Senior Citizens: A Malays...Waqas Tariq
There is growing ageing phenomena with the rise of ageing population throughout the world. According to the World Health Organization (2002), the growing ageing population indicates 694 million, or 223% is expected for people aged 60 and over, since 1970 and 2025.The growth is especially significant in some advanced countries such as North America, Japan, Italy, Germany, United Kingdom and so forth. This growing older adult population has significantly impact the social-culture, lifestyle, healthcare system, economy, infrastructure and government policy of a nation. However, there are limited research studies on the perception and usage of a mobile phone and its service for senior citizens in a developing nation like Malaysia. This paper explores the relationship between mobile phones and senior citizens in Malaysia from the perspective of a developing country. We conducted an exploratory study using contextual interviews with 5 senior citizens of how they perceive their mobile phones. This paper reveals 4 interesting themes from this preliminary study, in addition to the findings of the desirable mobile requirements for local senior citizens with respect of health, safety and communication purposes. The findings of this study bring interesting insight to local telecommunication industries as a whole, and will also serve as groundwork for more in-depth study in the future.
Principles of Good Screen Design in WebsitesWaqas Tariq
Visual techniques for proper arrangement of the elements on the user screen have helped the designers to make the screen look good and attractive. Several visual techniques emphasize the arrangement and ordering of the screen elements based on particular criteria for best appearance of the screen. This paper investigates few significant visual techniques in various web user interfaces and showcases the results for better understanding and their presence.
Virtual teams are used more and more by companies and other organizations to receive benefits. They are a great way to enable teamwork in situations where people are not sitting in the same physical place at the same time. As companies seek to increase the use of virtual teams, a need exists to explore the context of these teams, the virtuality of a team and software that may help these teams working virtualy. Virtual teams have the same basic principles as traditional teams, but there is one big difference. This difference is the way the team members communicate. Instead of using the dynamics of in-office face-to-face exchange, they now rely on special communication channels enabled by modern technologies, such as e-mails, faxes, phone calls and teleconferences, virtual meetings etc. This is why this paper is focused on the issues regarding virtual teams, and how these teams are created and progressing in Albania.
Cognitive Approach Towards the Maintenance of Web-Sites Through Quality Evalu...Waqas Tariq
It is a well established fact that the Web-Applications require frequent maintenance because of cutting– edge business competitions. The authors have worked on quality evaluation of web-site of Indian ecommerce domain. As a result of that work they have made a quality-wise ranking of these sites. According to their work and also the survey done by various other groups Futurebazaar web-site is considered to be one of the best Indian e-shopping sites. In this research paper the authors are assessing the maintenance of the same site by incorporating the problems incurred during this evaluation. This exercise gives a real world maintainability problem of web-sites. This work will give a clear picture of all the quality metrics which are directly or indirectly related with the maintainability of the web-site.
USEFul: A Framework to Mainstream Web Site Usability through Automated Evalua...Waqas Tariq
A paradox has been observed whereby web site usability is proven to be an essential element in a web site, yet at the same time there exist an abundance of web pages with poor usability. This discrepancy is the result of limitations that are currently preventing web developers in the commercial sector from producing usable web sites. In this paper we propose a framework whose objective is to alleviate this problem by automating certain aspects of the usability evaluation process. Mainstreaming comes as a result of automation, therefore enabling a non-expert in the field of usability to conduct the evaluation. This results in reducing the costs associated with such evaluation. Additionally, the framework allows the flexibility of adding, modifying or deleting guidelines without altering the code that references them since the guidelines and the code are two separate components. A comparison of the evaluation results carried out using the framework against published evaluations of web sites carried out by web site usability professionals reveals that the framework is able to automatically identify the majority of usability violations. Due to the consistency with which it evaluates, it identified additional guideline-related violations that were not identified by the human evaluators.
Robot Arm Utilized Having Meal Support System Based on Computer Input by Huma...Waqas Tariq
A robot arm utilized having meal support system based on computer input by human eyes only is proposed. The proposed system is developed for handicap/disabled persons as well as elderly persons and tested with able persons with several shapes and size of eyes under a variety of illumination conditions. The test results with normal persons show the proposed system does work well for selection of the desired foods and for retrieve the foods as appropriate as usersf requirements. It is found that the proposed system is 21% much faster than the manually controlled robotics.
Dynamic Construction of Telugu Speech Corpus for Voice Enabled Text EditorWaqas Tariq
In recent decades speech interactive systems have gained increasing importance. Performance of an ASR system mainly depends on the availability of large corpus of speech. The conventional method of building a large vocabulary speech recognizer for any language uses a top-down approach to speech. This approach requires large speech corpus with sentence or phoneme level transcription of the speech utterances. The transcriptions must also include different speech order so that the recognizer can build models for all the sounds present. But, for Telugu language, because of its complex nature, a very large, well annotated speech database is very difficult to build. It is very difficult, if not impossible, to cover all the words of any Indian language, where each word may have thousands and millions of word forms. A significant part of grammar that is handled by syntax in English (and other similar languages) is handled within morphology in Telugu. Phrases including several words (that is, tokens) in English would be mapped on to a single word in Telugu.Telugu language is phonetic in nature in addition to rich in morphology. That is why the speech technology developed for English cannot be applied to Telugu language. This paper highlights the work carried out in an attempt to build a voice enabled text editor with capability of automatic term suggestion. Main claim of the paper is the recognition enhancement process developed by us for suitability of highly inflecting, rich morphological languages. This method results in increased speech recognition accuracy with very much reduction in corpus size. It also adapts Telugu words to the database dynamically, resulting in growth of the corpus.
An Improved Approach for Word Ambiguity RemovalWaqas Tariq
Word ambiguity removal is a task of removing ambiguity from a word, i.e. correct sense of word is identified from ambiguous sentences. This paper describes a model that uses Part of Speech tagger and three categories for word sense disambiguation (WSD). Human Computer Interaction is very needful to improve interactions between users and computers. For this, the Supervised and Unsupervised methods are combined. The WSD algorithm is used to find the efficient and accurate sense of a word based on domain information. The accuracy of this work is evaluated with the aim of finding best suitable domain of word. Keywords: Human Computer Interaction, Supervised Training, Unsupervised Learning, Word Ambiguity, Word sense disambiguation
Parameters Optimization for Improving ASR Performance in Adverse Real World N...Waqas Tariq
From the existing research it has been observed that many techniques and methodologies are available for performing every step of Automatic Speech Recognition (ASR) system, but the performance (Minimization of Word Error Recognition-WER and Maximization of Word Accuracy Rate- WAR) of the methodology is not dependent on the only technique applied in that method. The research work indicates that, performance mainly depends on the category of the noise, the level of the noise and the variable size of the window, frame, frame overlap etc is considered in the existing methods. The main aim of the work presented in this paper is to use variable size of parameters like window size, frame size and frame overlap percentage to observe the performance of algorithms for various categories of noise with different levels and also train the system for all size of parameters and category of real world noisy environment to improve the performance of the speech recognition system. This paper presents the results of Signal-to-Noise Ratio (SNR) and Accuracy test by applying variable size of parameters. It is observed that, it is really very hard to evaluate test results and decide parameter size for ASR performance improvement for its resultant optimization. Hence, this study further suggests the feasible and optimum parameter size using Fuzzy Inference System (FIS) for enhancing resultant accuracy in adverse real world noisy environmental conditions. This work will be helpful to give discriminative training of ubiquitous ASR system for better Human Computer Interaction (HCI). Keywords: ASR Performance, ASR Parameters Optimization, Multi-Environmental Training, Fuzzy Inference System for ASR, ubiquitous ASR system, Human Computer Interaction (HCI)
Students, digital devices and success - Andreas Schleicher - 27 May 2024..pptxEduSkills OECD
Andreas Schleicher presents at the OECD webinar ‘Digital devices in schools: detrimental distraction or secret to success?’ on 27 May 2024. The presentation was based on findings from PISA 2022 results and the webinar helped launch the PISA in Focus ‘Managing screen time: How to protect and equip students against distraction’ https://www.oecd-ilibrary.org/education/managing-screen-time_7c225af4-en and the OECD Education Policy Perspective ‘Students, digital devices and success’ can be found here - https://oe.cd/il/5yV
Instructions for Submissions thorugh G- Classroom.pptxJheel Barad
This presentation provides a briefing on how to upload submissions and documents in Google Classroom. It was prepared as part of an orientation for new Sainik School in-service teacher trainees. As a training officer, my goal is to ensure that you are comfortable and proficient with this essential tool for managing assignments and fostering student engagement.
This is a presentation by Dada Robert in a Your Skill Boost masterclass organised by the Excellence Foundation for South Sudan (EFSS) on Saturday, the 25th and Sunday, the 26th of May 2024.
He discussed the concept of quality improvement, emphasizing its applicability to various aspects of life, including personal, project, and program improvements. He defined quality as doing the right thing at the right time in the right way to achieve the best possible results and discussed the concept of the "gap" between what we know and what we do, and how this gap represents the areas we need to improve. He explained the scientific approach to quality improvement, which involves systematic performance analysis, testing and learning, and implementing change ideas. He also highlighted the importance of client focus and a team approach to quality improvement.
The French Revolution, which began in 1789, was a period of radical social and political upheaval in France. It marked the decline of absolute monarchies, the rise of secular and democratic republics, and the eventual rise of Napoleon Bonaparte. This revolutionary period is crucial in understanding the transition from feudalism to modernity in Europe.
For more information, visit-www.vavaclasses.com
Read| The latest issue of The Challenger is here! We are thrilled to announce that our school paper has qualified for the NATIONAL SCHOOLS PRESS CONFERENCE (NSPC) 2024. Thank you for your unwavering support and trust. Dive into the stories that made us stand out!
We all have good and bad thoughts from time to time and situation to situation. We are bombarded daily with spiraling thoughts(both negative and positive) creating all-consuming feel , making us difficult to manage with associated suffering. Good thoughts are like our Mob Signal (Positive thought) amidst noise(negative thought) in the atmosphere. Negative thoughts like noise outweigh positive thoughts. These thoughts often create unwanted confusion, trouble, stress and frustration in our mind as well as chaos in our physical world. Negative thoughts are also known as “distorted thinking”.
2024.06.01 Introducing a competency framework for languag learning materials ...Sandy Millin
http://sandymillin.wordpress.com/iateflwebinar2024
Published classroom materials form the basis of syllabuses, drive teacher professional development, and have a potentially huge influence on learners, teachers and education systems. All teachers also create their own materials, whether a few sentences on a blackboard, a highly-structured fully-realised online course, or anything in between. Despite this, the knowledge and skills needed to create effective language learning materials are rarely part of teacher training, and are mostly learnt by trial and error.
Knowledge and skills frameworks, generally called competency frameworks, for ELT teachers, trainers and managers have existed for a few years now. However, until I created one for my MA dissertation, there wasn’t one drawing together what we need to know and do to be able to effectively produce language learning materials.
This webinar will introduce you to my framework, highlighting the key competencies I identified from my research. It will also show how anybody involved in language teaching (any language, not just English!), teacher training, managing schools or developing language learning materials can benefit from using the framework.
The Roman Empire A Historical Colossus.pdfkaushalkr1407
The Roman Empire, a vast and enduring power, stands as one of history's most remarkable civilizations, leaving an indelible imprint on the world. It emerged from the Roman Republic, transitioning into an imperial powerhouse under the leadership of Augustus Caesar in 27 BCE. This transformation marked the beginning of an era defined by unprecedented territorial expansion, architectural marvels, and profound cultural influence.
The empire's roots lie in the city of Rome, founded, according to legend, by Romulus in 753 BCE. Over centuries, Rome evolved from a small settlement to a formidable republic, characterized by a complex political system with elected officials and checks on power. However, internal strife, class conflicts, and military ambitions paved the way for the end of the Republic. Julius Caesar’s dictatorship and subsequent assassination in 44 BCE created a power vacuum, leading to a civil war. Octavian, later Augustus, emerged victorious, heralding the Roman Empire’s birth.
Under Augustus, the empire experienced the Pax Romana, a 200-year period of relative peace and stability. Augustus reformed the military, established efficient administrative systems, and initiated grand construction projects. The empire's borders expanded, encompassing territories from Britain to Egypt and from Spain to the Euphrates. Roman legions, renowned for their discipline and engineering prowess, secured and maintained these vast territories, building roads, fortifications, and cities that facilitated control and integration.
The Roman Empire’s society was hierarchical, with a rigid class system. At the top were the patricians, wealthy elites who held significant political power. Below them were the plebeians, free citizens with limited political influence, and the vast numbers of slaves who formed the backbone of the economy. The family unit was central, governed by the paterfamilias, the male head who held absolute authority.
Culturally, the Romans were eclectic, absorbing and adapting elements from the civilizations they encountered, particularly the Greeks. Roman art, literature, and philosophy reflected this synthesis, creating a rich cultural tapestry. Latin, the Roman language, became the lingua franca of the Western world, influencing numerous modern languages.
Roman architecture and engineering achievements were monumental. They perfected the arch, vault, and dome, constructing enduring structures like the Colosseum, Pantheon, and aqueducts. These engineering marvels not only showcased Roman ingenuity but also served practical purposes, from public entertainment to water supply.
Unit 8 - Information and Communication Technology (Paper I).pdfThiyagu K
This slides describes the basic concepts of ICT, basics of Email, Emerging Technology and Digital Initiatives in Education. This presentations aligns with the UGC Paper I syllabus.
Unit 8 - Information and Communication Technology (Paper I).pdf
An Optimal Approach For Knowledge Protection In Structured Frequent Patterns
1. Cynthia Selvi P & Mohamed Shanavas A.R
International Journal of Data Engineering (IJDE), Volume (5) : Issue (3) : 2014 22
An Optimal Approach For Knowledge Protection In Structured
Frequent Patterns
Cynthia Selvi P pselvi1501@gmail.com
Associate Professor, Dept. of Computer Science
Kunthavai Naacchiyar Govt. Arts College for Women(Autonomous), Thanjavur 613007
Affiliated to Bharathidasan University, Tiruchirapalli, TamilNadu, India.
Mohamed Shanavas A.R vas0699@yahoo.co.in
Associate Professor, Dept. of Computer Science,
Jamal Mohamed College, Tiruchirapalli 620 020
Affiliated to Bharathidasan University, Tiruchirapalli, TamilNadu, India.
Abstract
Data mining is valuable technology to facilitate the extraction of useful patterns and trends from
large volume of data. When these patterns are to be shared in a collaborative environment, they
must be protectively shared among the parties concerned in order to preserve the confidentiality
of the sensitive data. Sharing of information may be in the form of datasets or in any of the
structured patterns like trees, graphs, lattices, etc., This paper propose a sanitization algorithm for
protecting sensitive data in a structured frequent pattern(tree).
Keywords: Rank Function, Restricted Node, Sanitization, Structured Pattern, Victim States.
1. INTRODUCTION
Data mining is an emerging technology to provide various means for identifying the interesting
and important knowledge from large data collections. When this knowledge is to be shared
among various parties in decision making activities, the sensitive data is to be preserved by the
parties concerned. In particular, when multiple companies want to share the customer’s buying
behavior in a collaborative business environment that promote business, the sensitive information
of the individual company should be protected against sharing. The information to be shared may
be in the form of datasets, frequent itemsets, structured patterns or subsequences. Here
structured pattern refers to substructures like graphs, trees or lattices that contain frequent
itemsets[1]. Various approaches have been proposed so far to address this problem of preserving
sensitive patterns. This paper propose an algorithm that is aimed to sanitize sensitive information
in a frequent pattern tree(because trees exhibits the relationships among the itemsets more
clearly) which leaves no trace for the counterpart or an adversary to extract the hidden
information back, by blocking all possible inferences.
In this article, section-II briefs the literature review; section-III states the definitions needed for the
sanitization approach and algorithm presented in this article and section-IV gives the proposed
algorithm. In section-V illustration with sample graphs are given and in section-VI the
performance metrics are discussed with sample results.
2. LITERATURE REVIEW
Due to wide applicability of the field data mining and in particular for the task of association rules,
the focus has been more specific for the problem of protection of sensitive knowledge against
inference and it has been addressed by various researchers[2-12]. This task is referred to as
sanitization in [2] which blocks inference of sensitive rules that facilitate collaborators to mine
independently their own data and then sharing some of the resulting patterns. The above work
2. Cynthia Selvi P & Mohamed Shanavas A.R
International Journal of Data Engineering (IJDE), Volume (5) : Issue (3) : 2014 23
concentrate on hiding frequent itemsets in databases based on the support and/or confidence
framework. In many situations, it would be more comfortable to share the information in the form
of structured patterns like graphs, trees, lattices, etc instead of sharing the entire databases.
In structured patterns when a particular sensitive pattern is to be removed, its supersets and
subsets should also be removed in order to block the forward and backward references. The work
presented in [7], proposes an algorithm(DSA) that sanitize sensitive information in Graphs and
have compared the efficiency with that of Naïve approach. Naïve blocks only the forward
inference attack; but DSA blocks both forward and backward inference attacks. But in DSA, the
subsets of the sensitive pattern are chosen at random. In this situation, possibility is there for the
removal of more number of patterns; because, when a subset pattern is removed, the other
patterns associated with it would also be removed failing which would leave forward trace for the
counterpart to infer the details of the hidden pattern. Hence, this random removal would reduce
the data utility of the source dataset. To overcome this problem, the work proposed in this article
presents an algorithm(RSS) for sanitizing sensitive information in structured pattern tree that use
a rank function for reducing the computational complexity and legitimate information loss. In
comparison with DSA, this algorithm completely blocks the forward and backward inference
attacks by removing the sensitive information and its associated information in an optimized way
by means of the rank function.
3. BASIC DEFINITIONS
Tree: A Tree is a finite set of one or more nodes such that there is a specially designated node
called the root and the remaining nodes are partitioned into n ≥ 0 disjoint sets T1, …Tn, where
each of these nodes is a tree. The sets T1,..Tn are called the subtrees of the root[13].
Set-Enumeration (SE)-tree: It is a tool for representing and/or enumerating sets in a best-first
fashion. The complete SE-tree systematically enumerates elements of a power-set using a pre-
imposed order on the underlying set of elements.
Structured Pattern Tree: A structured pattern tree denoted by T=(N, L) consists of nonempty set
of frequent itemsets N, a set of links L that are ordered pairs of the elements of N such that for all
nodes a, b Є N there is a link from a to b if a ∩ b = a and |b| - |a| = 1, where |x| is the size of the
itemset x.
Level: Let T=(N, L) be a structured pattern(frequent itemset) tree. The level k of an itemset x such
that x Є N, is the length of the path connecting an 1-itemset (usually al level-0) to x.
Height: Let T=(N, L) be a structured pattern tree. The height, h of T is the length of the maximum
path connecting an 1-itemset a with any other itemset b, such that a,b Є N and a⊂ b.
Delete: The deletion of a node x from Ti, is denoted as Del(x). The resulting Ti
’
is the same as Ti
without the node x. In particular, if p1,..,pm , x, s1, ..., sn is the sibling sequence in a level of Ti, then
p1, ...,pm , s1, ..., sn is the sibling sequence in Ti
’
.
Negative Border Nodes: Negative border nodes possess the property of having all its members
(proper subsets) are frequent.
Problem state: Each node in the tree is a problem state.
R-nodes: Nodes that are sensitive and to be restricted before allowing the structured pattern to be
shared.
P-nodes: Predecessor node(subsets) of R-nodes which are to be identified (in order to block
forward inference) before selecting the particular nodes that are to be deleted.
3. Cynthia Selvi P & Mohamed Shanavas A.R
International Journal of Data Engineering (IJDE), Volume (5) : Issue (3) : 2014 24
S-nodes: Successor nodes(supersets) of R-nodes which are to be identified (in order to block
backward inference) before selecting the particular nodes that are to be deleted.
Victim states: Problem states for which the path from P-node(s) at level-1(containing 2-itemsets)
to R-node and/or to S-node(s) are to be searched to select nodes for deletion.
V-node(s): Node(s) to be deleted selectively (based on rank function) among the victim states.
Rank function - r(.): Choose P-node (of R-node) which leads to only one S-node (with single
Primary Link); choose one at random when tie occurs. The search for victim nodes can often be
speeded up by using the ranking function r(.) for all P-nodes. The ideal way to assign ranks would
be on the basis of minimum additional computational cost needed when this P-node is to be
removed.
4. ALGORITHM
Rank-based Structured-pattern Sanitization(RSS):
Input: Frequent Pattern Tree(T), Set of Restricted Nodes(R-nodes)
Output: Sanitized Tree(T’)
Begin
Obtain height h of the input tree;
identify ri Є R ( R-nodes to be restricted);
//Select victim nodes//
for each ri Є R
{
find level k;
V_nodes[ ] ← ri ;
if k > 0
{
do while(k<=h)
{
obtain S-nodes of ri (ie supersets);
V_nodes[ ] ← V_nodes[ ]+S-nodes(ri);
}
do while(k>=1)
{
obtain P-nodes of ri (ie subsets );
V_nodes[ ] ← V_nodes[ ]+P-node that satisfy r(.);
}
}
delete V_nodes[ ];
}
T’←T;
End
4. Cynthia Selvi P & Mohamed Shanavas A.R
International Journal of Data Engineering (IJDE), Volume (5) : Issue (3) : 2014 25
5. ILLUSTRATION
A sample frequent pattern SE-tree is given in Fig.1. Let the node to be protected(R-node) is the
one with itemset abc(dark-filled in Fig.2); When this is marked for deletion(V-nodes),
conventionaly it becomes infrequent. As per antimonotone property of frequent itemsets, if a set
cannot pass a test, all of its supersets will fail the same test as well. Hence all of its supersets(S-
nodes) are to be identified and deleted until the level equals the height of the tree. In this
example, node abcd (shaded) is the superset of abc and so it is marked for deletion(V-nodes).
Deletion of R-node and its supersets may completely hide the details of the sensitive
data(Restricted nodes) and this ensures the blocking of backward inference of R-node.
Morever, the negative border nodes are also to be deleted to completely block the future
inference of the sensitive data. This can be achieved by identifying the Predecessor nodes(P-
node) of R-node and suitably removing them by means of rank function-r(.) defined earlier. In this
example, abc has two P-nodes, ab and ac which are having primarly links and of them as
ac(shaded) has only one primary link, it is marked for deletion(V-nodes) with all its successors(in
this case, acd). Refer Fig.2.
Finally delete all victim nodes and thus the sanitized frequent pattern tree to be shared is resulted
(the one given Fig.3) and hence forward inference is also blocked.
On the contrary, if ab would have been chosen as victim node, then three more nodes would
have been additionally removed which would result in more information loss and utility loss.
Hence the rank function used in this approach sanitizes the structured frequent pattern tree with
reduced information loss and utility loss.
However, the nodes at level-0 (1-itemsets) are not deleted in any way and this preserves the
distinct items in the given structured frequent pattern tree.
FIGURE 1: Frequent Pattern Tree before Sanitization.
a b c d e
ab ac ad bc bd cd de
abc abd acd ade bde
be
abcd abde
ɸ
bcd
ae
abe
Primary Link Secondary Link
5. Cynthia Selvi P & Mohamed Shanavas A.R
International Journal of Data Engineering (IJDE), Volume (5) : Issue (3) : 2014 26
a b c d e
ab ad bc bd cd de
abd ade bde
be
abde
ɸ
bcd
ae
abe
FIGURE 3: Frequent Pattern Tree after Sanitization.
a b c d e
ab ac ad bc bd cd de
abc abd acd ade bde
be
abcd abde
ɸ
bcd
ae
abe
FIGURE 2: Frequent Pattern Tree with Victim Nodes.
6. Cynthia Selvi P & Mohamed Shanavas A.R
International Journal of Data Engineering (IJDE), Volume (5) : Issue (3) : 2014 27
6. EXPERIMENTAL ANALYSIS
The algorithm was tested for real dataset T10I4D100K[14] with number of transactions ranging
from 1K to 10K and number of restricted nodes from 1 to 5. The test run was made on Intel core
i5 processor with 2.3 GHz speed and 4GB RAM operating on 32 bit OS; The implementation of
the proposed algorithm was done with windows 7 - Netbeans 6.9.1 - SQL 2005. The frequent
patterns were obtained using Matrix Apriori[15], which requires only two scans of original
database and uses simpler data structures.
The efficiency of this approach is studied based on the measures given below and it has been
compared (Figures 4 to 7) with the previously proposed algorithms IMA, PMA, TMA[9-12] which
sanitizes the sensitive patterns(itemsets) in the source datasets.
Dissimilarity(dif) : The dissimilarity between the original(D) and sanitized(D’) databases is
measured in terms of their contents which can be measured by the formula,
dif(D, D’) = x
where fx(i) represents the i
th
item in the dataset X. This approach has very low percentage of
dissimilarity and this shows that information loss is very low and so the utility is well preserved.
From the fig.4 &5, it is observed that the proposed algorithm, RSS has very low dissimilarity in
comparison with previous algorithms. However, when the no. of transactions are increased, the
dissimilarity gets increased; this is due to the removal of subsets(with its associated nodes) of the
sensitive nodes for blocking backward inference attack and it is observed to be less than 5%.
CPU Time: The execution time is tested for the proposed algorithm by varying the number of
nodes to be restricted. Fig.6 & 7 shows that the execution time required for RSS algorithm is low
in comparison with the other algorithms. It is also observed that execution time is minimum, when
the no. of transactions in the source dataset is more. However, time is not a significant criteria as
the sanitization is done offline.
FIGURE 4: Dissimilarity
(varying no.of rules).
FIGURE 5: Dissimilarity
(varying no.of transactions).
FIGURE 6: Execution Time
(varying no.of rules).
FIGURE 7: Execution Time
(varying no.of transactions).
7. Cynthia Selvi P & Mohamed Shanavas A.R
International Journal of Data Engineering (IJDE), Volume (5) : Issue (3) : 2014 28
Scalability: In order to effectively hide sensitive knowledge in patterns, the sanitization
algorithms must be efficient and scalable which means the running time of any sanitizing
algorithm must be predictable and acceptable. The efficiency and scalability of the proposed
approach is proved below:
Theorem: The running time of the Rank-based Structured-pattern Sanitization(RSS) approach is
at least O[r(l+s)]; where r is the number of restrictive nodes(R-nodes), l is the number of
preceding levels that have subsets of R-nodes and s is the number of S-nodes(supersets) of
R-nodes.
Proof : Let T be a given Structured frequent pattern tree with N being the total number of nodes in
T; r be the number of sensitive nodes(R-nodes) to be restricted among N; l be the number of
preceding levels of R-nodes and s be the number of S-nodes(supersets) of R-nodes in T.
The proposed approach finds the height of the given tree. For every given R-node, find the victim
states which are the collection of its S-nodes(supersets) and P-nodes(subsets) that lead with only
one primary link for their own successors. As this approach satisfies anti-monotone property, all s
S-nodes are victim nodes and to be deleted to block the backward inferences. However
among the P-nodes(subsets), at each preceding level (other than level-0) the node(subset)
which forms as a single primary link for its successors is to be obtained and deleted (with all its
successors) in order to block all forward inferences; this selection process is quiet straightforward
and it gets repeated for all R-nodes.
This algorithm makes use of both depth-wise and breadth-wise search which requires atleast
O(l+s) computational complexity for every R-node.
Hence, the running time of proposed algorithm for k R-nodes is atleast O[r(l+s)], which is linear
and better than O(n
2
), O(n
3
), O(2
n
), O(n log N).
7. CONCLUSION
The proposed algorithm in this work sanitizes the structured frequent pattern tree in an optimal
way, by using a rank function that reduces the computational complexity as well as the
information loss and utility loss. Moreover, this approach blocks all the inference channels of the
restrictive patterns in both forward and backward directions leaving no trace of the nodes that are
restricted(removed) before sharing. This simulation process facilitates the task of sanitizing the
structured pattern with different set of restricted information when it is to be shared between
different set of collaborators. However, when the database is large, it is sometimes unrealistic to
construct a main-memory based pattern tree. The proposed algorithm sanitizes patterns in static
dataset and also the sanitization is done offline due to the offline decision analysis of the
restricted rules. But further effort is being taken to apply optimized heuristic approach to sanitize
continuous and dynamic dataset.
8. REFERENCES
[1] J.Han, M.Kamber, Data Mining Concepts and Techniches, Oxford University Press, 2009.
[2] M.Atallah, E.Bertino, A.Elmagarmid, M.Ibrahim and V.Verykios “Disclosure Limitation of
Sensitive Rules”, Proc. of IEEE Knowledge and Data Engineering Workshop, pages 45–52,
Chicago, Illinois, Nov 1999.
[3] E.Dasseni, V.S.Verykios, A.K.Elmagarmid & E.Bertino, “Hiding Association Rules by Using
Confidence and Support”, Proc. of the 4th Information Hiding Workshop, pages 369– 383,
Pittsburg, PA, Apr 2001.
[4] Y.Saygin, V.S.Verykios, and C.Clifton, “Using Unknowns to Prevent Discovery of Association
Rules”, SIGMOD Record, 30(4):45–54, Dec 2001.
8. Cynthia Selvi P & Mohamed Shanavas A.R
International Journal of Data Engineering (IJDE), Volume (5) : Issue (3) : 2014 29
[5] S.R.M.Oliveira, and O.R.Zaiane, “Privacy preserving Frequent Itemset Mining”, Proc. of the
IEEE ICDM Workshop on Privacy, Security, and Data Mining, Pages 43-54, Maebashi City,
Japan, Dec 2002.
[6] S.R.M.Oliveira, and O.R.Zaiane, “An Efficient One-Scan Sanitization for Improving the
Balance between Privacy and Knowledge Discovery”, Technical Report TR 03-15, Jun 2003.
[7] S.R.M.Oliveira, and O.R.Zaiane, “Secure Association Rule Mining”, Proc. of the 8
th
Pacific-
Asia Conference on Knowledge Discovery and Data Mining(PAKDD’04), Pages 74-85,
Sydney, Australia, May 2004..
[8] B.Yildz, and B.Ergenc, “Hiding Sensitive Predictive Frequent Itemsets”, Proc. of the
International MultiConference of Engineers and Computer Scientists 2011, Vol-I.
[9] P.Cynthia Selvi, A.R.Mohamed Shanavas, “An effective Heuristic Approach for Hiding
Sensitive Patterns in Databases”, International Organization of Scientific Research-Journal
of Computer Engineering(IOSRJCE) Vol. 5, Issue 1(Sep-Oct 2012), PP 06-11.
[10] P.Cynthia Selvi, A.R.Mohamed Shanavas, “An Improved Item-based Maxcover Algorithm to
protect Sensitive Patterns in Large Databases”, International Organization of Scientific
Research-Journal of Computer Engineering(IOSRJCE) Vol.14, Issue 4, Oct 2013, Pages 1-5.
[11] P.Cynthia Selvi, A.R.Mohamed Shanavas, “Output Privacy Protection With Pattern-Based
Heuristic Algorithm”, International Journal of Computer Science & Information
Technology(IJCSIT) Vol 6, No 2, Apr 2014, Pages 141 – 152.
[12] P.Cynthia Selvi, A.R.Mohamed Shanavas, “Towards Information Privacy Using Transaction-
Based Maxcover Algorithm”, World Applied Sciences Journal 29 (Data Mining and Soft
Computing Techniques): 06-11, 2014.
[13] Ellis Horowitz, Sartaj Sahni, Sanguthevar Rajasekaran, Fundamentals of Computer
Algorithms, Galgotia Pub. Pvt. Ltd, Delhi, 1999.
[14] The Dataset used in this work for experimental analysis was generated using the generator
from IBM Almaden Quest research group and is publicly available from
http://fimi.ua.ac.be/data/.
[15] J.Pavon, S.Viana, S.Gomez, “Matrix Apriori: speeding up the search for frequent patterns”,
Proc. 24th IASTED International Conference on Databases and Applications 2006, pp. 75-82.