The paper presents design of computer system application for the forest area estimation using the combination of genetic
algorithm (GA) and support vector machine SVM) methods in Kutai National park of East Borneo. The considering variables such as
reboization concerning natural green, forest fire, encroachment and illegal logging activities are the basic data for our proposed design. In
the development of design computer system application, the Unified Modeling Language is adopted with the use case, activity diagrams and
sequence diagrams are systematically followed in order to keep the purpose of design on track. The supporting instrument for this research is
the language programming of Borland Delphi 7 and MySQL database. The accuracy of area estimation result is compared with the actual
data using mean absolute percentage error (MAPE).
A BINARY BAT INSPIRED ALGORITHM FOR THE CLASSIFICATION OF BREAST CANCER DATAIJSCAI Journal
Advancement in information and technology has made a major impact on medical science where the
researchers come up with new ideas for improving the classification rate of various diseases. Breast cancer
is one such disease killing large number of people around the world. Diagnosing the disease at its earliest
instance makes a huge impact on its treatment. The authors propose a Binary Bat Algorithm (BBA) based
Feedforward Neural Network (FNN) hybrid model, where the advantages of BBA and efficiency of FNN is
exploited for the classification of three benchmark breast cancer datasets into malignant and benign cases.
Here BBA is used to generate a V-shaped hyperbolic tangent function for training the network and a fitness
function is used for error minimization. FNNBBA based classification produces 92.61% accuracy for
training data and 89.95% for testing data.
Pest Control in Agricultural Plantations Using Image ProcessingIOSR Journals
Abstract: Monocropped plantations are unique to India and a handful of countries throughout the globe. Essentially, the FOREST approach of growing coffee along with in India has enabled the plantation to fight many outbreaks of pests and diseases. Mono cropped Plantations are under constant threat of pest and disease incidence because it favours the build up of pest population. To cope with these problems, an automatic pest detection algorithm using image processing techniques in MATLAB has been proposed in this paper. Image acquisition devices are used to acquire images of plantations at regular intervals. These images are then subjected to pre-processing, transformation and clustering.
A Binary Bat Inspired Algorithm for the Classification of Breast Cancer Data ijscai
This document summarizes a research paper that proposes using a binary bat algorithm to classify breast cancer data. The researchers developed a hybrid model combining a binary bat algorithm and feedforward neural network. The binary bat algorithm was used to generate a activation function for training the neural network and minimize error. Testing of the model on three breast cancer datasets produced an accuracy of 92.61% for training data and 89.95% for testing data, showing potential for classifying breast cancer as malignant or benign.
The document summarizes a study that mapped forest types and plantation works in Bidar Forest Division, Karnataka, India using remote sensing and GIS techniques. Satellite imagery was analyzed which revealed that 84% of the area is covered by crop land, 1.778% by dry deciduous forest, and 1.38% by mixed plantation. Maps were generated showing forest type distribution and plantation works during 2011-2012 and planned for 2012-2013. Around 12 square kilometers were planted in 2011-2012, with most in Humnbad range and least in Aurad range. Proper forest management can be achieved using geoinformatics technologies.
A STUDY ON WEED DISCRIMINATION THROUGH WAVELET TRANSFORM, TEXTURE FEATURE EXT...ijcsit
Texture based weed classification has played an important role in agricultural applications. In the recent years weed classification based on wavelet transform is an effective method. But the feature extraction is main issue for proper classification of weed species. In this paper, the issue of statistical and texture
classification based on wavelet transform has been analysed. The efficient texture feature extraction
methods are developed for weed discrimination. Three group feature vector can be constructed by the mean
and standard deviation of the wavelet statistical features (WSF), Texture feature as Contrast, Cluster
Shade, Cluster Prominence and Local Homogeneity (WCSPH) and Energy, Correlation, Cluster Shade,
Cluster Prominence and Entropy features (WECSPE) which are derived from the sub-bands of the wavelet
decomposition and are used for classification. Experimental results show that Rbio33 Wavelet with
WECSPE texture feature obtaining high degree of success rate in classification.
SVM-PSO based Feature Selection for Improving Medical Diagnosis Reliability u...cscpconf
Improving accuracy of supervised classification algorithms in biomedical applications,
especially CADx, is one of active area of research. This paper proposes construction of rotation
forest (RF) ensemble using 20 learners over two clinical datasets namely lymphography and
backache. We propose a new feature selection strategy based on support vector machines
optimized by particle swarm optimization for relevant and minimum feature subset for obtaining
higher accuracy of ensembles. We have quantitatively analyzed 20 base learners over two
datasets and carried out the experiments with 10 fold cross validation leave-one-out strategy
and the performance of 20 classifiers are evaluated using performance metrics namely accuracy
(acc), kappa value (K), root mean square error (RMSE) and area under receiver operating
characteristics curve (ROC). Base classifiers succeeded 79.96% & 81.71% average accuracies
for lymphography & backache datasets respectively. As for RF ensembles, they produced
average accuracies of 83.72% & 85.77% for respective diseases. The paper presents promising
results using RF ensembles and provides a new direction towards construction of reliable and robust medical diagnosis systems.
Hybrid prediction model with missing value imputation for medical data 2015-g...Jitender Grover
The document presents a novel hybrid prediction model called HPM-MI that uses K-means clustering and multilayer perceptron (MLP) to improve predictive classification for medical data with missing values. The model first analyzes 11 different imputation techniques using K-means clustering to select the best one for filling missing values in the data. It then uses K-means clustering again to validate class labels and remove incorrectly classified instances before applying the MLP classifier. The model is tested on three medical datasets from the UCI repository and shows improved accuracy, sensitivity, specificity and other metrics compared to existing methods, particularly when datasets have large numbers of missing values.
This document summarizes an article that proposes using genetic algorithms to improve the effectiveness of a text to matrix generator. It begins with an overview of information retrieval and discusses the vector space model and genetic algorithms. It then proposes a genetic approach to optimize the objective function of a text to matrix generator to increase the average number of terms. The goal is to retrieve more relevant documents by obtaining the best combination of terms from document collections using genetic algorithms. Experimental results are presented to validate that the genetic approach improves the performance of the text to matrix generator.
A BINARY BAT INSPIRED ALGORITHM FOR THE CLASSIFICATION OF BREAST CANCER DATAIJSCAI Journal
Advancement in information and technology has made a major impact on medical science where the
researchers come up with new ideas for improving the classification rate of various diseases. Breast cancer
is one such disease killing large number of people around the world. Diagnosing the disease at its earliest
instance makes a huge impact on its treatment. The authors propose a Binary Bat Algorithm (BBA) based
Feedforward Neural Network (FNN) hybrid model, where the advantages of BBA and efficiency of FNN is
exploited for the classification of three benchmark breast cancer datasets into malignant and benign cases.
Here BBA is used to generate a V-shaped hyperbolic tangent function for training the network and a fitness
function is used for error minimization. FNNBBA based classification produces 92.61% accuracy for
training data and 89.95% for testing data.
Pest Control in Agricultural Plantations Using Image ProcessingIOSR Journals
Abstract: Monocropped plantations are unique to India and a handful of countries throughout the globe. Essentially, the FOREST approach of growing coffee along with in India has enabled the plantation to fight many outbreaks of pests and diseases. Mono cropped Plantations are under constant threat of pest and disease incidence because it favours the build up of pest population. To cope with these problems, an automatic pest detection algorithm using image processing techniques in MATLAB has been proposed in this paper. Image acquisition devices are used to acquire images of plantations at regular intervals. These images are then subjected to pre-processing, transformation and clustering.
A Binary Bat Inspired Algorithm for the Classification of Breast Cancer Data ijscai
This document summarizes a research paper that proposes using a binary bat algorithm to classify breast cancer data. The researchers developed a hybrid model combining a binary bat algorithm and feedforward neural network. The binary bat algorithm was used to generate a activation function for training the neural network and minimize error. Testing of the model on three breast cancer datasets produced an accuracy of 92.61% for training data and 89.95% for testing data, showing potential for classifying breast cancer as malignant or benign.
The document summarizes a study that mapped forest types and plantation works in Bidar Forest Division, Karnataka, India using remote sensing and GIS techniques. Satellite imagery was analyzed which revealed that 84% of the area is covered by crop land, 1.778% by dry deciduous forest, and 1.38% by mixed plantation. Maps were generated showing forest type distribution and plantation works during 2011-2012 and planned for 2012-2013. Around 12 square kilometers were planted in 2011-2012, with most in Humnbad range and least in Aurad range. Proper forest management can be achieved using geoinformatics technologies.
A STUDY ON WEED DISCRIMINATION THROUGH WAVELET TRANSFORM, TEXTURE FEATURE EXT...ijcsit
Texture based weed classification has played an important role in agricultural applications. In the recent years weed classification based on wavelet transform is an effective method. But the feature extraction is main issue for proper classification of weed species. In this paper, the issue of statistical and texture
classification based on wavelet transform has been analysed. The efficient texture feature extraction
methods are developed for weed discrimination. Three group feature vector can be constructed by the mean
and standard deviation of the wavelet statistical features (WSF), Texture feature as Contrast, Cluster
Shade, Cluster Prominence and Local Homogeneity (WCSPH) and Energy, Correlation, Cluster Shade,
Cluster Prominence and Entropy features (WECSPE) which are derived from the sub-bands of the wavelet
decomposition and are used for classification. Experimental results show that Rbio33 Wavelet with
WECSPE texture feature obtaining high degree of success rate in classification.
SVM-PSO based Feature Selection for Improving Medical Diagnosis Reliability u...cscpconf
Improving accuracy of supervised classification algorithms in biomedical applications,
especially CADx, is one of active area of research. This paper proposes construction of rotation
forest (RF) ensemble using 20 learners over two clinical datasets namely lymphography and
backache. We propose a new feature selection strategy based on support vector machines
optimized by particle swarm optimization for relevant and minimum feature subset for obtaining
higher accuracy of ensembles. We have quantitatively analyzed 20 base learners over two
datasets and carried out the experiments with 10 fold cross validation leave-one-out strategy
and the performance of 20 classifiers are evaluated using performance metrics namely accuracy
(acc), kappa value (K), root mean square error (RMSE) and area under receiver operating
characteristics curve (ROC). Base classifiers succeeded 79.96% & 81.71% average accuracies
for lymphography & backache datasets respectively. As for RF ensembles, they produced
average accuracies of 83.72% & 85.77% for respective diseases. The paper presents promising
results using RF ensembles and provides a new direction towards construction of reliable and robust medical diagnosis systems.
Hybrid prediction model with missing value imputation for medical data 2015-g...Jitender Grover
The document presents a novel hybrid prediction model called HPM-MI that uses K-means clustering and multilayer perceptron (MLP) to improve predictive classification for medical data with missing values. The model first analyzes 11 different imputation techniques using K-means clustering to select the best one for filling missing values in the data. It then uses K-means clustering again to validate class labels and remove incorrectly classified instances before applying the MLP classifier. The model is tested on three medical datasets from the UCI repository and shows improved accuracy, sensitivity, specificity and other metrics compared to existing methods, particularly when datasets have large numbers of missing values.
This document summarizes an article that proposes using genetic algorithms to improve the effectiveness of a text to matrix generator. It begins with an overview of information retrieval and discusses the vector space model and genetic algorithms. It then proposes a genetic approach to optimize the objective function of a text to matrix generator to increase the average number of terms. The goal is to retrieve more relevant documents by obtaining the best combination of terms from document collections using genetic algorithms. Experimental results are presented to validate that the genetic approach improves the performance of the text to matrix generator.
Parametric comparison based on split criterion on classification algorithmIAEME Publication
This document presents a comparison of different attribute selection criteria for classification algorithms in stream data mining. It analyzes two common criteria - information gain and Gini index - and evaluates their impact on classification accuracy using different datasets. The results show that information gain generally achieves higher accuracy than Gini index, especially for larger data sizes. The document aims to improve the performance of stream data classification algorithms by optimizing the split criterion selection approach.
GRC-MS: A GENETIC RULE-BASED CLASSIFIER MODEL FOR ANALYSIS OF MASS SPECTRA DATAcscpconf
Many studies uses different data mining techniques to analyze mass spectrometry data and
extract useful knowledge about biomarkers. These Biomarkers allow the medical experts to
determine whether an individual has a disease or not. Some of these studies have proposed
models that have obtained high accuracy. However, the black-box nature and complexity of the
proposed models have posed significant issues. Thus, to address this problem and build an
accurate model, we use a genetic algorithm for feature selection along with a rule-based
classifier, namely Genetic Rule-Based Classifier algorithm for Mass Spectra data (GRC-MS).
According to the literature, rule-based classifiers provide understandable rules, but not
accurate. In addition, genetic algorithms have achieved excellent results when used with
different classifiers for feature selection. Experiments are conducted on real dataset and the
proposed classifier GRC-MS achieves 99.7% accuracy. In addition, the generated rules are
more understandable than those of other classifier models
Analysis of crop yield prediction using data mining techniqueseSAT Journals
Abstract
Agrarian sector in India is facing rigorous problem to maximize the crop productivity. More than 60 percent of the crop still depends on monsoon rainfall. Recent developments in Information Technology for agriculture field has become an interesting research area to predict the crop yield. The problem of yield prediction is a major problem that remains to be solved based on available data. Data Mining techniques are the better choices for this purpose. Different Data Mining techniques are used and evaluated in agriculture for estimating the future year's crop production. This paper presents a brief analysis of crop yield prediction using Multiple Linear Regression (MLR) technique and Density based clustering technique for the selected region i.e. East Godavari district of Andhra Pradesh in India.
Keywords: Agrarian Sector, Crop Production, Data Mining, Density based clustering, Information Technology, Multiple Linear Regression, Yield Prediction.
IRJET- Agricultural Data Modeling and Yield Forecasting using Data Mining...IRJET Journal
This document proposes using data mining techniques to develop a predictive model for forecasting crop yields. It involves collecting agricultural data on factors like rainfall, temperature, seed quality, and sowing procedures. Data preprocessing and clustering techniques like K-means are applied. Classification algorithms like Support Vector Machine and Naive Bayes are used to predict crop yield as low, medium, or high. The predictive model aims to help farmers plan cultivation for high crop yields by identifying the best combinations of agricultural factors.
This document discusses using machine learning clustering algorithms to analyze stock market data. It compares the K-means, COBWEB, DBSCAN, EM and OPTICS clustering algorithms in the WEKA tool on a stock market dataset containing 420 instances and 6 attributes. The K-means algorithm had the best performance with the lowest error and fastest runtime. It clustered the data into 4 groups in 0.16 seconds. The COBWEB algorithm clustered the data into 107 groups in 27.88 seconds. The DBSCAN algorithm found 21 clusters in 3.97 seconds. The paper concludes that K-means is best suited for stock market data mining applications due to its simplicity and speed compared to other algorithms.
This document describes a study that used remote sensing to classify land use patterns in a region of India. Supervised and unsupervised classification algorithms were applied to a Sentinel-2 satellite image. Maximum likelihood classification achieved the highest overall accuracy of 72.99% among the methods. The classifications were validated using confusion matrices and kappa coefficients. The study aims to help farmers and policymakers with land management and crop production estimates.
IRJET- Agricultural Crop Yield Prediction using Deep Learning ApproachIRJET Journal
This document discusses using artificial neural networks to predict agricultural crop yields. It begins with an abstract that outlines using ANNs to predict crop yield given various input parameters like soil pH, nitrogen levels, temperature, rainfall, etc. It then provides an introduction on the importance of accurate crop yield prediction. The next sections discuss literature on previous ANN crop yield prediction models, the proposed ANN approach including network architecture and activation functions, the design process, and conclusions. The key points are that ANNs can accurately predict crop yields given various climatic and soil inputs, and providing farmers with these predictions could help maximize profits and minimize losses.
Comparative study on machine learning algorithms for early fire forest detect...IJECEIAES
Forest fires have become a great risk for countries. To minimize their impact and prevent this phenomenon, scientific methods have emerged. Notably machine learning algorithms and decision-making Geographical Information Systems. Therefore, a competitive spatial prediction model for early fire forest detection system using geodata can be proposed. This model can help researchers to predict forest fires and identify risk zonas. System using machine learning algorithm on geodata will be able to notify in real time the interested parts and authorities by providing alerts and presenting on maps based on geographical treatments for more efficacity and analyzing of the situation. This research extends the application of machine learning algorithms for early fire forest prediction to detection and representation in geographical information system (GIS) maps.
This paper proposes an automatic method for early pest detection in crops using image processing and machine learning. The method involves taking images of crop leaves and applying preprocessing techniques like color to grayscale conversion, resizing, filtering and segmentation to isolate the pests from the background. Features are then extracted from the images using gray level co-occurrence matrix and regional properties. A support vector machine classifier is trained on the image features to classify images as containing pests or not containing pests, allowing for early detection and prevention of crop damage from pests.
Neurocomputing is widely implemented in time series area, however the nearness of exceptions that for the most part happen in information time arrangement might be hurtful to the information organize preparing. This is on the grounds that the capacity to consequently discover any examples without earlier suppositions and loss of all-inclusive statement. In principle, the most well-known preparing calculation for Backpropagation calculations inclines toward lessening ordinary least squares estimator (OLS) or all the more particularly, the mean squared error (MSE). In any case, this calculation is not completely hearty when exceptions exist in preparing information, and it will prompt false estimate future esteem. Along these lines, in this paper, we show another calculation that control calculations firefly on slightest middle squares estimator (FFA-LMedS) for BFGS quasi-newton backpropagation neural network nonlinear autoregressive moving (BPNN-NARMA) model to lessen the effect of exceptions in time arrangement information. In the in the mean time, the monthly data of Malaysian Roof Materials cost index from January 1980 to December 2012 (base year 1980=100) with various level of exceptions issue is adjusted in this examination. Toward the finish of this paper, it was found that the upgraded BPNN-NARMA models utilizing FFA-LMedS performed extremely well with RMSE values just about zero errors. It is expected that the finding would help the specialists in Malaysian development activities to handle cost indices data accordingly.
Novel modelling of clustering for enhanced classification performance on gene...IJECEIAES
Gene expression data is popularized for its capability to disclose various disease conditions. However, the conventional procedure to extract gene expression data itself incorporates various artifacts that offer challenges in diagnosis a complex disease indication and classification like cancer. Review of existing research approaches indicates that classification approaches are few to proven to be standard with respect to higher accuracy and applicable to gene expression data apart from unaddresed problems of computational complexity. Therefore, the proposed manuscript introduces a novel and simplified model capable using Graph Fourier Transform, Eigen Value and vector for offering better classification performance considering case study of microarray database, which is one typical example of gene expressiondata. The study outcome shows that proposed system offers comparatively better accuracy and reduced computational complexity with the existing clustering approaches.
A Review on Associative Classification Data Mining Approach in Agricultural S...Editor IJMTER
Data mining in agriculture is a very recent research topic. It consists in the
application of data mining techniques to agriculture. Recent technologies are nowadays able to
provide a lot of information on agricultural-related activities, which can then be analyzed in
order to find important information. A related, but not equivalent term is precision agriculture.
This research aimed to assess the various classification techniques of data mining and apply
them to a soil science database to establish if meaningful relationships can be found. A large
data set of soil database is extracted from the Soil Science & Agricultural department, Bhopal
M.P and National Informatics Centre, The application of data mining techniques has never
been conducted for Bhopal soil data sets. The research compares the different classifiers and
the outcome of this research could improve the management and systems of soil uses
throughout a large number of fields that include agriculture, horticulture, environmental and
land use management.
DATA MINING ATTRIBUTE SELECTION APPROACH FOR DROUGHT MODELLING: A CASE STUDY ...IJDKP
ABSTRACT
The objectives of this paper were to 1) develop an empirical method for selecting relevant attributes for modelling drought and 2) select the most relevant attribute for drought modelling and predictions in the Greater Horn of Africa (GHA). Twenty four attributes from different domain areas were used for this experimental analysis. Two attribute selection algorithms were used for the current study: Principal Component Analysis (PCA) and correlation-based attribute selection (CAS). Using the PCA and CAS algorithms, the 24 attributes were ranked by their merit value. Accordingly, 15 attributes were selected for modelling drought in GHA. The average merit values for the selected attributes ranged from 0.5 to 0.9. The methodology developed here helps to avoid the uncertainty of domain experts’ attribute selection
challenges, which are unsystematic and dominated by somewhat arbitrary trial. Future research may evaluate the developed methodology using relevant classification techniques and quantify the actual information gain from the developed approach.
Classification of Breast Cancer Diseases using Data Mining Techniquesinventionjournals
Medical data mining has great deal for exploring new knowledge from large amount of data. Classification is one of the important data mining techniques for classification of data. In this research work, we have used various data mining based classification techniques for classification of cancer diseases patient or not. We applied the Breast Cancer-Wisconsin (Original) data set into different data mining techniques and compared the accuracy of models with two different data partitions. BayesNet achieved highest accuracy as 97.13% in case of 10-fold data partitions. We have also applied the info gain feature selection technique on BayesNet and Support Vector Machine (SVM) and achieved best accuracy 97.28% accuracy with BayesNet in case of 6 feature subset.
Privacy Preservation and Restoration of Data Using Unrealized Data SetsIJERA Editor
In today’s world, there is an improved advance in hardware technology which increases the capability to store and record personal data about consumers and individuals. Data mining extracts knowledge to support a variety of areas as marketing, medical diagnosis, weather forecasting, national security etc successfully. Still there is a challenge to extract certain kinds of data without violating the data owners’ privacy. As data mining becomes more enveloping, such privacy concerns are increasing. This gives birth to a new category of data mining method called privacy preserving data mining algorithm (PPDM). The aim of this algorithm is to protect the easily affected information in data from the large amount of data set. The privacy preservation of data set can be expressed in the form of decision tree. This paper proposes a privacy preservation based on data set complement algorithms which store the information of the real dataset. So that the private data can be safe from the unauthorized party, if some portion of the data can be lost, then we can recreate the original data set from the unrealized dataset and the perturbed data set.
Efficiency of Prediction Algorithms for Mining Biological DatabasesIOSR Journals
This document analyzes the efficiency of various prediction algorithms for mining biological databases. It discusses prediction through mining biological databases to identify disease risks. It then evaluates several prediction algorithms (ZeroR, OneR, JRip, PART, Decision Table) on a breast cancer dataset using measures like accuracy, sensitivity, specificity, and predictive values. The results show that the JRip and PART algorithms generally had the highest accuracy rates, around 70%, while ZeroR had the lowest accuracy. However, ZeroR had a perfect positive predictive value. The study aims to assess the most efficient algorithms for predictive mining of biological data.
Predictive Data Mining with Normalized Adaptive Training Method for Neural Ne...IJERDJOURNAL
Abstract:- Predictive data mining is an upcoming and fast-growing field and offers a competitive edge for the benefit of organization. In recent decades, researchers have developed new techniques and intelligent algorithms for predictive data mining. In this research paper, we have proposed a novel training algorithm for optimizing neural networks for prediction purpose and to utilize it for the development of prediction models. Models developed in MATLAB Neural Network Toolbox have been tested for insurance datasets taken from a live data warehouse. A comparative study of the proposed algorithm with other popular first and second order algorithms has been presented to judge the predictive accuracy of the suggested technique. Various graphs have been presented to analyse the convergence behaviour of different algorithms towards point of minimum error.
This document discusses key concepts in vector calculus including:
1) The gradient of a scalar, which is a vector representing the directional derivative/rate of change.
2) Divergence of a vector, which measures the outward flux density at a point.
3) Divergence theorem, relating the outward flux through a closed surface to the volume integral of the divergence.
4) Curl of a vector, which measures the maximum circulation and tendency for rotation.
Formulas are provided for calculating these quantities in Cartesian, cylindrical, and spherical coordinate systems. Examples are worked through applying the concepts and formulas.
Application of Gauss,Green and Stokes TheoremSamiul Ehsan
Gauss' law, Stokes' theorem, and Green's theorem are used to relate line integrals, surface integrals, and volume integrals. Gauss' law relates the electric flux through a closed surface to the enclosed charge. Stokes' theorem converts a line integral around a closed curve into a surface integral over the enclosed surface. Green's theorem converts a line integral around a closed curve into a double integral over the enclosed area. These theorems have applications in electrostatics, electrodynamics, calculating mass and momentum, and deriving Kepler's laws of planetary motion.
This document discusses how vectors are used in 2D and 3D gaming. Vectors store positions, directions, and velocities of objects in games. They allow for modeling movement over time by adding velocities to positions and applying accelerations. Vector addition, subtraction, and calculating length are used for tasks like determining the path of projectiles and calculating damage. Vector graphics use control points and strokes defined by vectors to render graphics efficiently at variable sizes. 3D graphics similarly use vectors to define triangle meshes that are rendered to create 3D scenes.
This document provides an overview of key concepts in vector calculus and linear algebra, including:
- The gradient of a scalar field, which describes the direction of steepest ascent/descent.
- Curl, which describes infinitesimal rotation of a 3D vector field.
- Divergence, which measures the magnitude of a vector field's source or sink.
- Solenoidal fields have zero divergence, while irrotational fields have zero curl.
- The directional derivative describes the rate of change of a function at a point in a given direction.
Vectors have both magnitude and direction, represented by arrows. The sum of two vectors is obtained by placing the tail of one vector at the head of the other. If the vectors are at right angles, their dot product is zero, while their cross product is maximum. Scalar multiplication scales the magnitude but not the direction of a vector.
Parametric comparison based on split criterion on classification algorithmIAEME Publication
This document presents a comparison of different attribute selection criteria for classification algorithms in stream data mining. It analyzes two common criteria - information gain and Gini index - and evaluates their impact on classification accuracy using different datasets. The results show that information gain generally achieves higher accuracy than Gini index, especially for larger data sizes. The document aims to improve the performance of stream data classification algorithms by optimizing the split criterion selection approach.
GRC-MS: A GENETIC RULE-BASED CLASSIFIER MODEL FOR ANALYSIS OF MASS SPECTRA DATAcscpconf
Many studies uses different data mining techniques to analyze mass spectrometry data and
extract useful knowledge about biomarkers. These Biomarkers allow the medical experts to
determine whether an individual has a disease or not. Some of these studies have proposed
models that have obtained high accuracy. However, the black-box nature and complexity of the
proposed models have posed significant issues. Thus, to address this problem and build an
accurate model, we use a genetic algorithm for feature selection along with a rule-based
classifier, namely Genetic Rule-Based Classifier algorithm for Mass Spectra data (GRC-MS).
According to the literature, rule-based classifiers provide understandable rules, but not
accurate. In addition, genetic algorithms have achieved excellent results when used with
different classifiers for feature selection. Experiments are conducted on real dataset and the
proposed classifier GRC-MS achieves 99.7% accuracy. In addition, the generated rules are
more understandable than those of other classifier models
Analysis of crop yield prediction using data mining techniqueseSAT Journals
Abstract
Agrarian sector in India is facing rigorous problem to maximize the crop productivity. More than 60 percent of the crop still depends on monsoon rainfall. Recent developments in Information Technology for agriculture field has become an interesting research area to predict the crop yield. The problem of yield prediction is a major problem that remains to be solved based on available data. Data Mining techniques are the better choices for this purpose. Different Data Mining techniques are used and evaluated in agriculture for estimating the future year's crop production. This paper presents a brief analysis of crop yield prediction using Multiple Linear Regression (MLR) technique and Density based clustering technique for the selected region i.e. East Godavari district of Andhra Pradesh in India.
Keywords: Agrarian Sector, Crop Production, Data Mining, Density based clustering, Information Technology, Multiple Linear Regression, Yield Prediction.
IRJET- Agricultural Data Modeling and Yield Forecasting using Data Mining...IRJET Journal
This document proposes using data mining techniques to develop a predictive model for forecasting crop yields. It involves collecting agricultural data on factors like rainfall, temperature, seed quality, and sowing procedures. Data preprocessing and clustering techniques like K-means are applied. Classification algorithms like Support Vector Machine and Naive Bayes are used to predict crop yield as low, medium, or high. The predictive model aims to help farmers plan cultivation for high crop yields by identifying the best combinations of agricultural factors.
This document discusses using machine learning clustering algorithms to analyze stock market data. It compares the K-means, COBWEB, DBSCAN, EM and OPTICS clustering algorithms in the WEKA tool on a stock market dataset containing 420 instances and 6 attributes. The K-means algorithm had the best performance with the lowest error and fastest runtime. It clustered the data into 4 groups in 0.16 seconds. The COBWEB algorithm clustered the data into 107 groups in 27.88 seconds. The DBSCAN algorithm found 21 clusters in 3.97 seconds. The paper concludes that K-means is best suited for stock market data mining applications due to its simplicity and speed compared to other algorithms.
This document describes a study that used remote sensing to classify land use patterns in a region of India. Supervised and unsupervised classification algorithms were applied to a Sentinel-2 satellite image. Maximum likelihood classification achieved the highest overall accuracy of 72.99% among the methods. The classifications were validated using confusion matrices and kappa coefficients. The study aims to help farmers and policymakers with land management and crop production estimates.
IRJET- Agricultural Crop Yield Prediction using Deep Learning ApproachIRJET Journal
This document discusses using artificial neural networks to predict agricultural crop yields. It begins with an abstract that outlines using ANNs to predict crop yield given various input parameters like soil pH, nitrogen levels, temperature, rainfall, etc. It then provides an introduction on the importance of accurate crop yield prediction. The next sections discuss literature on previous ANN crop yield prediction models, the proposed ANN approach including network architecture and activation functions, the design process, and conclusions. The key points are that ANNs can accurately predict crop yields given various climatic and soil inputs, and providing farmers with these predictions could help maximize profits and minimize losses.
Comparative study on machine learning algorithms for early fire forest detect...IJECEIAES
Forest fires have become a great risk for countries. To minimize their impact and prevent this phenomenon, scientific methods have emerged. Notably machine learning algorithms and decision-making Geographical Information Systems. Therefore, a competitive spatial prediction model for early fire forest detection system using geodata can be proposed. This model can help researchers to predict forest fires and identify risk zonas. System using machine learning algorithm on geodata will be able to notify in real time the interested parts and authorities by providing alerts and presenting on maps based on geographical treatments for more efficacity and analyzing of the situation. This research extends the application of machine learning algorithms for early fire forest prediction to detection and representation in geographical information system (GIS) maps.
This paper proposes an automatic method for early pest detection in crops using image processing and machine learning. The method involves taking images of crop leaves and applying preprocessing techniques like color to grayscale conversion, resizing, filtering and segmentation to isolate the pests from the background. Features are then extracted from the images using gray level co-occurrence matrix and regional properties. A support vector machine classifier is trained on the image features to classify images as containing pests or not containing pests, allowing for early detection and prevention of crop damage from pests.
Neurocomputing is widely implemented in time series area, however the nearness of exceptions that for the most part happen in information time arrangement might be hurtful to the information organize preparing. This is on the grounds that the capacity to consequently discover any examples without earlier suppositions and loss of all-inclusive statement. In principle, the most well-known preparing calculation for Backpropagation calculations inclines toward lessening ordinary least squares estimator (OLS) or all the more particularly, the mean squared error (MSE). In any case, this calculation is not completely hearty when exceptions exist in preparing information, and it will prompt false estimate future esteem. Along these lines, in this paper, we show another calculation that control calculations firefly on slightest middle squares estimator (FFA-LMedS) for BFGS quasi-newton backpropagation neural network nonlinear autoregressive moving (BPNN-NARMA) model to lessen the effect of exceptions in time arrangement information. In the in the mean time, the monthly data of Malaysian Roof Materials cost index from January 1980 to December 2012 (base year 1980=100) with various level of exceptions issue is adjusted in this examination. Toward the finish of this paper, it was found that the upgraded BPNN-NARMA models utilizing FFA-LMedS performed extremely well with RMSE values just about zero errors. It is expected that the finding would help the specialists in Malaysian development activities to handle cost indices data accordingly.
Novel modelling of clustering for enhanced classification performance on gene...IJECEIAES
Gene expression data is popularized for its capability to disclose various disease conditions. However, the conventional procedure to extract gene expression data itself incorporates various artifacts that offer challenges in diagnosis a complex disease indication and classification like cancer. Review of existing research approaches indicates that classification approaches are few to proven to be standard with respect to higher accuracy and applicable to gene expression data apart from unaddresed problems of computational complexity. Therefore, the proposed manuscript introduces a novel and simplified model capable using Graph Fourier Transform, Eigen Value and vector for offering better classification performance considering case study of microarray database, which is one typical example of gene expressiondata. The study outcome shows that proposed system offers comparatively better accuracy and reduced computational complexity with the existing clustering approaches.
A Review on Associative Classification Data Mining Approach in Agricultural S...Editor IJMTER
Data mining in agriculture is a very recent research topic. It consists in the
application of data mining techniques to agriculture. Recent technologies are nowadays able to
provide a lot of information on agricultural-related activities, which can then be analyzed in
order to find important information. A related, but not equivalent term is precision agriculture.
This research aimed to assess the various classification techniques of data mining and apply
them to a soil science database to establish if meaningful relationships can be found. A large
data set of soil database is extracted from the Soil Science & Agricultural department, Bhopal
M.P and National Informatics Centre, The application of data mining techniques has never
been conducted for Bhopal soil data sets. The research compares the different classifiers and
the outcome of this research could improve the management and systems of soil uses
throughout a large number of fields that include agriculture, horticulture, environmental and
land use management.
DATA MINING ATTRIBUTE SELECTION APPROACH FOR DROUGHT MODELLING: A CASE STUDY ...IJDKP
ABSTRACT
The objectives of this paper were to 1) develop an empirical method for selecting relevant attributes for modelling drought and 2) select the most relevant attribute for drought modelling and predictions in the Greater Horn of Africa (GHA). Twenty four attributes from different domain areas were used for this experimental analysis. Two attribute selection algorithms were used for the current study: Principal Component Analysis (PCA) and correlation-based attribute selection (CAS). Using the PCA and CAS algorithms, the 24 attributes were ranked by their merit value. Accordingly, 15 attributes were selected for modelling drought in GHA. The average merit values for the selected attributes ranged from 0.5 to 0.9. The methodology developed here helps to avoid the uncertainty of domain experts’ attribute selection
challenges, which are unsystematic and dominated by somewhat arbitrary trial. Future research may evaluate the developed methodology using relevant classification techniques and quantify the actual information gain from the developed approach.
Classification of Breast Cancer Diseases using Data Mining Techniquesinventionjournals
Medical data mining has great deal for exploring new knowledge from large amount of data. Classification is one of the important data mining techniques for classification of data. In this research work, we have used various data mining based classification techniques for classification of cancer diseases patient or not. We applied the Breast Cancer-Wisconsin (Original) data set into different data mining techniques and compared the accuracy of models with two different data partitions. BayesNet achieved highest accuracy as 97.13% in case of 10-fold data partitions. We have also applied the info gain feature selection technique on BayesNet and Support Vector Machine (SVM) and achieved best accuracy 97.28% accuracy with BayesNet in case of 6 feature subset.
Privacy Preservation and Restoration of Data Using Unrealized Data SetsIJERA Editor
In today’s world, there is an improved advance in hardware technology which increases the capability to store and record personal data about consumers and individuals. Data mining extracts knowledge to support a variety of areas as marketing, medical diagnosis, weather forecasting, national security etc successfully. Still there is a challenge to extract certain kinds of data without violating the data owners’ privacy. As data mining becomes more enveloping, such privacy concerns are increasing. This gives birth to a new category of data mining method called privacy preserving data mining algorithm (PPDM). The aim of this algorithm is to protect the easily affected information in data from the large amount of data set. The privacy preservation of data set can be expressed in the form of decision tree. This paper proposes a privacy preservation based on data set complement algorithms which store the information of the real dataset. So that the private data can be safe from the unauthorized party, if some portion of the data can be lost, then we can recreate the original data set from the unrealized dataset and the perturbed data set.
Efficiency of Prediction Algorithms for Mining Biological DatabasesIOSR Journals
This document analyzes the efficiency of various prediction algorithms for mining biological databases. It discusses prediction through mining biological databases to identify disease risks. It then evaluates several prediction algorithms (ZeroR, OneR, JRip, PART, Decision Table) on a breast cancer dataset using measures like accuracy, sensitivity, specificity, and predictive values. The results show that the JRip and PART algorithms generally had the highest accuracy rates, around 70%, while ZeroR had the lowest accuracy. However, ZeroR had a perfect positive predictive value. The study aims to assess the most efficient algorithms for predictive mining of biological data.
Predictive Data Mining with Normalized Adaptive Training Method for Neural Ne...IJERDJOURNAL
Abstract:- Predictive data mining is an upcoming and fast-growing field and offers a competitive edge for the benefit of organization. In recent decades, researchers have developed new techniques and intelligent algorithms for predictive data mining. In this research paper, we have proposed a novel training algorithm for optimizing neural networks for prediction purpose and to utilize it for the development of prediction models. Models developed in MATLAB Neural Network Toolbox have been tested for insurance datasets taken from a live data warehouse. A comparative study of the proposed algorithm with other popular first and second order algorithms has been presented to judge the predictive accuracy of the suggested technique. Various graphs have been presented to analyse the convergence behaviour of different algorithms towards point of minimum error.
This document discusses key concepts in vector calculus including:
1) The gradient of a scalar, which is a vector representing the directional derivative/rate of change.
2) Divergence of a vector, which measures the outward flux density at a point.
3) Divergence theorem, relating the outward flux through a closed surface to the volume integral of the divergence.
4) Curl of a vector, which measures the maximum circulation and tendency for rotation.
Formulas are provided for calculating these quantities in Cartesian, cylindrical, and spherical coordinate systems. Examples are worked through applying the concepts and formulas.
Application of Gauss,Green and Stokes TheoremSamiul Ehsan
Gauss' law, Stokes' theorem, and Green's theorem are used to relate line integrals, surface integrals, and volume integrals. Gauss' law relates the electric flux through a closed surface to the enclosed charge. Stokes' theorem converts a line integral around a closed curve into a surface integral over the enclosed surface. Green's theorem converts a line integral around a closed curve into a double integral over the enclosed area. These theorems have applications in electrostatics, electrodynamics, calculating mass and momentum, and deriving Kepler's laws of planetary motion.
This document discusses how vectors are used in 2D and 3D gaming. Vectors store positions, directions, and velocities of objects in games. They allow for modeling movement over time by adding velocities to positions and applying accelerations. Vector addition, subtraction, and calculating length are used for tasks like determining the path of projectiles and calculating damage. Vector graphics use control points and strokes defined by vectors to render graphics efficiently at variable sizes. 3D graphics similarly use vectors to define triangle meshes that are rendered to create 3D scenes.
This document provides an overview of key concepts in vector calculus and linear algebra, including:
- The gradient of a scalar field, which describes the direction of steepest ascent/descent.
- Curl, which describes infinitesimal rotation of a 3D vector field.
- Divergence, which measures the magnitude of a vector field's source or sink.
- Solenoidal fields have zero divergence, while irrotational fields have zero curl.
- The directional derivative describes the rate of change of a function at a point in a given direction.
Vectors have both magnitude and direction, represented by arrows. The sum of two vectors is obtained by placing the tail of one vector at the head of the other. If the vectors are at right angles, their dot product is zero, while their cross product is maximum. Scalar multiplication scales the magnitude but not the direction of a vector.
Application of coordinate system and vectors in the real lifeАлиакбар Рахимов
1. The document discusses the application of coordinate systems and vectors in real life contexts. It provides several examples, such as using coordinates to describe the location of objects or aircraft, map projections, latitude and longitude coordinates, their use in economics, and by the military.
2. Vectors are also discussed as applied to cannons, sports like baseball and football, wind vectors, and sailing. Calculating forces, torques, acceleration, and velocity also require the use of vectors.
3. Coordinate systems and vectors have many practical applications, from describing positions to applications in transportation, navigation, military targeting, sports, and physics concepts like motion. They underlie many areas of science, engineering, and daily life.
Application of coordinate system and vectors in the real life
Similar to Forest Area Estimation in Kutai Nasional Park of East Kalimantan Using Computer System Aplication based Genetic Algorithm-Support Vector Machine
Feature selection for multiple water quality status: integrated bootstrapping...IJECEIAES
STORET is one method to determine the river water quality, and to classify them into four classes (very good, good, medium and bad) based on the data of water for each attribute or feature. The success of the formation of pattern recognition model much depends on the quality of data. There are two issues as the concern of this research as follows, the data having disproportionate amount among the classes (imbalance class) and the finding of noise on its attribute. Therefore, this research integrates the SMOTE Technique and bootstrapping to handle the problem of imbalance class. While an experiment is conducted to eliminate the noise on the attribute by using some feature selection algorithms with filter approach (information gain, rule, derivation, correlation and chi square). This research has some stages as follows: data understanding, pre-processing, imbalance class, feature selection, classification and performance evaluation. Based on the result of testing using 10-fold cross validation, it shows that the use of the SMOTE-bootstrapping technique is able to increase the accuracy from 83.3% to be 98.8%. While the process of noise elimination onthe data attribute is also able to increase the accuracy to be 99.5% (the use of feature subset produced by the information gain algorithm and the decision tree classification algorithm).
Agricultural research has strengthened the optimized
economical profit, internationally and is very vast and
important field to gain more benefits.
In future agriculture is the only scope for all the people. But
today number of people having land, but they don’t know how
to yield the crops.
So many of people are doing useless agriculture by
cultivating the crop on improper soil. To implement the
application to identify the types of soil,water source of that
land whether that land is based on rain or bore water. And
suggest what of crop is suitable for that soil. So through this
application provide application for the people to know about
the agriculture. There is no any application to know about the
cultivation. However, it can be enhanced by the use of different
technological resources, tool, and procedures. Predict the type
of crop which one is suitable for that particular soil, weather
condition, temperature and so on. So for, using machine
learning with the set of data set are identified the crop for the
corresponding soil.
Agriculture crop yield prediction using inertia based cat swarm optimizationIJECEIAES
Crop yield prediction is among the most important and main sources of income in the Indian economy. In this paper, the improved cat swarm optimization (ICSO) based recurrent neural network (RNN) model is proposed for crop yield prediction using time series data. The inertia weight parameter is added to position equation that is selected randomly, and a new velocity equation is produced which enhances the searching ability in the best cat area. By using inertia weight, the ICSO enhances performance of feature selection and obtains better convergence in minimum iteration. The RNN is applied to produce direct graph using sequence of data and decides current layer output by involving all other existing calculations. The performance of the model is estimated using coefficient of determination (R2), root mean square error (RMSE), mean squared error (MSE), and mean absolute error (MAE) on the yield from the years 2011 to 2021 with an annual prediction for 120 records of approximately 8 million nuts. The evaluated result shows that the proposed ICSO-RNN model delivers metrics such as R2, MAE, MSE, and RMSE values of 0.99, 0.77, 0.68, and 0.82 correspondingly, which ensures accurate yield prediction when compared with the existing methods which are hybrid reinforcement learning-random forest (RL-RF) and machine learning (ML) methods.
A new model for iris data set classification based on linear support vector m...IJECEIAES
1. The authors propose a new model for classifying the iris data set using a linear support vector machine (SVM) classifier with genetic algorithm optimization of the SVM's C and gamma parameters.
2. Principal component analysis was used to reduce the iris data set features from four to three before classification.
3. The genetic algorithm was shown to optimize the SVM parameters, achieving 98.7% accuracy on the iris data set classification compared to 95.3% accuracy without parameter optimization.
Efficient Disease Classifier Using Data Mining Techniques: Refinement of Rand...IOSR Journals
The document describes research on improving the termination criteria for the random forest algorithm when classifying disease data. The random forest algorithm constructs a forest of decision trees to predict disease classification. Typically, random forest terminates tree construction based on accuracy and correlation metrics. The proposed method uses binomial distribution, multinomial distribution, and sequential probability ratio testing to determine when to terminate tree construction, with the goal of stopping earlier than typical random forest approaches. Experimental results on five disease datasets show the proposed method with probability distributions achieves better accuracy than classical random forest while constructing fewer trees, improving the efficiency of the algorithm.
IRJET- Gene Mutation Data using Multiplicative Adaptive Algorithm and Gene On...IRJET Journal
This document presents a methodology for analyzing gene mutation data using ontologies and association rule mining. It aims to develop a common knowledge base for genomic and proteomic analysis by integrating multiple data sources. The methodology involves using k-nearest neighbors algorithm to find similar genes, an iterative multiplicative updating algorithm to solve optimization problems, and SNCoNMF to identify co-regulatory modules between genes, microRNAs and transcription factors. The results are represented using a Bayesian rose tree for efficient visualization of associations between genetic components and diseases.
IRJET - An User Friendly Interface for Data Preprocessing and Visualizati...IRJET Journal
This document presents a tool for preprocessing and visualizing data using machine learning models. It aims to simplify the preprocessing steps for users by performing tasks like data cleaning, transformation, and reduction. The tool takes in a raw dataset, cleans it by removing missing values, outliers, etc. It then allows users to apply machine learning algorithms like linear regression, KNN, random forest for analysis. The processed and predicted data can be visualized. The tool is intended to save time by automating preprocessing and providing visual outputs for analysis using machine learning models on large datasets.
IRJET- Expert Independent Bayesian Data Fusion and Decision Making Model for ...IRJET Journal
This document proposes an expert independent Bayesian data fusion and decision making model for multi-sensor systems smart control. The model uses a Naive Bayes classifier to predict the system state based only on prior and current sensor data. Simulations of a three sensor system (soil temperature, air temperature, and moisture) achieved an overall prediction accuracy of more than 96%. However, real-world implementation of the proposed algorithm is still needed.
This document summarizes a study that used satellite images and supervised classification to monitor forest land cover in Gisoom forest park in Iran. Land samples were taken using GPS and classified using ENVI software. Maximum likelihood classification of satellite images from 2007 achieved a total accuracy of 75.98% and kappa coefficient of 74.73%, indicating good classification. The study found that residential development, construction of recreational structures, roads, and tourism led to decreases in forest areas over time.
IRJET - Enlightening Farmers on Crop YieldIRJET Journal
This document discusses using data mining techniques to predict crop yields to help farmers. It proposes using a random forest regression algorithm on past agricultural data from 2000-2014 to build a prediction model. The model would help farmers select optimal crops, understand weather patterns, and maximize yields. The system is described as gathering data, preprocessing it, training a random forest model on 60% of the data and testing it on 20%. It would then provide yield predictions and recommendations to farmers through a visualization tool. The goal is to help guide farmers' decisions around fertilizer use, soil management, and crop selection to improve production levels.
This document summarizes different analytical models that can be used for crop prediction, including classification models. It discusses feature selection methods like principal component analysis and information gain that are important for crop prediction models. The document reviews different machine learning techniques used in previous studies for crop yield prediction, such as linear regression, k-nearest neighbors, neural networks, support vector machines, and decision trees. It aims to compare the performance of these classification techniques for predicting crop yields based on parameters like temperature, rainfall, soil characteristics, and more.
This document summarizes a study that uses a genetic algorithm to optimize imputing missing cost data for fans used in road tunnels in Sweden. The genetic algorithm is used to impute the missing cost data by optimizing the valid data period used. The results show highly correlated data (R-squared of 0.99) after imputing the missing values, indicating the genetic algorithm provides an effective way to optimize imputing and create complete data that can then be used for forecasting and life cycle cost analysis. The document also reviews other methods for data imputation and discusses experimental results comparing the proposed two-stage approach using K-means clustering and multilayer perceptron on several datasets.
Random forest model for forecasting vegetable prices: a case study in Nakhon ...IJECEIAES
The objectives of this research were developing a model for forecasting vegetable prices in Nakhon Si Thammarat Province using random forest and comparing the forecast results of different crops. The information used in this paper were monthly climate data and average monthly vegetable prices collected between 2011 – 2020 from Nakhon Si Thammarat meteorological station and Nakhon Si Thammarat Provincial Commercial Office, respectively. We evaluated model performance based on mean absolute percentage error (MAPE), root mean squared error (RMSE), and mean absolute error (MAE). The experimental results showed that the random forest model was able to predict the prices of vegetables, including pumpkin, eggplant, and lentils with high accuracy with MAPE values of 0.09, 0.07, and 0.15, with RMSE values of 1.82, 1.46, and 2.33, and with MAE values of 3.32, 2.15, and 5.42, respectively. The forecast model derived from this research can be beneficial for vegetable planting planning in the Pak Phanang River Basin of Nakhon Si Thammarat Province, Thailand.
Handling Imbalanced Data: SMOTE vs. Random UndersamplingIRJET Journal
This document compares different techniques for handling imbalanced data, including random undersampling and SMOTE (Synthetic Minority Over-sampling Technique), using two machine learning classifiers - Random Forest and XGBoost. It finds that XGBoost generally performs better than Random Forest across different sampling techniques in terms of metrics like AUC, sensitivity and specificity. Random undersampling with XGBoost provided the most balanced results on the randomly generated imbalanced dataset used in this study. SMOTE without proper validation gave the best numerical results but likely overfit the data.
An automated speech recognition and feature selection approach based on impro...IAESIJAI
Automatic speech recognition (ASR) approach is dependent on optimal
speech feature extraction, which attempts to get a parametric depiction of an
input speech signal. Feature extraction (FE) strategy combined with a feature
selection (FS) approach should capture the most important features of the
signal while discarding the rest. FS is a crucial process that can affect the
pattern classification and recognition system's performance. In this research,
we introduce a hybrid supervised learning using metaheuristic technique for
optimum FE and FS termed Northern Goshawk optimization (NGO) and
opposition-based learning (OBL). Pre-processing, feature extraction and
selection, and recognition are the three steps of the proposed technique. The
pre-processing is done first to lessen the amount of noise. In the FE stage, we
extract features. The OBL-NGO method is used to pick the best collection of
extracted characteristics. Finally, these optimised features are utilised to train
the k-nearest neighbour (KNN) classifier, and the matching text is shown as
the output based on these optimised characteristics of the provided input audio
signal. The system's performance is outstanding, and the suggested OBLNGO is best suited for ASR, according to the testing data.
IRJET- Agricultural Crop Classification Models in Data Mining TechniquesIRJET Journal
This document discusses using Naive Bayes classification to classify agricultural crop types using Landsat satellite imagery data. The researchers analyzed Landsat data using Naive Bayes and evaluated the accuracy of crop classification using various performance metrics. On the training dataset, Naive Bayes correctly classified 211 instances and incorrectly classified 238 instances. On the testing dataset, it correctly classified 49 instances and incorrectly classified 45 instances. The researchers concluded Naive Bayes can achieve reasonably high accuracy for crop classification but plan to compare its performance to other algorithms in future work.
This document discusses a study that aims to identify predictive biomarkers of metabolic syndrome from metabolomic data using knowledge discovery and machine learning methods. The study will use support vector machines and random forests to analyze metabolomic datasets and identify predictive biomarkers. These machine learning algorithms will be compared and the relevant features obtained from supervised approaches will be visualized. Formal concept analysis will be used to construct a concept lattice and discover association rules between emerging features. The overall goal is to develop a workflow to extract knowledge from complex metabolomic data and help experts identify relevant biomarkers for metabolic syndrome.
TUPLE VALUE BASED MULTIPLICATIVE DATA PERTURBATION APPROACH TO PRESERVE PRIVA...IJDKP
Huge volume of data from domain specific applications such as medical, financial, library, telephone,
shopping records and individual are regularly generated. Sharing of these data is proved to be beneficial
for data mining application. On one hand such data is an important asset to business decision making by
analyzing it. On the other hand data privacy concerns may prevent data owners from sharing information
for data analysis. In order to share data while preserving privacy, data owner must come up with a solution
which achieves the dual goal of privacy preservation as well as an accuracy of data mining task –
clustering and classification. An efficient and effective approach has been proposed that aims to protect
privacy of sensitive information and obtaining data clustering with minimum information loss
Comparative study of various supervisedclassification methodsforanalysing def...eSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
Performance evaluation of random forest with feature selection methods in pre...IJECEIAES
Data mining is nothing but the process of viewing data in different angle and compiling it into appropriate information. Recent improvements in the area of data mining and machine learning have empowered the research in biomedical field to improve the condition of general health care. Since the wrong classification may lead to poor prediction, there is a need to perform the better classification which further improves the prediction rate of the medical datasets. When medical data mining is applied on the medical datasets the important and difficult challenges are the classification and prediction. In this proposed work we evaluate the PIMA Indian Diabtes data set of UCI repository using machine learning algorithm like Random Forest along with feature selection methods such as forward selection and backward elimination based on entropy evaluation method using percentage split as test option. The experiment was conducted using R studio platform and we achieved classification accuracy of 84.1%. From results we can say that Random Forest predicts diabetes better than other techniques with less number of attributes so that one can avoid least important test for identifying diabetes.
Similar to Forest Area Estimation in Kutai Nasional Park of East Kalimantan Using Computer System Aplication based Genetic Algorithm-Support Vector Machine (20)
Text Mining in Digital Libraries using OKAPI BM25 ModelEditor IJCATR
The emergence of the internet has made vast amounts of information available and easily accessible online. As a result, most libraries have digitized their content in order to remain relevant to their users and to keep pace with the advancement of the internet. However, these digital libraries have been criticized for using inefficient information retrieval models that do not perform relevance ranking to the retrieved results. This paper proposed the use of OKAPI BM25 model in text mining so as means of improving relevance ranking of digital libraries. Okapi BM25 model was selected because it is a probability-based relevance ranking algorithm. A case study research was conducted and the model design was based on information retrieval processes. The performance of Boolean, vector space, and Okapi BM25 models was compared for data retrieval. Relevant ranked documents were retrieved and displayed at the OPAC framework search page. The results revealed that Okapi BM 25 outperformed Boolean model and Vector Space model. Therefore, this paper proposes the use of Okapi BM25 model to reward terms according to their relative frequencies in a document so as to improve the performance of text mining in digital libraries.
Green Computing, eco trends, climate change, e-waste and eco-friendlyEditor IJCATR
This document discusses green computing practices and sustainable IT services. It provides an overview of factors driving adoption of green computing to reduce costs and environmental impact of data centers, such as rising energy costs and density. Green strategies discussed include improving infrastructure efficiency, power management, thermal management, efficient product design, and virtualization to optimize resource utilization. The document examines how green computing aims to lower costs and environmental footprint, and how sustainable IT services take a broader approach considering economic, environmental and social impacts.
Policies for Green Computing and E-Waste in NigeriaEditor IJCATR
Computers today are an integral part of individuals’ lives all around the world, but unfortunately these devices are toxic to the environment given the materials used, their limited battery life and technological obsolescence. Individuals are concerned about the hazardous materials ever present in computers, even if the importance of various attributes differs, and that a more environment -friendly attitude can be obtained through exposure to educational materials. In this paper, we aim to delineate the problem of e-waste in Nigeria and highlight a series of measures and the advantage they herald for our country and propose a series of action steps to develop in these areas further. It is possible for Nigeria to have an immediate economic stimulus and job creation while moving quickly to abide by the requirements of climate change legislation and energy efficiency directives. The costs of implementing energy efficiency and renewable energy measures are minimal as they are not cash expenditures but rather investments paid back by future, continuous energy savings.
Performance Evaluation of VANETs for Evaluating Node Stability in Dynamic Sce...Editor IJCATR
Vehicular ad hoc networks (VANETs) are a favorable area of exploration which empowers the interconnection amid the movable vehicles and between transportable units (vehicles) and road side units (RSU). In Vehicular Ad Hoc Networks (VANETs), mobile vehicles can be organized into assemblage to promote interconnection links. The assemblage arrangement according to dimensions and geographical extend has serious influence on attribute of interaction .Vehicular ad hoc networks (VANETs) are subclass of mobile Ad-hoc network involving more complex mobility patterns. Because of mobility the topology changes very frequently. This raises a number of technical challenges including the stability of the network .There is a need for assemblage configuration leading to more stable realistic network. The paper provides investigation of various simulation scenarios in which cluster using k-means algorithm are generated and their numbers are varied to find the more stable configuration in real scenario of road.
Optimum Location of DG Units Considering Operation ConditionsEditor IJCATR
The optimal sizing and placement of Distributed Generation units (DG) are becoming very attractive to researchers these days. In this paper a two stage approach has been used for allocation and sizing of DGs in distribution system with time varying load model. The strategic placement of DGs can help in reducing energy losses and improving voltage profile. The proposed work discusses time varying loads that can be useful for selecting the location and optimizing DG operation. The method has the potential to be used for integrating the available DGs by identifying the best locations in a power system. The proposed method has been demonstrated on 9-bus test system.
Analysis of Comparison of Fuzzy Knn, C4.5 Algorithm, and Naïve Bayes Classifi...Editor IJCATR
Early detection of diabetes mellitus (DM) can prevent or inhibit complication. There are several laboratory test that must be done to detect DM. The result of this laboratory test then converted into data training. Data training used in this study generated from UCI Pima Database with 6 attributes that were used to classify positive or negative diabetes. There are various classification methods that are commonly used, and in this study three of them were compared, which were fuzzy KNN, C4.5 algorithm and Naïve Bayes Classifier (NBC) with one identical case. The objective of this study was to create software to classify DM using tested methods and compared the three methods based on accuracy, precision, and recall. The results showed that the best method was Fuzzy KNN with average and maximum accuracy reached 96% and 98%, respectively. In second place, NBC method had respective average and maximum accuracy of 87.5% and 90%. Lastly, C4.5 algorithm had average and maximum accuracy of 79.5% and 86%, respectively.
Web Scraping for Estimating new Record from Source SiteEditor IJCATR
Study in the Competitive field of Intelligent, and studies in the field of Web Scraping, have a symbiotic relationship mutualism. In the information age today, the website serves as a main source. The research focus is on how to get data from websites and how to slow down the intensity of the download. The problem that arises is the website sources are autonomous so that vulnerable changes the structure of the content at any time. The next problem is the system intrusion detection snort installed on the server to detect bot crawler. So the researchers propose the use of the methods of Mining Data Records and the method of Exponential Smoothing so that adaptive to changes in the structure of the content and do a browse or fetch automatically follow the pattern of the occurrences of the news. The results of the tests, with the threshold 0.3 for MDR and similarity threshold score 0.65 for STM, using recall and precision values produce f-measure average 92.6%. While the results of the tests of the exponential estimation smoothing using ? = 0.5 produces MAE 18.2 datarecord duplicate. It slowed down to 3.6 datarecord from 21.8 datarecord results schedule download/fetch fix in an average time of occurrence news.
Evaluating Semantic Similarity between Biomedical Concepts/Classes through S...Editor IJCATR
Most of the existing semantic similarity measures that use ontology structure as their primary source can measure semantic similarity between concepts/classes using single ontology. The ontology-based semantic similarity techniques such as structure-based semantic similarity techniques (Path Length Measure, Wu and Palmer’s Measure, and Leacock and Chodorow’s measure), information content-based similarity techniques (Resnik’s measure, Lin’s measure), and biomedical domain ontology techniques (Al-Mubaid and Nguyen’s measure (SimDist)) were evaluated relative to human experts’ ratings, and compared on sets of concepts using the ICD-10 “V1.0” terminology within the UMLS. The experimental results validate the efficiency of the SemDist technique in single ontology, and demonstrate that SemDist semantic similarity techniques, compared with the existing techniques, gives the best overall results of correlation with experts’ ratings.
Semantic Similarity Measures between Terms in the Biomedical Domain within f...Editor IJCATR
The techniques and tests are tools used to define how measure the goodness of ontology or its resources. The similarity between biomedical classes/concepts is an important task for the biomedical information extraction and knowledge discovery. However, most of the semantic similarity techniques can be adopted to be used in the biomedical domain (UMLS). Many experiments have been conducted to check the applicability of these measures. In this paper, we investigate to measure semantic similarity between two terms within single ontology or multiple ontologies in ICD-10 “V1.0” as primary source, and compare my results to human experts score by correlation coefficient.
A Strategy for Improving the Performance of Small Files in Openstack Swift Editor IJCATR
This is an effective way to improve the storage access performance of small files in Openstack Swift by adding an aggregate storage module. Because Swift will lead to too much disk operation when querying metadata, the transfer performance of plenty of small files is low. In this paper, we propose an aggregated storage strategy (ASS), and implement it in Swift. ASS comprises two parts which include merge storage and index storage. At the first stage, ASS arranges the write request queue in chronological order, and then stores objects in volumes. These volumes are large files that are stored in Swift actually. During the short encounter time, the object-to-volume mapping information is stored in Key-Value store at the second stage. The experimental results show that the ASS can effectively improve Swift's small file transfer performance.
Integrated System for Vehicle Clearance and RegistrationEditor IJCATR
Efficient management and control of government's cash resources rely on government banking arrangements. Nigeria, like many low income countries, employed fragmented systems in handling government receipts and payments. Later in 2016, Nigeria implemented a unified structure as recommended by the IMF, where all government funds are collected in one account would reduce borrowing costs, extend credit and improve government's fiscal policy among other benefits to government. This situation motivated us to embark on this research to design and implement an integrated system for vehicle clearance and registration. This system complies with the new Treasury Single Account policy to enable proper interaction and collaboration among five different level agencies (NCS, FRSC, SBIR, VIO and NPF) saddled with vehicular administration and activities in Nigeria. Since the system is web based, Object Oriented Hypermedia Design Methodology (OOHDM) is used. Tools such as Php, JavaScript, css, html, AJAX and other web development technologies were used. The result is a web based system that gives proper information about a vehicle starting from the exact date of importation to registration and renewal of licensing. Vehicle owner information, custom duty information, plate number registration details, etc. will also be efficiently retrieved from the system by any of the agencies without contacting the other agency at any point in time. Also number plate will no longer be the only means of vehicle identification as it is presently the case in Nigeria, because the unified system will automatically generate and assigned a Unique Vehicle Identification Pin Number (UVIPN) on payment of duty in the system to the vehicle and the UVIPN will be linked to the various agencies in the management information system.
Assessment of the Efficiency of Customer Order Management System: A Case Stu...Editor IJCATR
The Supermarket Management System deals with the automation of buying and selling of good and services. It includes both sales and purchase of items. The project Supermarket Management System is to be developed with the objective of making the system reliable, easier, fast, and more informative.
Energy-Aware Routing in Wireless Sensor Network Using Modified Bi-Directional A*Editor IJCATR
Energy is a key component in the Wireless Sensor Network (WSN)[1]. The system will not be able to run according to its function without the availability of adequate power units. One of the characteristics of wireless sensor network is Limitation energy[2]. A lot of research has been done to develop strategies to overcome this problem. One of them is clustering technique. The popular clustering technique is Low Energy Adaptive Clustering Hierarchy (LEACH)[3]. In LEACH, clustering techniques are used to determine Cluster Head (CH), which will then be assigned to forward packets to Base Station (BS). In this research, we propose other clustering techniques, which utilize the Social Network Analysis approach theory of Betweeness Centrality (BC) which will then be implemented in the Setup phase. While in the Steady-State phase, one of the heuristic searching algorithms, Modified Bi-Directional A* (MBDA *) is implemented. The experiment was performed deploy 100 nodes statically in the 100x100 area, with one Base Station at coordinates (50,50). To find out the reliability of the system, the experiment to do in 5000 rounds. The performance of the designed routing protocol strategy will be tested based on network lifetime, throughput, and residual energy. The results show that BC-MBDA * is better than LEACH. This is influenced by the ways of working LEACH in determining the CH that is dynamic, which is always changing in every data transmission process. This will result in the use of energy, because they always doing any computation to determine CH in every transmission process. In contrast to BC-MBDA *, CH is statically determined, so it can decrease energy usage.
Security in Software Defined Networks (SDN): Challenges and Research Opportun...Editor IJCATR
In networks, the rapidly changing traffic patterns of search engines, Internet of Things (IoT) devices, Big Data and data centers has thrown up new challenges for legacy; existing networks; and prompted the need for a more intelligent and innovative way to dynamically manage traffic and allocate limited network resources. Software Defined Network (SDN) which decouples the control plane from the data plane through network vitalizations aims to address these challenges. This paper has explored the SDN architecture and its implementation with the OpenFlow protocol. It has also assessed some of its benefits over traditional network architectures, security concerns and how it can be addressed in future research and related works in emerging economies such as Nigeria.
Measure the Similarity of Complaint Document Using Cosine Similarity Based on...Editor IJCATR
Report handling on "LAPOR!" (Laporan, Aspirasi dan Pengaduan Online Rakyat) system depending on the system administrator who manually reads every incoming report [3]. Read manually can lead to errors in handling complaints [4] if the data flow is huge and grows rapidly, it needs at least three days to prepare a confirmation and it sensitive to inconsistencies [3]. In this study, the authors propose a model that can measure the identities of the Query (Incoming) with Document (Archive). The authors employed Class-Based Indexing term weighting scheme, and Cosine Similarities to analyse document similarities. CoSimTFIDF, CoSimTFICF and CoSimTFIDFICF values used in classification as feature for K-Nearest Neighbour (K-NN) classifier. The optimum result evaluation is pre-processing employ 75% of training data ratio and 25% of test data with CoSimTFIDF feature. It deliver a high accuracy 84%. The k = 5 value obtain high accuracy 84.12%
Hangul Recognition Using Support Vector MachineEditor IJCATR
The recognition of Hangul Image is more difficult compared with that of Latin. It could be recognized from the structural arrangement. Hangul is arranged from two dimensions while Latin is only from the left to the right. The current research creates a system to convert Hangul image into Latin text in order to use it as a learning material on reading Hangul. In general, image recognition system is divided into three steps. The first step is preprocessing, which includes binarization, segmentation through connected component-labeling method, and thinning with Zhang Suen to decrease some pattern information. The second is receiving the feature from every single image, whose identification process is done through chain code method. The third is recognizing the process using Support Vector Machine (SVM) with some kernels. It works through letter image and Hangul word recognition. It consists of 34 letters, each of which has 15 different patterns. The whole patterns are 510, divided into 3 data scenarios. The highest result achieved is 94,7% using SVM kernel polynomial and radial basis function. The level of recognition result is influenced by many trained data. Whilst the recognition process of Hangul word applies to the type 2 Hangul word with 6 different patterns. The difference of these patterns appears from the change of the font type. The chosen fonts for data training are such as Batang, Dotum, Gaeul, Gulim, Malgun Gothic. Arial Unicode MS is used to test the data. The lowest accuracy is achieved through the use of SVM kernel radial basis function, which is 69%. The same result, 72 %, is given by the SVM kernel linear and polynomial.
Application of 3D Printing in EducationEditor IJCATR
This paper provides a review of literature concerning the application of 3D printing in the education system. The review identifies that 3D Printing is being applied across the Educational levels [1] as well as in Libraries, Laboratories, and Distance education systems. The review also finds that 3D Printing is being used to teach both students and trainers about 3D Printing and to develop 3D Printing skills.
Survey on Energy-Efficient Routing Algorithms for Underwater Wireless Sensor ...Editor IJCATR
In underwater environment, for retrieval of information the routing mechanism is used. In routing mechanism there are three to four types of nodes are used, one is sink node which is deployed on the water surface and can collect the information, courier/super/AUV or dolphin powerful nodes are deployed in the middle of the water for forwarding the packets, ordinary nodes are also forwarder nodes which can be deployed from bottom to surface of the water and source nodes are deployed at the seabed which can extract the valuable information from the bottom of the sea. In underwater environment the battery power of the nodes is limited and that power can be enhanced through better selection of the routing algorithm. This paper focuses the energy-efficient routing algorithms for their routing mechanisms to prolong the battery power of the nodes. This paper also focuses the performance analysis of the energy-efficient algorithms under which we can examine the better performance of the route selection mechanism which can prolong the battery power of the node
Comparative analysis on Void Node Removal Routing algorithms for Underwater W...Editor IJCATR
The designing of routing algorithms faces many challenges in underwater environment like: propagation delay, acoustic channel behaviour, limited bandwidth, high bit error rate, limited battery power, underwater pressure, node mobility, localization 3D deployment, and underwater obstacles (voids). This paper focuses the underwater voids which affects the overall performance of the entire network. The majority of the researchers have used the better approaches for removal of voids through alternate path selection mechanism but still research needs improvement. This paper also focuses the architecture and its operation through merits and demerits of the existing algorithms. This research article further focuses the analytical method of the performance analysis of existing algorithms through which we found the better approach for removal of voids
Decay Property for Solutions to Plate Type Equations with Variable CoefficientsEditor IJCATR
In this paper we consider the initial value problem for a plate type equation with variable coefficients and memory in
1 n R n ), which is of regularity-loss property. By using spectrally resolution, we study the pointwise estimates in the spectral
space of the fundamental solution to the corresponding linear problem. Appealing to this pointwise estimates, we obtain the global
existence and the decay estimates of solutions to the semilinear problem by employing the fixed point theorem
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
Building Production Ready Search Pipelines with Spark and MilvusZilliz
Spark is the widely used ETL tool for processing, indexing and ingesting data to serving stack for search. Milvus is the production-ready open-source vector database. In this talk we will show how to use Spark to process unstructured data to extract vector representations, and push the vectors to Milvus vector database for search serving.
How to Interpret Trends in the Kalyan Rajdhani Mix Chart.pdfChart Kalyan
A Mix Chart displays historical data of numbers in a graphical or tabular form. The Kalyan Rajdhani Mix Chart specifically shows the results of a sequence of numbers over different periods.
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
Generating privacy-protected synthetic data using Secludy and MilvusZilliz
During this demo, the founders of Secludy will demonstrate how their system utilizes Milvus to store and manipulate embeddings for generating privacy-protected synthetic data. Their approach not only maintains the confidentiality of the original data but also enhances the utility and scalability of LLMs under privacy constraints. Attendees, including machine learning engineers, data scientists, and data managers, will witness first-hand how Secludy's integration with Milvus empowers organizations to harness the power of LLMs securely and efficiently.
Main news related to the CCS TSI 2023 (2023/1695)Jakub Marek
An English 🇬🇧 translation of a presentation to the speech I gave about the main changes brought by CCS TSI 2023 at the biggest Czech conference on Communications and signalling systems on Railways, which was held in Clarion Hotel Olomouc from 7th to 9th November 2023 (konferenceszt.cz). Attended by around 500 participants and 200 on-line followers.
The original Czech 🇨🇿 version of the presentation can be found here: https://www.slideshare.net/slideshow/hlavni-novinky-souvisejici-s-ccs-tsi-2023-2023-1695/269688092 .
The videorecording (in Czech) from the presentation is available here: https://youtu.be/WzjJWm4IyPk?si=SImb06tuXGb30BEH .
Unlock the Future of Search with MongoDB Atlas_ Vector Search Unleashed.pdfMalak Abu Hammad
Discover how MongoDB Atlas and vector search technology can revolutionize your application's search capabilities. This comprehensive presentation covers:
* What is Vector Search?
* Importance and benefits of vector search
* Practical use cases across various industries
* Step-by-step implementation guide
* Live demos with code snippets
* Enhancing LLM capabilities with vector search
* Best practices and optimization strategies
Perfect for developers, AI enthusiasts, and tech leaders. Learn how to leverage MongoDB Atlas to deliver highly relevant, context-aware search results, transforming your data retrieval process. Stay ahead in tech innovation and maximize the potential of your applications.
#MongoDB #VectorSearch #AI #SemanticSearch #TechInnovation #DataScience #LLM #MachineLearning #SearchTechnology
Taking AI to the Next Level in Manufacturing.pdfssuserfac0301
Read Taking AI to the Next Level in Manufacturing to gain insights on AI adoption in the manufacturing industry, such as:
1. How quickly AI is being implemented in manufacturing.
2. Which barriers stand in the way of AI adoption.
3. How data quality and governance form the backbone of AI.
4. Organizational processes and structures that may inhibit effective AI adoption.
6. Ideas and approaches to help build your organization's AI strategy.
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).
Digital Marketing Trends in 2024 | Guide for Staying AheadWask
https://www.wask.co/ebooks/digital-marketing-trends-in-2024
Feeling lost in the digital marketing whirlwind of 2024? Technology is changing, consumer habits are evolving, and staying ahead of the curve feels like a never-ending pursuit. This e-book is your compass. Dive into actionable insights to handle the complexities of modern marketing. From hyper-personalization to the power of user-generated content, learn how to build long-term relationships with your audience and unlock the secrets to success in the ever-shifting digital landscape.
Project Management Semester Long Project - Acuityjpupo2018
Acuity is an innovative learning app designed to transform the way you engage with knowledge. Powered by AI technology, Acuity takes complex topics and distills them into concise, interactive summaries that are easy to read & understand. Whether you're exploring the depths of quantum mechanics or seeking insight into historical events, Acuity provides the key information you need without the burden of lengthy texts.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
Best 20 SEO Techniques To Improve Website Visibility In SERPPixlogix Infotech
Boost your website's visibility with proven SEO techniques! Our latest blog dives into essential strategies to enhance your online presence, increase traffic, and rank higher on search engines. From keyword optimization to quality content creation, learn how to make your site stand out in the crowded digital landscape. Discover actionable tips and expert insights to elevate your SEO game.
Best 20 SEO Techniques To Improve Website Visibility In SERP
Forest Area Estimation in Kutai Nasional Park of East Kalimantan Using Computer System Aplication based Genetic Algorithm-Support Vector Machine
1. International Journal of Computer Applications Technology and Research
Volume 5– Issue 5, 253 - 259, 2016, ISSN:- 2319–8656
www.ijcat.com 253
Forest Area Estimation in Kutai Nasional Park
of East Kalimantan Using Computer System Aplication
based Genetic Algorithm-Support Vector Machine
Lapu Tombilayuk
Informatics Department, Sekolah Tinggi Teknologi Bontang,
Bontang East Kalimantan, Indonesia
Abstract : The paper presents design of computer system application for the forest area estimation using the combination of genetic
algorithm (GA) and support vector machine SVM) methods in Kutai National park of East Borneo. The considering variables such as
reboization concerning natural green, forest fire, encroachment and illegal logging activities are the basic data for our proposed design. In
the development of design computer system application, the Unified Modeling Language is adopted with the use case, activity diagrams and
sequence diagrams are systematically followed in order to keep the purpose of design on track. The supporting instrument for this research is
the language programming of Borland Delphi 7 and MySQL database. The accuracy of area estimation result is compared with the actual
data using mean absolute percentage error (MAPE).
Keywords: GA-SVM methods, UML, Borland Delphi, MySQL database, white box testing.
1. Introduction
Indonesia has continuously experienced in deforestation area
where the majority of forest damaging occurs in Borneo and
Sumatera. The causes of deforestation are forest mismanagement,
illegal logging, forest fire and forest opening for farming and
mining purposes. The effects of forest damaging can be locally,
such as ecologic disaster, flood, soil avalanche and can be globally
as well, such as drought and other global warming effects. The
government is actually proactive to do the forest conservation
program with policies, for instance the protection of primary
natural and peat forests and the strict regulation of forest
utilization for mining purpose. The government regulation on
forest conservation is highly protected by law; in fact,
deforestation seems uncontrollable. The organization of forest
watch Indonesia (FWI) reported that the forest in Java will be used
up in 2020 with small width area remaining in Bali and Nusa
Tenggara (0.08 million h.a), Sulawesi (7.2 million h.a), Sumatera
(7.72 million h.a), Borneo (21.29 million h.a) and Papua (33.45
million h.a). The rate of deforestation is about 1.51 million h.a per
year in the period of 2000-2009 with the highest rate of 0.55
million h.a per year occurs in Borneo.
Forest area estimation method has gained important awareness of
researchers with variety methods. The statistic methods are still
dominant in this topic. With the hierarchy of Bayesian model, the
forest of area can be accurately classified up to 88%. Another
approach by the forest sampling method is used to estimate the
forest canopy cove according to probability theory. In addition,
the forest cover estimation based the importance vegetation type
has been proposed using K-means method on the time series data.
The research is focused on the clustering of different type of
vegetation. Meanwhile, the estimation of forest canopy by
comparison of the field measurement technique has been
conducted by pictures result of digital camera.
The results of the previous research motivate us to find the best
method to estimate the forest area based on the conditions of
reboization concerning natural green, forest fire, encroachment
and illegal logging activities. The parametric method by mean
mathematic model and statistic, such as autoregressive integrated
moving average (ARIMA) is less suitable for this case study due
to the difficulty in modeling irregular and variable number of data.
On the other hand, the non-parametric method with exponential
technique is only superior for the short-term forecasting and the
results may not be confirmed optimal. In addition, the artificial
neural network is facing difficulty to provide solution with time-
series data.
2. Configuration of proposed system
The time series estimation method has continuously attracted
serious attention from scientific community. The method is
basically part of computational intelligence utilizing historical
data to solve estimation problems. This kind of technique is
possible supporting fromthe advanced computational technique
and information technology. One of the computational techniques
based machine learning is the Support Vector Machine (SVM).
This method is superior to classify the input data set from the
minimum to the maximum values. The classification results are
used as chromosome for the genetic algorithm (GA) operation.
This combination accelerates the computational efforts and
provides high accuracy estimation compared to the only GA
process. In this case, the output genetic algorithm (GA) is the
optimal
estimated value of forest area. It also implies that the uncorrelated
data classification is avoided, as results only the important inputs
are considered by the implementation of GA-SVM method. In this
research, the main configuration of the estimation method for the
forest area in Kutai National park, East Borneo is divided into the
input database of including positive, like reboization concerning
natural green and negative causes, such as forest fire,
encroachment and illegal logging activities that affecting to the
forest area, processing database using GA-SVM methods and the
area forest estimation output. The database system is stored using
MySQL software system. The genetic algorithm utilizes these data
to initialize parameter and to generate the population. Later, the
support vector machine method evaluates the fitness function in
order obtain the data set. Then, the data set is reselected to obtain
two chromosomes with the best fitness function by the genetic
algorithm. By this approach, the only correlated data is used for
the GA process, results in high accuracy estimation. The process
and evaluation continue with crossover and mutation of new
generation until it convergences at the best intent of fitness
function.
The input database is taken directly from Kutai National Park
office during the last 10 years (between 2003 and 2012) about data
record of reboization, forest fire, encroachment and illegal logging
activities.
These data is quite random and irregular due to the cause
combination between the nature and human activities. In these
2. International Journal of Computer Applications Technology and Research
Volume 5– Issue 5, 253 - 259, 2016, ISSN:- 2319–8656
www.ijcat.com 254
data, the average forest damaging area based on the negative
causes is about 125.68 h.a, with the mean area of reboization is
about 54.5 h.a. In 2013, the total forest area of Kutai National ark,
East Borneo is about 198,629 h.a with damaging area about 711.8
h.a. Based on this reason, it is important to provide some tools to
estimate the forest area for some time in the future, so that it
becomes reference to measure and to accelerate the green
activities inside the National Park.
The forest area estimation is performed by the implementation of
combination between the genetic algorithm (GA) and support
vector machine (SVM) method based on the storing data in the
database system. The superiority of this method is in the capability
of SVM method to divide vector space into hyperplane according
to the data trend from small area to wide area classes. The
algorithm combination for such estimation technique is able to
provide better solution compared to other estimation techniques,
such as artificial neural network because the error and
generalization of the SVM method is not depending on the input
vector. The complete process of GA-SVM methods is shown by
the pseudo-code as follows:
1) Initialization process by the genetic algorithm to select the
chromosome candidates of data stored in database.
2) Initialization of data set by running the SVM method to
search all data set about reboization, forest fire,
encroachment and illegal logging activities between 2003
and 2012 in the data base system.
3) Run the polynomial Kernel function to classify the non-
linear data into two classes by (XT.Xi + 1)P. The results are
about the influence causes, measured from the most to the
less value impacts.
4) The training process of all selected data set by means the
influenced parameters to the forest area. The selection is
aimed to obtain the hyperplane of y(x)f(x)=1 that separates 2
classes. The candidates of support vector are:
y(x)f(x)+β+1 and y(x)f(x)1 (1)
where Xa=XaXb and β is the arbitrarily determined value
by users.
If the re-training process is conducted, then the previous
training results are improved by taking only some data
(XXa) as the support vector candidates.
5) The data classification after training process is divided into
xi.w+b1 for the 1st class and xi.w+b-1 for the 2nd class.
6) Chromosomes selection. The best chromosome is usually
selected by objective function evaluation with defined high
probability. The roulette-wheel method is used in this step.
7) Fitness function evaluation by:
Fitness = C – f(x) or Fitness =
𝐶
𝐹(𝑥)+ Є
(2) (2)
where C is a constant and Є is small number to avoid zero
division. In this research, The fitness function = forest fire +
encroachment + illegal logging – rebozation
8) Selection process with Linier Fitness Ranking (LFR).
9) Cross-Over process. One-cut point method is used to
exchange the gen of the parent chromosome with cross-over
probability (Pc) of 0.25.
Begin
k 0;
While (k < populasi) do
R[k] random [0 - 1]; (3) (4)
If (R[k] < Pc) then
Select Chromosome[k] as parent;
End;
10) Mutation process. The mutation rate is specified at 0.1.
11) Population replacement by the new generation
12) If the new generation convergences to the optimal solution,
the overall process stop; otherwise the process is repeated
from no. 10.
The above pseudocode is translated into language
programming of Borland Delphi 7. The mean absolute percentage
error (MAPE) is used to validate the accuracy in output
measurement. In addition, the white box testing is used to evaluate
all computer logic programming by checking the logical iteration
and to assess the overall data used in the simulation.
3. Design of Computer Application for Forest
Area Estimation
In this section, it will be explained about the designs information
related to the unified modeling language (UML), database and
user interface. The detailed information of such design is provided
as follows:
3.1 Design of Unified Modeling Language
(UML)
The unified Modeling Language (UML) is the standar design and
documentation of software system with the capability of software
application modelling according to the hardware configuration,
operating system and it has the flexibility ain network design and
language programming. The UML diagram can be built-up using
use case, class, activity, sequence and statechart diagrams. With
the UML design, the programmers may expect the model runs
perfectly and accurately with considering the scalability,
robustness and security. The use case diagram represents the
expected functionality of designed system and also can be
considered as the interaction mechanism between the user and the
system. It consists of case login to access the system, case view
menu with several menu options, case input data of such the
causes that influences to forest area, case analysis by means the
GA-SVM method and the case output as the estimated forest area.
The input data processing may be performed by an administrator
and the output data is utilized by the National Park Authority in
order to anticipate the negative causes and continuing promote the
green activities. The use case diagram is shown in Figure 1.
Figure 1. Use Case Diagram
3. International Journal of Computer Applications Technology and Research
Volume 5– Issue 5, 253 - 259, 2016, ISSN:- 2319–8656
www.ijcat.com 255
The class diagram represents the class of structure and its
description, package and object including the connectivity within
the application design. Figure 2 shows the class description of our
proposed application. There are 3 data classes, i.e input data,
analysis and report. The class of input data consists of causes
identity, year, location, causes and width area fields. Meanwhile,
the class of analysis is the fields of number, year, causes and
width area. The last class is the class of report which comprises of
fields of report identity and its type. The chart of class diagram is
shown in Figure 2.
Figure 2. Class diagram
The next diagram in UML development is the activity diagram
that describes the activity path inside the designed system, to how
the system starts to end including the possibility decision that
might be appear.
Amongst the developed diagram, the activity diagram is the
special state diagram where most of the states are an action,
triggered transition and internal processing by the previous states.
In this research, the activity diagram explains the system process
from the data inputting by an administrator, data processing using
GA-SVM method and the estimation output utilization by
the National park authority. The activity diagram is shown in
Figure 3.
Figure 3. Actifity Diagram
The object interaction inside and surrounding the system including
the user, display and so on is described by sequential diagram in
the forms of messages per time value. The sequential diagram can
also be used to express the scenarios or steps that should be
followed according to the output events. It may start from the
causes that trigger the next activity, to what kind of internal
changes and typical. Each object including administrator has
vertical life line. The message is depicted by arrows from an
object to other objects. In the next phase design, such messages
are mapped into operation or method of classes. Figure 4 explains
the object behavior when the operator interacts with application
system. Administrator needs login to input the causes that
influence to the total forest area into the database. Later, this data
is processed by the GA-SVM method to yield total forest area
estimation which can be used by the National park authority.
Figure 4. Sequential diagram
To support all diagrams process, it requires the state chart diagram
input data that describes transition states or changes of object
inside the systems. The state chart diagram is shown in Figure 5.
In this figure, if the operators makes mistake during data inputting
process, they have opportunity to do editing then resave the data
into the database. If the cancelation of inputting data occurs, the
system will terminate. The inputting data process into the database
keeps going until the data is finished and saved. The remained
process is just waiting the execution of GA-SVM method to yield
the estimated forest area. In addition, the state chart diagram may
describe the data analysis using GA-SVM method and report
analysis as the outcome of algorithm.
4. International Journal of Computer Applications Technology and Research
Volume 5– Issue 5, 253 - 259, 2016, ISSN:- 2319–8656
www.ijcat.com 256
Figure 5. Statechart diagram of input and edit data
3.2 Design of database
The database design is used to map the conceptual model
into the basis data model. In this research, the structure data
file about the causes affecting the forest area located in the
database is shown in Table 1. There are five field names but
the most important field (primary) is the field of number.
When transforming these data into terms of generation data
in Genetic Algorithm called the data identity of generation,
there are four field names. Again, the most important field
(primary) is the field of number. The data identity of
generation is shown in Table 2.
Table 1. Data identity of causes influenced to the forest area
Table 2. Data identity of generation
In addition, all data in the database system in cooperating with
software system is stored with MySQL basis management system.
MySQL system is free software (open source) under the license of
general public license (GPL) both the source code and executable
program. The other advantage of MySQL are portability,
multiuser, flexible tuning performance, high security and scalable.
Therefore, design such computer application system is nowadays
very convenient with maximal performance.
3.3. Design of User interface
The proposed user interface is the displays of the main menu,
input data, initial data, generation and estimation. These displays
are explained as follows. The display of main menu for the forest
area estimation in Kutai National Park of East Borneo as shown in
Figure 6 contains the communicative information related to input
data, analysis and report. The display explains the main front of
application on how the users interact to the system application
simply. When the main display is active, the ser may input data by
simple click. If they continue for the analysis and report, they may
just precede the previous activity.
Figure 6. Display of main menu
In the inputting data process, the users may input data of year,
region, and causes both positive and negative. The year data is
specified from 2003 to 2012 with three definite regions where
region I is called Suka Rahmat, region II is Sangatta and region III
is Manamang. The last required data are the causes and their
affected total area. The causes influenced to the forest area in the
park are reboization, forest fire, encroachment and illegal logging
activities. The page display for the input data process is shown in
Figure 7.
Figure 7. Interface display for input data
After the input data process is complete, the user may check their
data stored in the database. The typical initial data stored in
database system is shown in Figure 8. It is clearly shown that
there are certain damaged forest area affected, forest fire,
encroachment and illegal logging activities even though the
reboization contributes positively in this matter. Total forest area
in Kutai National Park, East Borneo is 198,629 h.a. However, it
has been recently found that the total forest area reduced to
197,917 h.a by considering the above causes. In the current
display, the users may continue the process until they obtain the
estimated area in future years.
5. International Journal of Computer Applications Technology and Research
Volume 5– Issue 5, 253 - 259, 2016, ISSN:- 2319–8656
www.ijcat.com 257
Figure 8. Initial input data display in database system
If the process continues, then the users may receive information
about the estimation of forest area in the following years. In this
simulation, the maximum width of forest area (the chromosome
value) of the generation is the estimation output. It is due the
consideration of the 10 years previous data to predict the forest
area in the year after. For example in Figure 9, the chromosome
value change from the initial value (data from 2003 to 2012) to
other number specified that the damaged forest area in 2013 is
784.7 h.a. With the same consideration in the following
generation, it will be the estimation results of the year 2014, and
so on. With this application system, the estimated damaged
forest area in the years of 2014, 2015 and 2016 are 905.1 h.a,
1,444.7 h.a and 2,211.2 h.a, respectively. These results are
obtained with assumptions that there is no updating data from
the initial data condition. The estimated results in 2016 may less
than the above estimated number if the reboization activity
increases and one of the negative causes can be pressed down.
Figure 9. Typical display of damage forest area
estimation in 2013
If the process continues, then the users may receive information
about the estimation of forest area in the following years. In this
simulation, the maximum width of forest area (the chromosome
value) of the generation is the estimation output. It is due the
consideration of the 10 years previous data to predict the forest
area in the year after. For example in Figure 10, the chromosome
value change from the initial value (data from 2003 to 2012) to
other number specified that the damaged forest area in 2013 is
784.7 h.a. With the same consideration in the following
generation, it will be the estimation results of the year 2014, and
so on. With this application system, the estimated damaged forest
area in the years of 2014, 2015 and 2016 are 905.1 h.a, 1,444.7 h.a
and 2,211.2 h.a, respectively. These results are obtained with
assumptions that there is no updating data from the initial data
condition. The estimated results in 2016 may less than the above
estimated number if the reboization activity increases and one of
the negative causes can be pressed down.
Figure 10. Typical display of damage forest area
estimation in 2013
The design of application program is complete with the display of
graph estimation as shown in Figure 11. In this figure, it is shown
that in certain period the damage forest area decreases. For
instance, there is significant reduction of forest area to about 939
h.a in 2017 and 358.4 h.a in 2020 from the previous year
conditions. However, this condition is some kind of transition
states because the total damaged area rises again to 1,883.9 h.a
and 2,341.3 h.a in the years of 2018 and 2019, respectively. As
previously mentioned that the current forest area of Kutai National
Park, East Borneo is 198,629 h.a. If the prediction process
continues without any changes in the input data by means there is
no positive action for the green activity, the forest area is used up
in 2061 because the estimated damage area in this year has
reached 200,129 h.a which is far beyond the current forest area.
Obtaining such number is important for the local people, park
authority and global society as the reference to do positive action
for the forest conservation since the Kutai National Park of East
Borneo is ‘the lung of world’.
Figure 11. Typical display of estimation results
with graph estimation
6. International Journal of Computer Applications Technology and Research
Volume 5– Issue 5, 253 - 259, 2016, ISSN:- 2319–8656
www.ijcat.com 258
4. Simulation results and discussion
In the design of system application, the high accuracy estimation
or prediction is the most important aspect to be considered in
order to guarantee the outcomes are on the right value even the
data input collection is abruptly changed. The accuracy
assessment in this research is by measuring the Mean Absolute
Percentage Error (MAPE) between the estimated area as the
output of the designed system application and the actual area. In
addition, the estimated output area is also compared with the
conventional linear regression with similar MAPE performance
index measurement.
The first scenario is the comparison between the estimated area
and the actual area. For the easy comparison, the data of previous
five years from 2003 to 2007 is arbitrarily selected as the initial
input data (training data) because we have the figure actual data
from 2008 to 2012 for the validation data. The result of simulation
under scenario is shown in Fig. 11. In the simulation results, we
have comparing data between estimated area and actual area from
2008 to 2012. These data is evaluated by Mean Absolute
Percentage Error (MAPE) equation as follows:
where At is the actual area in the t year, Et is the estimate area in
the t year and N is the period of data evaluation (N=5) in this
research. The MAPE index performance calculation is
summarized in Table III. For the five years period of estimation,
the average of MAPE is about 2.1 % with the maximum
difference of 102.8 h.a in 2008 and minimum difference of 8.8 h.a
in 2011. It means that the proposed system application to estimate
the forest area of Kutai National Park, East Borneo is guarantee
small. Therefore, the estimated forest area until the year of 2020
may present proper results. Typical display of estimation results
for the following years has been shown in Figure 12.
Figure 12. Simulation result for the estimated
and actual area comparison
Table 3. MAPE performance index measurement
The last performance testing to our proposed system design is the
White Box Testing. The test is conducted to see the module
contents in order to evaluate the property of coding program. The
test may cover the logical statements and their decision, iteration
process within its constraints and the overall internal data structure
to guarantee the validity of program. Basically, the white box
testing is the path testing to allow the programmers to measure the
logical complexity of procedural design and to use this
measurement as the guidance to define the path set. The white box
testing of this research is shown in Figure 13. There are 5 regions,
denoted with R1= initial population, R2= chromosome selection,
R3= crossover, R4= mutation and R5 = generation iteration.
According to the simulator Flowgraph application, the Edge
number is 14 and the Node number is 11; therefore the
Cyclometric Complexity (CC) is equal to 5. Also, we obtain 5
paths from Fig. 12. which are Path 1: A – B – C – D – E – D – E –
F – G – H – I – J – K; Path 2:A – B – C – D – E – F – G – F – G –
H – I – J – K; Path 3: A – B – C – D – E – F – G – H – I – H – I –
J – K; Path 4: A – B – C – D – E – F – G – H – I – J – C – D – E –
F –G – H – I – J – K and Path 5: A – B – C – D – E – F – G – H –
I – K. Because of the proposed system application has 5 regions, 5
CCs and 5 paths, the analysis application can be claimed to be
properly correct.
Figure 13. The flow graph in white box testing
(4)
7. International Journal of Computer Applications Technology and Research
Volume 5– Issue 5, 253 - 259, 2016, ISSN:- 2319–8656
www.ijcat.com 259
5. Conclusion
The paper presents the design computer system application to
estimate the forest area in Kutai National Park, East Borneo based
the combination of genetic algorithm (GA) and support vector
machine (SVM). In the development of design system application,
the Unified Modeling Language is adopted with the case, activity
and sequence diagrams are systematically followed in order to
keep the purpose of design on track considering variables such as
reboization concerning natural green, forest fire, encroachment
and illegal logging as the input data in database system. The
supporting instrument for this research is the language
programming of Borland Delphi 7 and MySQL database. Without
any actions to the initial data by means there is no significant
effort to do the forest conservation program by the park authority,
the forest area will be finished by the year of 2061. It is very
dangerous situation to the world ecosystem because this park is
‘the lung of the world’. Provision initial data information is very
urgently to save our environment. The accuracy of area estimation
result is compared with the actual data using mean absolute
percentage error (MAPE) with the average error of 2.1%. In
addition, the white box testing to calculate region, cyclometric
complexity and path is implemented to confirm the correctness of
the logic algorithm of design application program.
References
[1] Wirenro Sumargo, Soelthon Gussedya Nanggara, Frionny
Nainggolan, Isnenti Apriani: “Potret Keadaan Hutan
Indonesia Periode tahun 2000-2009”, Forest Watch
Indonesia, 2011, pp. 20-22.
[2] Peraturan Pemerintah Republik Indonesia Nomor 60 Tahun
2012: “Perubahan Atas Peraturan Pemerintah Nomor 10
Tahun 2010 Tentang Tata Cara Perubahan Peruntukan dan
Fungsi Kawasan Hutan”, 2012.
[3] Dudy Subagdja: “Nasib Hutan Kita dan Kebijakan Ekonomi
Hijauh”, Berita Kompasiana, 29 Maret 2013.
[4] Christina Basaria S.: “Kajian Kelestarian Tegakan Dan
Produksi Kayu Jati Jangka Panjang KPH Bojonegoro
Perum Perhutani Unit II Jawa Timur”, Institut Pertanian
Bogor, 2009.
[5] Dyah Pratiwi, et.al: “Penghitungan laju luas area hutan
berbasis algoritma segmentasi warna local”, Konferensi
Nasional Sistem Informasi 2013, pp. 471-475.
[6] F. Deppe: "Forest Area Estimation Using Sample Survey and
Landsat MSS and TM Data”, Photogrammetric Engineering
& Remote Sensing, Vol.6+, No.4, 1998, pp. 285-292.
[7] Dang Khoi dan Yuji Murayama: “Forecasting Areas
Vulnerable to Forest Conversion in the Tam Dao
National Park Region”, Remote Sens. Vol. 2, 2010, pp.
1249-1272.
[8] Loghman Ghahramany, Pariz Fatehi, Hedayat Ghazanfari:
“Estimation of Basal Area in West Oak Forests of Iran
Using Remote Sensing Imagery”, International Journal of
Geosciences, Vol. 3, 2012, pp. 398-403.
[9] Oliver Diedershagen, Barbara Koch: “Automatic Estimation
Of Forest Inventory Parameters Based
on Lidar, Multi-Spectral and Fogis Data”, Holgen
Weinacker, University Freiburg, Germany, 2003,
pp. 4-13.
[10] Andrew O. Finley, et.al.: “A Bayesian approach to multi-
source forest area estimation”, USA, Environ Ecol Stat, Vol.
15, 2008, pp.241–258.
[11] Farshad Keivan Behjou, Mahbobeh Foshat: “Using A
Sampling Method for Estimation of Forest
Canopy cover”, International Journal of Agriculture:
Research and Review. Vol., 3 (2), 2013, pp. 217-222.
[12] Anuj Karpatne, et.al.:, “Importance of Vegetation Type in
Forest Cover Estimation”, University of Minnesota, 2010.
[13] Lauri Korhonen, Kari T. Khorhonen, Miina Rautianen and
Pauline Stenberg: “Estimation of Forest Canopy Cover: a
Comparison of field measurement Techniques”, Silva
Fennica, Vol. 40, No. 4, 2006, pp.577-588.
[14] Thi Nguyen, Lee Gordon-Brown, Peter Wheeler, Jim
Peterson: “GA-SVM Based Framework for Time Series
Forecasting”, the fifth International Conference on Natural
Computation 2009, pp.493-498.