Automatically fruit recognition by using machine vision is considered as challenging task due to similarities between various types of fruits and external environmental changes e-g lighting. In this paper, fruit recognition algorithm based on Deep Convolution Neural Network(DCNN) is proposed. Most of the previous techniques have some limitations because they were examined and evaluated under limited dataset, furthermore they have not considered external environmental changes. Another major contribution in this paper is that we established fruit images database having 15 different categories comprising of 44406 images which were collected within a period of 6 months by keeping in view the limitations of existing dataset under different real-world conditions. Images were directly used as input to DCNN for training and recognition without extracting features, besides this DCNN learn optimal features from images through adaptation process. The final decision was totally based on a fusion of all regional classification using probability mechanism. Experimental results exhibit that the proposed approach have efficient capability of automatically recognizing the fruit with a high accuracy of 99% and it can also effectively meet real world application requirements.
Analysis of pancreas histological images for glucose intollerance identificat...Tathagata Bandyopadhyay
This document presents a method for analyzing pancreas histological images to identify glucose intolerance using wavelet decomposition. 134 pancreas images from rats were used, with 56 normal and 78 pre-diabetic. Wavelet analysis was used to segment beta cell regions. Features like area ratios were extracted and classifiers like SVM, MLP, Naive Bayes were applied, achieving up to 84% accuracy. Future work could introduce more robust features and faster segmentation methods to improve classification.
This document presents a study that uses machine learning techniques to classify lumbar intervertebral disc degeneration in MRI images. 181 MRI images were analyzed to extract texture features and train a decision tree classifier. The classifier achieved 93.33% accuracy in multi-class multi-label classification of discs as normal or degenerated, and identifying the specific affected disc. This automated classification approach could help with medical diagnosis and image retrieval for orthopedists.
This document proposes a method to detect diseases in tree leaves and fruits using convolutional neural networks. The method involves preprocessing images by resizing and converting them to grayscale. Noise is removed using median filtering. Images are then segmented and features are extracted. A convolutional neural network is trained on labeled image features and used to classify new images and detect diseases. Results on test images show the preprocessing steps and segmentation being applied before classification with CNN. The method is presented as an effective way to accurately detect diseases at early stages for successful agriculture.
Segmentation of Brain MR Images for Tumor Extraction by Combining Kmeans Clus...CSCJournals
Segmentation of images holds an important position in the area of image processing. It becomes more important while typically dealing with medical images where pre-surgery and post surgery decisions are required for the purpose of initiating and speeding up the recovery process [5] Computer aided detection of abnormal growth of tissues is primarily motivated by the necessity of achieving maximum possible accuracy. Manual segmentation of these abnormal tissues cannot be compared with modern day’s high speed computing machines which enable us to visually observe the volume and location of unwanted tissues. A well known segmentation problem within MRI is the task of labeling voxels according to their tissue type which include White Matter (WM), Grey Matter (GM) , Cerebrospinal Fluid (CSF) and sometimes pathological tissues like tumor etc. This paper describes an efficient method for automatic brain tumor segmentation for the extraction of tumor tissues from MR images. It combines Perona and Malik anisotropic diffusion model for image enhancement and Kmeans clustering technique for grouping tissues belonging to a specific group. The proposed method uses T1, T2 and PD weighted gray level intensity images. The proposed technique produced appreciative results
Alzheimer’s disease(AD) is a neurological disease. It affects memory. The livelihood of the people that are
diagnosed with AD. In this paper, we have discussed various imaging modalities, feature selection and
extraction, segmentation and classification techniques.
1) The document discusses various medical image fusion techniques including pixel level, feature level, and decision level fusion.
2) It proposes a novel pixel level fusion method called Iterative Block Level Principal Component Averaging fusion that divides images into blocks and calculates principal components for each block.
3) Experimental results on fusing noise free and noise filtered MR images show that the proposed method performs well in terms of average mutual information and structural similarity compared to other algorithms.
ENHANCED SYSTEM FOR COMPUTER-AIDED DETECTION OF MRI BRAIN TUMORSsipij
This paper presents a system for detecting and classifying brain tumors in MRI images. Features were extracted from 105 brain images, including mean, standard deviation, and derivative. Two classifiers, SVM and KNN, were tested on the features to classify images as normal or abnormal. The SVM classifier achieved 100% accuracy on the test set, demonstrating the system's ability to successfully separate the two classes.
Analysis of pancreas histological images for glucose intollerance identificat...Tathagata Bandyopadhyay
This document presents a method for analyzing pancreas histological images to identify glucose intolerance using wavelet decomposition. 134 pancreas images from rats were used, with 56 normal and 78 pre-diabetic. Wavelet analysis was used to segment beta cell regions. Features like area ratios were extracted and classifiers like SVM, MLP, Naive Bayes were applied, achieving up to 84% accuracy. Future work could introduce more robust features and faster segmentation methods to improve classification.
This document presents a study that uses machine learning techniques to classify lumbar intervertebral disc degeneration in MRI images. 181 MRI images were analyzed to extract texture features and train a decision tree classifier. The classifier achieved 93.33% accuracy in multi-class multi-label classification of discs as normal or degenerated, and identifying the specific affected disc. This automated classification approach could help with medical diagnosis and image retrieval for orthopedists.
This document proposes a method to detect diseases in tree leaves and fruits using convolutional neural networks. The method involves preprocessing images by resizing and converting them to grayscale. Noise is removed using median filtering. Images are then segmented and features are extracted. A convolutional neural network is trained on labeled image features and used to classify new images and detect diseases. Results on test images show the preprocessing steps and segmentation being applied before classification with CNN. The method is presented as an effective way to accurately detect diseases at early stages for successful agriculture.
Segmentation of Brain MR Images for Tumor Extraction by Combining Kmeans Clus...CSCJournals
Segmentation of images holds an important position in the area of image processing. It becomes more important while typically dealing with medical images where pre-surgery and post surgery decisions are required for the purpose of initiating and speeding up the recovery process [5] Computer aided detection of abnormal growth of tissues is primarily motivated by the necessity of achieving maximum possible accuracy. Manual segmentation of these abnormal tissues cannot be compared with modern day’s high speed computing machines which enable us to visually observe the volume and location of unwanted tissues. A well known segmentation problem within MRI is the task of labeling voxels according to their tissue type which include White Matter (WM), Grey Matter (GM) , Cerebrospinal Fluid (CSF) and sometimes pathological tissues like tumor etc. This paper describes an efficient method for automatic brain tumor segmentation for the extraction of tumor tissues from MR images. It combines Perona and Malik anisotropic diffusion model for image enhancement and Kmeans clustering technique for grouping tissues belonging to a specific group. The proposed method uses T1, T2 and PD weighted gray level intensity images. The proposed technique produced appreciative results
Alzheimer’s disease(AD) is a neurological disease. It affects memory. The livelihood of the people that are
diagnosed with AD. In this paper, we have discussed various imaging modalities, feature selection and
extraction, segmentation and classification techniques.
1) The document discusses various medical image fusion techniques including pixel level, feature level, and decision level fusion.
2) It proposes a novel pixel level fusion method called Iterative Block Level Principal Component Averaging fusion that divides images into blocks and calculates principal components for each block.
3) Experimental results on fusing noise free and noise filtered MR images show that the proposed method performs well in terms of average mutual information and structural similarity compared to other algorithms.
ENHANCED SYSTEM FOR COMPUTER-AIDED DETECTION OF MRI BRAIN TUMORSsipij
This paper presents a system for detecting and classifying brain tumors in MRI images. Features were extracted from 105 brain images, including mean, standard deviation, and derivative. Two classifiers, SVM and KNN, were tested on the features to classify images as normal or abnormal. The SVM classifier achieved 100% accuracy on the test set, demonstrating the system's ability to successfully separate the two classes.
SEGMENTATION OF LUNG GLANDULAR CELLS USING MULTIPLE COLOR SPACESIJCSEA Journal
Early detection of lung cancer is a challenging problem, the world faces today. Prior to classify glandular cells as malignant or benign a reliable segmentation technique is required. In this paper we present a novel lung glandular cell segmentation technique. The technique uses a combination of multiple color spaces and various clustering algorithms to automatically find the best possible segmentation result. Unsupervised clustering methods of K-means and Fuzzy C-means were used on multiple color spaces such as HSV, LAB, LUV, xyY. Experimental results of segmentation using various color spaces are provided to show the performance of the proposed system.
Utilization of Super Pixel Based Microarray Image Segmentationijtsrd
In the division of PC vision pictures, Super pixels are go probably as key part from 10 years prior. There are various counts and methodology to separate the Super pixels anyway whole all of them the best super pixel looking at strategy is Simple Linear Iterative Clustering SLIC have come to pivot continuously recently. The concentrating of small scale group quality verbalization from MRI imaging is more useful to perceive tumors or some other dangerous development contaminations, so the fundamental DNA cDNA microarray is a grounded device for analyzing the same. The division of microarray pictures is the essential development in a microarray assessment. In this paper, we proposed a figuring to dividing the cDNA small show picture using Simple Linear Iterative Clustering SLIC based Self Organizing Maps SOM method. In any case, the proposed figuring is taken up a moving task to look at the bad quality of pictures in addition. There are two phases to separate the image, introductory, a pre setting up the applied picture to diminish fuss levels and second, to piece the image using SLIC based SOM approach. Mr. Davu Manikanta | Mr. Parasurama N | K Keerthi "Utilization of Super Pixel Based Microarray Image Segmentation" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-5 , August 2021, URL: https://www.ijtsrd.com/papers/ijtsrd46274.pdf Paper URL: https://www.ijtsrd.com/engineering/electronics-and-communication-engineering/46274/utilization-of-super-pixel-based-microarray-image-segmentation/mr-davu-manikanta
Multimodal Medical Image Fusion Based On SVDIOSR Journals
Image fusion is a promising process in the field of medical image processing, the idea behind is to
improve the content of medical image by combining two or more multimodal medical images. In this paper a
novel fusion framework based on singular value decomposition - based image fusion algorithm is proposed.
SVD is an image adaptive transform, it transforms the matrix of the given image into product USVT
, which
allows to refactor a digital image into three matrices called tensors. The proposed algorithm picks out
informative image patches of source images to constitute the fused image by processing the divided subtensors
rather than the whole tensor and a novel sigmoid-function-like coefficient-combining scheme is applied to
construct the fused result. Experimental results show that the proposed algorithm is an alternative image fusion
approach.
Flower Classification Using Neural Network Based Image ProcessingIOSR Journals
Abstract: In this paper, it is proposed to have a method for classification of flowers using Artificial Neural Network (ANN) classifier. The proposed method is based on textural features such as Gray level co-occurrence matrix (GLCM) and discrete wavelet transform (DWT). A flower image is segmented using a threshold based method. The data set has different flower images with similar appearance .The database of flower images is a mixture of images taken from World Wide Web and the images taken by us. The ANN has been trained by 50 samples to classify 5 classes of flowers and achieved classification accuracy more than 85% using GLCM features only. Keywords: Artificial Neural Network, DWT, GLCM, Segmentation.
Classification of MR medical images Based Rough-Fuzzy KMeansIOSRJM
The document summarizes a proposed algorithm for classifying MR medical images using Rough-Fuzzy K-Means (FRKM). It begins with an introduction to the challenges of medical image classification and a literature review of previous techniques. It then provides background on rough set theory, fuzzy set theory, and K-means clustering. The proposed FRKM algorithm is described as using rough set theory for feature selection and dimensionality reduction, followed by a K-means clustering with probabilities assigned based on rough set approximations to classify ambiguous areas. Experimental results show the FRKM approach achieves 94.4% accuracy, higher than other techniques.
Quality Compression for Medical Big Data X-Ray Image using Biorthogonal 5.5 W...IJERA Editor
Medical Big Data (MBD) consists of very useful type of information. It is very important for a physician for decision making and treatments to cure the patient. For accurate diagnosis, data availability is the most important factor. MBD over network needs intelligent compression schemes so that it is transferred to the destination by utilizing available bandwidth. Biorthogonal 5.5 Wavelet Compression scheme compress the MBD without losing the important information, thus making the information reliable and less in size; transference by efficient bandwidth utilization from source to destination.
Object Recogniton Based on Undecimated Wavelet TransformIJCOAiir
Object Recognition (OR) is the mission of finding a specified object in an image or video sequence
in computer vision. An efficient method for recognizing object in an image based on Undecimated Wavelet
Transform (UWT) is proposed. In this system, the undecimated coefficients are used as features to recognize the
objects. The given original image is decomposed by using the UWT. All coefficients are taken as features for
the classification process. This method is applied to all the training images and the extracted features of
unknown object are used as an input to the K-Nearest Neighbor (K-NN) classifier to recognize the object. The
assessment of the system is agreed on using Columbia Object Image Library Dataset (COIL-100) database.
An Efficient Brain Tumor Detection Algorithm based on Segmentation for MRI Sy...ijtsrd
A collection, or mass, of abnormal cells in the brain is called as Brain Tumor . The skull, which encloses your brain, is very rigid. Growth inside such a restricted space can cause problems. Brain tumors can be malignant or benign. Segmentation in magnetic resonance imaging (MRI) was an emergent research area in the field of medical imaging system. In this an efficient algorithm is proposed for tumor detection based on segmentation and morphological operators. Quality of scanned image is enhanced and then morphological operators are applied to detect the tumor in the scanned image. Merlin Asha. M | G. Naveen Balaji | S. Mythili | A. Karthikeyan | N. Thillaiarasu"An Efficient Brain Tumor Detection Algorithm based on Segmentation for MRI System" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-2 | Issue-2 , February 2018, URL: http://www.ijtsrd.com/papers/ijtsrd9667.pdf http://www.ijtsrd.com/engineering/electronics-and-communication-engineering/9667/an-efficient-brain-tumor-detection-algorithm-based-on-segmentation-for-mri-system/merlin-asha-m
A BINARY BAT INSPIRED ALGORITHM FOR THE CLASSIFICATION OF BREAST CANCER DATAIJSCAI Journal
Advancement in information and technology has made a major impact on medical science where the
researchers come up with new ideas for improving the classification rate of various diseases. Breast cancer
is one such disease killing large number of people around the world. Diagnosing the disease at its earliest
instance makes a huge impact on its treatment. The authors propose a Binary Bat Algorithm (BBA) based
Feedforward Neural Network (FNN) hybrid model, where the advantages of BBA and efficiency of FNN is
exploited for the classification of three benchmark breast cancer datasets into malignant and benign cases.
Here BBA is used to generate a V-shaped hyperbolic tangent function for training the network and a fitness
function is used for error minimization. FNNBBA based classification produces 92.61% accuracy for
training data and 89.95% for testing data.
Implementing Tumor Detection and Area Calculation in Mri Image of Human Brain...IJERA Editor
This paper is based on the research on Human Brain Tumor which uses the MRI imaging technique to capture the image. In this proposed work Brain Tumor area is calculated to define the Stage or level of seriousness of the tumor. Image Processing techniques are used for the brain tumor area calculation and Neural Network algorithms for the tumor position calculation. Also in the further advancement the classification of the tumor based on few parameters is also expected. Proposed work is divided in to following Modules: Module 1: Image Pre-Processing Module 2: Feature Extraction, Segmentation using K-Means Algorithm and Fuzzy C-Means Algorithm Module 3: Tumor Area calculation & Stage detection Module 4: Classification and position calculation of tumor using Neural Network
Brain tumour segmentation based on local independent projection based classif...eSAT Journals
This document summarizes a research paper on brain tumour segmentation using local independent projection based classification. The proposed method uses MRI images and consists of four main steps: preprocessing using median filtering, feature extraction using patches, tumour segmentation using local independent projection classification, and post processing to analyze the tumour region. Local independent projection classification treats segmentation as a classification problem and uses local anchor embedding and softmax regression to improve performance. The method was able to classify tumour and edema regions and calculate the tumour area and perimeter pixels.
THE EFFECT OF PHYSICAL BASED FEATURES FOR RECOGNITION OF RECAPTURED IMAGESijcsit
It is very simple and easier to recapture a high quality images from LCD screens with the development of multimedia technology and digital devices. In authentication, the use of such recaptured images can be very dangerous. So, it is very important to recognize the recaptured images in order to increase authenticity. Even though, there are a number of features that have been proposed in various state-of-theart
visual recognition tasks, but it is still difficult to decide which feature or combination of features have more significant impact on this task. In this paper an image recapture detection method based on set of physical based features including texture, HSV colour and blurriness is proposed. Also, this paper evaluates the performance of different distinctive featuresin the context of recognition of recaptured
images. Several experimental setups have been conducted in order to demonstrate the performance of the proposed method. In all these experimental results, the proposed method is efficient with good recognition rate. Among the combination of low-level features, CS-LBP detection is to operator which is used to extract the texture feature is the most robust feature.
IRJET - Clustering Algorithm for Brain Image SegmentationIRJET Journal
The document presents a clustering algorithm for brain image segmentation using fuzzy c-means clustering. It aims to optimize the segmentation process and achieve higher accuracy rates when segmenting human MRI brain images. The fuzzy c-means algorithm is combined with rough set theory for segmentation. The algorithm segments images into homogeneous regions where adjacent regions are heterogeneous. This approach is evaluated on a set of brain images and demonstrates effectiveness as well as a comparison to other related algorithms. The goal of the algorithm is to simplify images and extract useful information for detecting brain tumors.
Image Fusion and Image Quality Assessment of Fused ImagesCSCJournals
Accurate diagnosis of tumor extent is important in radiotherapy. This paper presents the use of image fusion of PET and MRI image. Multi-sensor image fusion is the process of combining information from two or more images into a single image. The resulting image contains more information as compared to individual images. PET delivers high-resolution molecular imaging with a resolution down to 2.5 mm full width at half maximum (FWHM), which allows us to observe the brain\'s molecular changes using the specific reporter genes and probes. On the other hand, the 7.0 T-MRI, with sub-millimeter resolution images of the cortical areas down to 250 m, allows us to visualize the fine details of the brainstem areas as well as the many cortical and sub-cortical areas. The PET-MRI fusion imaging system provides complete information on neurological diseases as well as cognitive neurosciences. The paper presents PCA based image fusion and also focuses on image fusion algorithm based on wavelet transform to improve resolution of the images in which two images to be fused are firstly decomposed into sub-images with different frequency and then the information fusion is performed and finally these sub-images are reconstructed into result image with plentiful information. . We also propose image fusion in Radon space. This paper presents assessment of image fusion by measuring the quantity of enhanced information in fused images. We use entropy, mean, standard deviation and Fusion Mutual Information, cross correlation , Mutual Information Root Mean Square Error, Universal Image Quality Index and Relative shift in mean to compare fused image quality. Comparative evaluation of fused images is a critical step to evaluate the relative performance of different image fusion algorithms. In this paper, we also propose image quality metric based on the human vision system (HVS).
MIP AND UNSUPERVISED CLUSTERING FOR THE DETECTION OF BRAIN TUMOUR CELLSAM Publications
Image processing is widely used in biomedical applications. Image processing can be used to analyze
different MRI brain images in order to get the abnormality in the image .The objective is to extract meaningful
information from the imaged signals. Image segmentation is a process of partitioning an image in to different parts.
The division in to parts is often based on the characteristics of the pixels in the image. In our paper the segmentation
of the tumour tissues is carried out using k-means and fuzzy c-means clustering.Tumour can be found and faster
detection is achieved with only few seconds for execution. The input image of the brain is taken from the available
database and the presence of tumourin input image can be detected.
This document describes a study that used deep learning models to diagnose Alzheimer's disease using MRI scans. Specifically, it used convolutional neural networks (CNNs) trained on MRI data from the Alzheimer's Disease Neuroimaging Initiative (ADNI) dataset. The study achieved an F1-score of 89% for classifying Alzheimer's vs normal cognition. It also used a CycleGAN for data augmentation, which generated additional synthetic MRI images and improved the CNN classification accuracy to 95% F1-score. The document provides background on Alzheimer's diagnosis, CNNs, GANs including CycleGAN, related work applying machine learning to Alzheimer's diagnosis, and the methodology used in this study for data preprocessing, model training, and evaluating the impact of GAN
SELF-LEARNING AI FRAMEWORK FOR SKIN LESION IMAGE SEGMENTATION AND CLASSIFICATIONijcsit
Image segmentation and classification are the two main fundamental steps in pattern recognition. To perform medical image segmentation or classification with deep learning models, it requires training on large image dataset with annotation. The dermoscopy images (ISIC archive) considered for this work does not have ground truth information for lesion segmentation. Performing manual labelling on this dataset is time-consuming. To overcome this issue, self-learning annotation scheme was proposed in the two-stage deep learning algorithm. The two-stage deep learning algorithm consists of U-Net segmentation model with the annotation scheme and CNN classifier model. The annotation scheme uses a K-means clustering algorithm along with merging conditions to achieve initial labelling information for training the U-Net model. The classifier models namely ResNet-50 and LeNet-5 were trained and tested on the image dataset without segmentation for comparison and with the U-Net segmentation for implementing the proposed self-learning Artificial Intelligence (AI) framework. The classification results of the proposed AI framework achieved training accuracy of 93.8% and testing accuracy of 82.42% when compared with the two classifier models directly trained on the input images.
A NOVEL DATA DICTIONARY LEARNING FOR LEAF RECOGNITIONsipij
Automatic leaf recognition via image processing has been greatly important for a number of professionals, such as botanical taxonomic, environmental protectors, and foresters. Learn an over-complete leaf dictionary is an essential step for leaf image recognition. Big leaf images dimensions and training images number is facing of fast and complete data leaves dictionary. In this work an efficient approach applies to construct over-complete data leaves dictionary to set of big images diminutions based on sparse representation. In the proposed method a new cropped-contour method has used to crop the training image. The experiments are testing using correlation between the sparse representation and data dictionary and with focus on the computing time.
Plant Disease Detection using Convolution Neural Network (CNN)IRJET Journal
This document describes a study that used a convolutional neural network (CNN) to detect plant diseases from images with high accuracy. The researchers trained a CNN model on a dataset of plant leaf images labeled with 38 different disease classes. The CNN was able to automatically extract features from the input images and classify them into the respective disease classes. The proposed system achieved an average accuracy of 92%, demonstrating that neural networks can effectively detect plant diseases even with limited computing resources. The document provides details on how CNNs work, including their typical layers of convolution, max pooling, and fully connected layers, and discusses previous related work applying deep learning to plant disease detection.
Image segmentation and classification tasks in computer vision have proven to be highly effective using neural networks, specifically Convolutional Neural Networks (CNNs). These tasks have numerous
practical applications, such as in medical imaging, autonomous driving, and surveillance. CNNs are capable
of learning complex features directly from images and achieving outstanding performance across several
datasets. In this work, we have utilized three different datasets to investigate the efficacy of various preprocessing and classification techniques in accurssedately segmenting and classifying different structures
within the MRI and natural images. We have utilized both sample gradient and Canny Edge Detection
methods for pre-processing, and K-means clustering have been applied to segment the images. Image
augmentation improves the size and diversity of datasets for training the models for image classification
Image segmentation and classification tasks in computer vision have proven to be highly effective using neural networks, specifically Convolutional Neural Networks (CNNs). These tasks have numerous
practical applications, such as in medical imaging, autonomous driving, and surveillance. CNNs are capable
of learning complex features directly from images and achieving outstanding performance across several
datasets. In this work, we have utilized three different datasets to investigate the efficacy of various preprocessing and classification techniques in accurssedately segmenting and classifying different structures
within the MRI and natural images. We have utilized both sample gradient and Canny Edge Detection
methods for pre-processing, and K-means clustering have been applied to segment the images. Image
augmentation improves the size and diversity of datasets for training the models for image classification.
This work highlights transfer learning’s effectiveness in image classification using CNNs and VGG 16 that
provides insights into the selection of pre-trained models and hyper parameters for optimal performance.
We have proposed a comprehensive approach for image segmentation and classification, incorporating preprocessing techniques, the K-means algorithm for segmentation, and employing deep learning models such
as CNN and VGG 16 for classification.
SEGMENTATION OF LUNG GLANDULAR CELLS USING MULTIPLE COLOR SPACESIJCSEA Journal
Early detection of lung cancer is a challenging problem, the world faces today. Prior to classify glandular cells as malignant or benign a reliable segmentation technique is required. In this paper we present a novel lung glandular cell segmentation technique. The technique uses a combination of multiple color spaces and various clustering algorithms to automatically find the best possible segmentation result. Unsupervised clustering methods of K-means and Fuzzy C-means were used on multiple color spaces such as HSV, LAB, LUV, xyY. Experimental results of segmentation using various color spaces are provided to show the performance of the proposed system.
Utilization of Super Pixel Based Microarray Image Segmentationijtsrd
In the division of PC vision pictures, Super pixels are go probably as key part from 10 years prior. There are various counts and methodology to separate the Super pixels anyway whole all of them the best super pixel looking at strategy is Simple Linear Iterative Clustering SLIC have come to pivot continuously recently. The concentrating of small scale group quality verbalization from MRI imaging is more useful to perceive tumors or some other dangerous development contaminations, so the fundamental DNA cDNA microarray is a grounded device for analyzing the same. The division of microarray pictures is the essential development in a microarray assessment. In this paper, we proposed a figuring to dividing the cDNA small show picture using Simple Linear Iterative Clustering SLIC based Self Organizing Maps SOM method. In any case, the proposed figuring is taken up a moving task to look at the bad quality of pictures in addition. There are two phases to separate the image, introductory, a pre setting up the applied picture to diminish fuss levels and second, to piece the image using SLIC based SOM approach. Mr. Davu Manikanta | Mr. Parasurama N | K Keerthi "Utilization of Super Pixel Based Microarray Image Segmentation" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-5 , August 2021, URL: https://www.ijtsrd.com/papers/ijtsrd46274.pdf Paper URL: https://www.ijtsrd.com/engineering/electronics-and-communication-engineering/46274/utilization-of-super-pixel-based-microarray-image-segmentation/mr-davu-manikanta
Multimodal Medical Image Fusion Based On SVDIOSR Journals
Image fusion is a promising process in the field of medical image processing, the idea behind is to
improve the content of medical image by combining two or more multimodal medical images. In this paper a
novel fusion framework based on singular value decomposition - based image fusion algorithm is proposed.
SVD is an image adaptive transform, it transforms the matrix of the given image into product USVT
, which
allows to refactor a digital image into three matrices called tensors. The proposed algorithm picks out
informative image patches of source images to constitute the fused image by processing the divided subtensors
rather than the whole tensor and a novel sigmoid-function-like coefficient-combining scheme is applied to
construct the fused result. Experimental results show that the proposed algorithm is an alternative image fusion
approach.
Flower Classification Using Neural Network Based Image ProcessingIOSR Journals
Abstract: In this paper, it is proposed to have a method for classification of flowers using Artificial Neural Network (ANN) classifier. The proposed method is based on textural features such as Gray level co-occurrence matrix (GLCM) and discrete wavelet transform (DWT). A flower image is segmented using a threshold based method. The data set has different flower images with similar appearance .The database of flower images is a mixture of images taken from World Wide Web and the images taken by us. The ANN has been trained by 50 samples to classify 5 classes of flowers and achieved classification accuracy more than 85% using GLCM features only. Keywords: Artificial Neural Network, DWT, GLCM, Segmentation.
Classification of MR medical images Based Rough-Fuzzy KMeansIOSRJM
The document summarizes a proposed algorithm for classifying MR medical images using Rough-Fuzzy K-Means (FRKM). It begins with an introduction to the challenges of medical image classification and a literature review of previous techniques. It then provides background on rough set theory, fuzzy set theory, and K-means clustering. The proposed FRKM algorithm is described as using rough set theory for feature selection and dimensionality reduction, followed by a K-means clustering with probabilities assigned based on rough set approximations to classify ambiguous areas. Experimental results show the FRKM approach achieves 94.4% accuracy, higher than other techniques.
Quality Compression for Medical Big Data X-Ray Image using Biorthogonal 5.5 W...IJERA Editor
Medical Big Data (MBD) consists of very useful type of information. It is very important for a physician for decision making and treatments to cure the patient. For accurate diagnosis, data availability is the most important factor. MBD over network needs intelligent compression schemes so that it is transferred to the destination by utilizing available bandwidth. Biorthogonal 5.5 Wavelet Compression scheme compress the MBD without losing the important information, thus making the information reliable and less in size; transference by efficient bandwidth utilization from source to destination.
Object Recogniton Based on Undecimated Wavelet TransformIJCOAiir
Object Recognition (OR) is the mission of finding a specified object in an image or video sequence
in computer vision. An efficient method for recognizing object in an image based on Undecimated Wavelet
Transform (UWT) is proposed. In this system, the undecimated coefficients are used as features to recognize the
objects. The given original image is decomposed by using the UWT. All coefficients are taken as features for
the classification process. This method is applied to all the training images and the extracted features of
unknown object are used as an input to the K-Nearest Neighbor (K-NN) classifier to recognize the object. The
assessment of the system is agreed on using Columbia Object Image Library Dataset (COIL-100) database.
An Efficient Brain Tumor Detection Algorithm based on Segmentation for MRI Sy...ijtsrd
A collection, or mass, of abnormal cells in the brain is called as Brain Tumor . The skull, which encloses your brain, is very rigid. Growth inside such a restricted space can cause problems. Brain tumors can be malignant or benign. Segmentation in magnetic resonance imaging (MRI) was an emergent research area in the field of medical imaging system. In this an efficient algorithm is proposed for tumor detection based on segmentation and morphological operators. Quality of scanned image is enhanced and then morphological operators are applied to detect the tumor in the scanned image. Merlin Asha. M | G. Naveen Balaji | S. Mythili | A. Karthikeyan | N. Thillaiarasu"An Efficient Brain Tumor Detection Algorithm based on Segmentation for MRI System" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-2 | Issue-2 , February 2018, URL: http://www.ijtsrd.com/papers/ijtsrd9667.pdf http://www.ijtsrd.com/engineering/electronics-and-communication-engineering/9667/an-efficient-brain-tumor-detection-algorithm-based-on-segmentation-for-mri-system/merlin-asha-m
A BINARY BAT INSPIRED ALGORITHM FOR THE CLASSIFICATION OF BREAST CANCER DATAIJSCAI Journal
Advancement in information and technology has made a major impact on medical science where the
researchers come up with new ideas for improving the classification rate of various diseases. Breast cancer
is one such disease killing large number of people around the world. Diagnosing the disease at its earliest
instance makes a huge impact on its treatment. The authors propose a Binary Bat Algorithm (BBA) based
Feedforward Neural Network (FNN) hybrid model, where the advantages of BBA and efficiency of FNN is
exploited for the classification of three benchmark breast cancer datasets into malignant and benign cases.
Here BBA is used to generate a V-shaped hyperbolic tangent function for training the network and a fitness
function is used for error minimization. FNNBBA based classification produces 92.61% accuracy for
training data and 89.95% for testing data.
Implementing Tumor Detection and Area Calculation in Mri Image of Human Brain...IJERA Editor
This paper is based on the research on Human Brain Tumor which uses the MRI imaging technique to capture the image. In this proposed work Brain Tumor area is calculated to define the Stage or level of seriousness of the tumor. Image Processing techniques are used for the brain tumor area calculation and Neural Network algorithms for the tumor position calculation. Also in the further advancement the classification of the tumor based on few parameters is also expected. Proposed work is divided in to following Modules: Module 1: Image Pre-Processing Module 2: Feature Extraction, Segmentation using K-Means Algorithm and Fuzzy C-Means Algorithm Module 3: Tumor Area calculation & Stage detection Module 4: Classification and position calculation of tumor using Neural Network
Brain tumour segmentation based on local independent projection based classif...eSAT Journals
This document summarizes a research paper on brain tumour segmentation using local independent projection based classification. The proposed method uses MRI images and consists of four main steps: preprocessing using median filtering, feature extraction using patches, tumour segmentation using local independent projection classification, and post processing to analyze the tumour region. Local independent projection classification treats segmentation as a classification problem and uses local anchor embedding and softmax regression to improve performance. The method was able to classify tumour and edema regions and calculate the tumour area and perimeter pixels.
THE EFFECT OF PHYSICAL BASED FEATURES FOR RECOGNITION OF RECAPTURED IMAGESijcsit
It is very simple and easier to recapture a high quality images from LCD screens with the development of multimedia technology and digital devices. In authentication, the use of such recaptured images can be very dangerous. So, it is very important to recognize the recaptured images in order to increase authenticity. Even though, there are a number of features that have been proposed in various state-of-theart
visual recognition tasks, but it is still difficult to decide which feature or combination of features have more significant impact on this task. In this paper an image recapture detection method based on set of physical based features including texture, HSV colour and blurriness is proposed. Also, this paper evaluates the performance of different distinctive featuresin the context of recognition of recaptured
images. Several experimental setups have been conducted in order to demonstrate the performance of the proposed method. In all these experimental results, the proposed method is efficient with good recognition rate. Among the combination of low-level features, CS-LBP detection is to operator which is used to extract the texture feature is the most robust feature.
IRJET - Clustering Algorithm for Brain Image SegmentationIRJET Journal
The document presents a clustering algorithm for brain image segmentation using fuzzy c-means clustering. It aims to optimize the segmentation process and achieve higher accuracy rates when segmenting human MRI brain images. The fuzzy c-means algorithm is combined with rough set theory for segmentation. The algorithm segments images into homogeneous regions where adjacent regions are heterogeneous. This approach is evaluated on a set of brain images and demonstrates effectiveness as well as a comparison to other related algorithms. The goal of the algorithm is to simplify images and extract useful information for detecting brain tumors.
Image Fusion and Image Quality Assessment of Fused ImagesCSCJournals
Accurate diagnosis of tumor extent is important in radiotherapy. This paper presents the use of image fusion of PET and MRI image. Multi-sensor image fusion is the process of combining information from two or more images into a single image. The resulting image contains more information as compared to individual images. PET delivers high-resolution molecular imaging with a resolution down to 2.5 mm full width at half maximum (FWHM), which allows us to observe the brain\'s molecular changes using the specific reporter genes and probes. On the other hand, the 7.0 T-MRI, with sub-millimeter resolution images of the cortical areas down to 250 m, allows us to visualize the fine details of the brainstem areas as well as the many cortical and sub-cortical areas. The PET-MRI fusion imaging system provides complete information on neurological diseases as well as cognitive neurosciences. The paper presents PCA based image fusion and also focuses on image fusion algorithm based on wavelet transform to improve resolution of the images in which two images to be fused are firstly decomposed into sub-images with different frequency and then the information fusion is performed and finally these sub-images are reconstructed into result image with plentiful information. . We also propose image fusion in Radon space. This paper presents assessment of image fusion by measuring the quantity of enhanced information in fused images. We use entropy, mean, standard deviation and Fusion Mutual Information, cross correlation , Mutual Information Root Mean Square Error, Universal Image Quality Index and Relative shift in mean to compare fused image quality. Comparative evaluation of fused images is a critical step to evaluate the relative performance of different image fusion algorithms. In this paper, we also propose image quality metric based on the human vision system (HVS).
MIP AND UNSUPERVISED CLUSTERING FOR THE DETECTION OF BRAIN TUMOUR CELLSAM Publications
Image processing is widely used in biomedical applications. Image processing can be used to analyze
different MRI brain images in order to get the abnormality in the image .The objective is to extract meaningful
information from the imaged signals. Image segmentation is a process of partitioning an image in to different parts.
The division in to parts is often based on the characteristics of the pixels in the image. In our paper the segmentation
of the tumour tissues is carried out using k-means and fuzzy c-means clustering.Tumour can be found and faster
detection is achieved with only few seconds for execution. The input image of the brain is taken from the available
database and the presence of tumourin input image can be detected.
This document describes a study that used deep learning models to diagnose Alzheimer's disease using MRI scans. Specifically, it used convolutional neural networks (CNNs) trained on MRI data from the Alzheimer's Disease Neuroimaging Initiative (ADNI) dataset. The study achieved an F1-score of 89% for classifying Alzheimer's vs normal cognition. It also used a CycleGAN for data augmentation, which generated additional synthetic MRI images and improved the CNN classification accuracy to 95% F1-score. The document provides background on Alzheimer's diagnosis, CNNs, GANs including CycleGAN, related work applying machine learning to Alzheimer's diagnosis, and the methodology used in this study for data preprocessing, model training, and evaluating the impact of GAN
SELF-LEARNING AI FRAMEWORK FOR SKIN LESION IMAGE SEGMENTATION AND CLASSIFICATIONijcsit
Image segmentation and classification are the two main fundamental steps in pattern recognition. To perform medical image segmentation or classification with deep learning models, it requires training on large image dataset with annotation. The dermoscopy images (ISIC archive) considered for this work does not have ground truth information for lesion segmentation. Performing manual labelling on this dataset is time-consuming. To overcome this issue, self-learning annotation scheme was proposed in the two-stage deep learning algorithm. The two-stage deep learning algorithm consists of U-Net segmentation model with the annotation scheme and CNN classifier model. The annotation scheme uses a K-means clustering algorithm along with merging conditions to achieve initial labelling information for training the U-Net model. The classifier models namely ResNet-50 and LeNet-5 were trained and tested on the image dataset without segmentation for comparison and with the U-Net segmentation for implementing the proposed self-learning Artificial Intelligence (AI) framework. The classification results of the proposed AI framework achieved training accuracy of 93.8% and testing accuracy of 82.42% when compared with the two classifier models directly trained on the input images.
A NOVEL DATA DICTIONARY LEARNING FOR LEAF RECOGNITIONsipij
Automatic leaf recognition via image processing has been greatly important for a number of professionals, such as botanical taxonomic, environmental protectors, and foresters. Learn an over-complete leaf dictionary is an essential step for leaf image recognition. Big leaf images dimensions and training images number is facing of fast and complete data leaves dictionary. In this work an efficient approach applies to construct over-complete data leaves dictionary to set of big images diminutions based on sparse representation. In the proposed method a new cropped-contour method has used to crop the training image. The experiments are testing using correlation between the sparse representation and data dictionary and with focus on the computing time.
Plant Disease Detection using Convolution Neural Network (CNN)IRJET Journal
This document describes a study that used a convolutional neural network (CNN) to detect plant diseases from images with high accuracy. The researchers trained a CNN model on a dataset of plant leaf images labeled with 38 different disease classes. The CNN was able to automatically extract features from the input images and classify them into the respective disease classes. The proposed system achieved an average accuracy of 92%, demonstrating that neural networks can effectively detect plant diseases even with limited computing resources. The document provides details on how CNNs work, including their typical layers of convolution, max pooling, and fully connected layers, and discusses previous related work applying deep learning to plant disease detection.
Image segmentation and classification tasks in computer vision have proven to be highly effective using neural networks, specifically Convolutional Neural Networks (CNNs). These tasks have numerous
practical applications, such as in medical imaging, autonomous driving, and surveillance. CNNs are capable
of learning complex features directly from images and achieving outstanding performance across several
datasets. In this work, we have utilized three different datasets to investigate the efficacy of various preprocessing and classification techniques in accurssedately segmenting and classifying different structures
within the MRI and natural images. We have utilized both sample gradient and Canny Edge Detection
methods for pre-processing, and K-means clustering have been applied to segment the images. Image
augmentation improves the size and diversity of datasets for training the models for image classification
Image segmentation and classification tasks in computer vision have proven to be highly effective using neural networks, specifically Convolutional Neural Networks (CNNs). These tasks have numerous
practical applications, such as in medical imaging, autonomous driving, and surveillance. CNNs are capable
of learning complex features directly from images and achieving outstanding performance across several
datasets. In this work, we have utilized three different datasets to investigate the efficacy of various preprocessing and classification techniques in accurssedately segmenting and classifying different structures
within the MRI and natural images. We have utilized both sample gradient and Canny Edge Detection
methods for pre-processing, and K-means clustering have been applied to segment the images. Image
augmentation improves the size and diversity of datasets for training the models for image classification.
This work highlights transfer learning’s effectiveness in image classification using CNNs and VGG 16 that
provides insights into the selection of pre-trained models and hyper parameters for optimal performance.
We have proposed a comprehensive approach for image segmentation and classification, incorporating preprocessing techniques, the K-means algorithm for segmentation, and employing deep learning models such
as CNN and VGG 16 for classification.
Accuracy study of image classification for reverse vending machine waste segr...IJECEIAES
This study aims to create a sorting system with high accuracy that can classify various beverage containers based on types and separate them accordingly. This reverse vending machine (RVM) provides an image classification method and allows for recycling three types of beverage containers: drink carton boxes, polyethylene terephthalate (PET) bottles, and aluminium cans. The image classification method used in this project is transfer learning with convolutional neural networks (CNN). AlexNet, GoogLeNet, DenseNet201, InceptionResNetV2, InceptionV3, MobileNetV2, XceptionNet, ShuffleNet, ResNet 18, ResNet 50, and ResNet 101 are the neural networks that used in this project. This project will compare the F1- score and computational time among the eleven networks. The F1-score and computational time of image classification differs for each neural network. In this project, the AlexNet network gave the best F1-score, 97.50% with the shortest computational time, 2229.235 s among the eleven neural networks.
Plastic waste has become a pressing global concern in recent decades, posing significant challenges to our
environment due to its non-biodegradable nature and causing significant pollution and damage to our planet.
Recycling plastic waste is one of the most effective solutions to this dilemma, which is why the aim of our project
was to create a system that detects plastic waste using a large dataset with labeled data and one of the most
famous deep learning neural networks, “Convolutional Neural Networks,” to classify and speed up the waste
collection process and provide an easier recycling process. Thanks to our work, we have achieved 97% accuracy.
CNN FEATURES ARE ALSO GREAT AT UNSUPERVISED CLASSIFICATION cscpconf
This paper aims at providing insight on the transferability of deep CNN features to
unsupervised problems. We study the impact of different pretrained CNN feature extractors on
the problem of image set clustering for object classification as well as fine-grained
classification. We propose a rather straightforward pipeline combining deep-feature extraction
using a CNN pretrained on ImageNet and a classic clustering algorithm to classify sets of
images. This approach is compared to state-of-the-art algorithms in image-clustering and
provides better results. These results strengthen the belief that supervised training of deep CNN
on large datasets, with a large variability of classes, extracts better features than most carefully
designed engineering approaches, even for unsupervised tasks. We also validate our approach
on a robotic application, consisting in sorting and storing objects smartly based on clustering
A Survey of Convolutional Neural Network Architectures for Deep Learning via ...ijtsrd
Convolutional Neural Network CNN designs can successfully classify, predict and cluster in many artificial intelligence applications. In the health sector, intensive studies continue for disease classification. When the literature in this field is examined, it is seen that the studies are concentrated on the health sector. Thanks to these studies, doctors can make an accurate diagnosis by examining radiological images more consistently. In addition, doctors can save time to do other patient work by using CNN. In this study, related current manuscripts in the health sector were examined. The contributions of these publications to the literature were explained and evaluated. Complementary and contradictory arguments of the presented perspectives were revealed. It has been stated that the current status of the studies carried out and in which direction the future studies should evolve and that they can make an important contribution to the literature. Suggestions have been made for the guidance for future studies. Ahmet Özcan | Mahmut Ünver | Atilla Ergüzen "A Survey of Convolutional Neural Network Architectures for Deep Learning via Health Images" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-6 | Issue-2 , February 2022, URL: https://www.ijtsrd.com/papers/ijtsrd49156.pdf Paper URL: https://www.ijtsrd.com/computer-science/artificial-intelligence/49156/a-survey-of-convolutional-neural-network-architectures-for-deep-learning-via-health-images/ahmet-özcan
PADDY CROP DISEASE DETECTION USING SVM AND CNN ALGORITHMIRJET Journal
- The document discusses a study on detecting diseases in paddy/rice crops using deep learning algorithms like convolutional neural networks (CNN) and support vector machines (SVM).
- A dataset of rice leaf images was created and a CNN model using transfer learning with MobileNet was developed and trained on the dataset to classify rice diseases.
- The proposed method aims to automatically classify rice disease images to help farmers more accurately identify diseases, as manual identification can be difficult and inaccurate. This could help improve treatment and support farmers.
Dilated Inception U-Net for Nuclei Segmentation in Multi-Organ Histology ImagesIRJET Journal
The document summarizes a study that used a Dilated Inception U-Net model for nuclei segmentation in histology images. Key points:
1. A Dilated Inception U-Net model was used to segment nuclei in histology images, which employs dilated convolutions to efficiently generate feature maps over a large input area.
2. The model was tested on the MoNuSeg dataset containing H&E stained images. Preprocessing included color normalization, data augmentation, and extracting 256x256 patches.
3. The Dilated Inception U-Net modifies the classic U-Net by replacing convolutional blocks with dilated inception blocks containing 1x1 and 3x3 filters with different dilation rates, allowing it to
The document proposes a method for face recognition using deep learning and data augmentation. It cleans and pre-processes existing face datasets to remove noise and extracts faces. It then uses image processing techniques to add masks to the faces to create a new masked face dataset. An Inception Resnet-v1 model is trained on the new dataset. The method is applied to build a face recognition application for employee timekeeping that achieves high accuracy even when faces are masked.
A Review on Food Classification using Convolutional Neural NetworksIRJET Journal
This document provides a review of food classification using convolutional neural networks (CNNs). It discusses how CNNs can automatically extract complex features from food images to classify foods, unlike traditional machine learning approaches. The document reviews several CNN models that have been used for food classification, including AlexNet, VGG-16, ResNet-50, Inception V3, LeNet-5, SqueezeNet, and CaffeNet. It also summarizes various food classification systems that have employed these CNNs and compares their performance based on accuracy. Overall, the document outlines the methodology of food classification systems using CNNs and highlights several CNN architectures for this task.
IRJET-Multiclass Classification Method Based On Deep Learning For Leaf Identi...IRJET Journal
This document discusses a multiclass classification method using deep learning for leaf identification to help farmers. It proposes using a convolutional neural network (CNN) model for feature extraction and classification of leaf images. The CNN model is trained on labeled leaf image data and can then be used to classify new unlabeled leaf images. The method involves preprocessing leaf images, extracting features using the CNN model, and classifying the leaves into different plant categories. The researchers tested their method on 13 plant leaf categories and 4 disease categories, achieving 95.25% accuracy. They conclude CNNs are well-suited for leaf identification and classification tasks due to their ability to handle large image datasets.
Transfer learning with multiple pre-trained network for fundus classificationTELKOMNIKA JOURNAL
Transfer learning (TL) is a technique of reuse and modify a pre-trained network.
It reuses feature extraction layer at a pre-trained network. A target domain in TL
obtains the features knowledge from the source domain. TL modified
classification layer at a pre-trained network. The target domain can do new tasks
according to a purpose. In this article, the target domain is fundus image
classification includes normal and neovascularization. Data consist of
100 patches. The comparison of training and validation data was 70:30.
The selection of training and validation data is done randomly. Steps of TL i.e
load pre-trained networks, replace final layers, train the network, and assess
network accuracy. First, the pre-trained network is a layer configuration of
the convolutional neural network architecture. Pre-trained network used are
AlexNet, VGG16, VGG19, ResNet50, ResNet101, GoogLeNet, Inception-V3,
InceptionResNetV2, and squeezenet. Second, replace the final layer is to replace
the last three layers. They are fully connected layer, softmax, and output layer.
The layer is replaced with a fully connected layer that classifies according to
number of classes. Furthermore, it's followed by a softmax and output layer that
matches with the target domain. Third, we trained the network. Networks were
trained to produce optimal accuracy. In this section, we use gradient descent
algorithm optimization. Fourth, assess network accuracy. The experiment results
show a testing accuracy between 80% and 100%.
: It is an End to End deep learning project to classify
disease in plants .I have built a web application in this project that can take a
picture of the plant and tell the farmer if the plant has a disease or not.
Convolutional neural networks (CNN) trained using deep learning (DL) have advanced dramatically in recent years. Researchers from a variety of fields have been motivated by the success of CNNs in computer vision to develop better CNN models for use in other visually-rich settings. Successes in image classification and research have been achieved in a wide variety of domains throughout the past year. Among the many popularized image classification techniques, the detection of plant leaf diseases has received extensive research. As a result of the nature of the procedure, image quality is often degraded and distortions are introduced during the capturing of the image. In this study, we look into how various CNN models are affected by distortions. Corn-maze leaf photos from the 4,188-image corn or maize leaf Dataset (split into four categories) are under consideration. To evaluate how well they handle noise and blur, researchers have deployed pre-trained deep CNN models like visual geometry group (VGG), InceptionV3, ResNet50, and EfficientNetB0. Classification accuracy and metrics like as recall and f1-score are used to evaluate CNN performance.
The document describes a study that used a convolutional neural network with a ConvNeXtLarge architecture to classify skin cancer images into benign and malignant classes. The CNN model was trained on a dataset of 3,297 skin cancer images from Kaggle. It achieved an AUC of 0.91 for classifying the images, demonstrating the ConvNeXtLarge architecture is effective for this task. The study aims to help early diagnosis and treatment of skin cancers.
This document summarizes a research paper that proposes using a convolutional neural network (CNN) model to detect rice leaf blast disease from images. The CNN model would be trained on a dataset of images of rice leaves with different levels of the disease. Farmers could then input images of infected rice leaves into a web application, and the CNN model would analyze the images to determine the severity of the disease and provide treatment recommendations. The goal is to automate disease detection and provide farmers with timely recommendations to help reduce rice crop losses from leaf blast disease.
Evaluation of deep neural network architectures in the identification of bone...TELKOMNIKA JOURNAL
This document evaluates the performance of three deep neural network architectures - ResNet, DenseNet, and NASNet - in identifying bone fissures in radiological images. The networks were trained on a dataset of 1000 labeled images of fissured and seamless bones. NASNet achieved the best performance with 75% accuracy, outperforming ResNet and DenseNet. While all networks reduced classification errors, NASNet did so with the fewest parameters. The document concludes NASNet is the best solution for this bone fissure identification task.
Development of 3D convolutional neural network to recognize human activities ...journalBEEI
This document describes the development of a 3D convolutional neural network (CNN) model to recognize human activities using moderate computation capabilities. The model is trained on the KTH dataset, which contains activities like walking, running, jogging, handwaving, handclapping, and boxing. The proposed model uses 3D CNN layers and max pooling layers to extract both spatial and temporal features from video frames. Testing achieved an accuracy of 93.33% for activity recognition. The number of model parameters and operations are also calculated to show the model can perform human activity recognition with reasonable computational requirements suitable for devices with moderate capabilities.
Detection of Diabetic Retinopathy using Convolutional Neural NetworkIRJET Journal
This document presents a system for detecting diabetic retinopathy using convolutional neural networks. The system inputs retinal images, extracts features like microaneurysms and hemorrhages, and classifies the images as normal or diseased and predicts the severity level. The proposed method uses a ResNet-18 convolutional neural network and achieves an accuracy of 83% on a dataset of over 8,500 images. The system is intended to allow for quicker and easier detection of diabetic retinopathy compared to manual examination methods.
Similar to AUTOMATIC FRUIT RECOGNITION BASED ON DCNN FOR COMMERCIAL SOURCE TRACE SYSTEM (20)
Applications of artificial Intelligence in Mechanical Engineering.pdfAtif Razi
Historically, mechanical engineering has relied heavily on human expertise and empirical methods to solve complex problems. With the introduction of computer-aided design (CAD) and finite element analysis (FEA), the field took its first steps towards digitization. These tools allowed engineers to simulate and analyze mechanical systems with greater accuracy and efficiency. However, the sheer volume of data generated by modern engineering systems and the increasing complexity of these systems have necessitated more advanced analytical tools, paving the way for AI.
AI offers the capability to process vast amounts of data, identify patterns, and make predictions with a level of speed and accuracy unattainable by traditional methods. This has profound implications for mechanical engineering, enabling more efficient design processes, predictive maintenance strategies, and optimized manufacturing operations. AI-driven tools can learn from historical data, adapt to new information, and continuously improve their performance, making them invaluable in tackling the multifaceted challenges of modern mechanical engineering.
Discover the latest insights on Data Driven Maintenance with our comprehensive webinar presentation. Learn about traditional maintenance challenges, the right approach to utilizing data, and the benefits of adopting a Data Driven Maintenance strategy. Explore real-world examples, industry best practices, and innovative solutions like FMECA and the D3M model. This presentation, led by expert Jules Oudmans, is essential for asset owners looking to optimize their maintenance processes and leverage digital technologies for improved efficiency and performance. Download now to stay ahead in the evolving maintenance landscape.
DEEP LEARNING FOR SMART GRID INTRUSION DETECTION: A HYBRID CNN-LSTM-BASED MODELijaia
As digital technology becomes more deeply embedded in power systems, protecting the communication
networks of Smart Grids (SG) has emerged as a critical concern. Distributed Network Protocol 3 (DNP3)
represents a multi-tiered application layer protocol extensively utilized in Supervisory Control and Data
Acquisition (SCADA)-based smart grids to facilitate real-time data gathering and control functionalities.
Robust Intrusion Detection Systems (IDS) are necessary for early threat detection and mitigation because
of the interconnection of these networks, which makes them vulnerable to a variety of cyberattacks. To
solve this issue, this paper develops a hybrid Deep Learning (DL) model specifically designed for intrusion
detection in smart grids. The proposed approach is a combination of the Convolutional Neural Network
(CNN) and the Long-Short-Term Memory algorithms (LSTM). We employed a recent intrusion detection
dataset (DNP3), which focuses on unauthorized commands and Denial of Service (DoS) cyberattacks, to
train and test our model. The results of our experiments show that our CNN-LSTM method is much better
at finding smart grid intrusions than other deep learning algorithms used for classification. In addition,
our proposed approach improves accuracy, precision, recall, and F1 score, achieving a high detection
accuracy rate of 99.50%.
Prediction of Electrical Energy Efficiency Using Information on Consumer's Ac...PriyankaKilaniya
Energy efficiency has been important since the latter part of the last century. The main object of this survey is to determine the energy efficiency knowledge among consumers. Two separate districts in Bangladesh are selected to conduct the survey on households and showrooms about the energy and seller also. The survey uses the data to find some regression equations from which it is easy to predict energy efficiency knowledge. The data is analyzed and calculated based on five important criteria. The initial target was to find some factors that help predict a person's energy efficiency knowledge. From the survey, it is found that the energy efficiency awareness among the people of our country is very low. Relationships between household energy use behaviors are estimated using a unique dataset of about 40 households and 20 showrooms in Bangladesh's Chapainawabganj and Bagerhat districts. Knowledge of energy consumption and energy efficiency technology options is found to be associated with household use of energy conservation practices. Household characteristics also influence household energy use behavior. Younger household cohorts are more likely to adopt energy-efficient technologies and energy conservation practices and place primary importance on energy saving for environmental reasons. Education also influences attitudes toward energy conservation in Bangladesh. Low-education households indicate they primarily save electricity for the environment while high-education households indicate they are motivated by environmental concerns.
Generative AI Use cases applications solutions and implementation.pdfmahaffeycheryld
Generative AI solutions encompass a range of capabilities from content creation to complex problem-solving across industries. Implementing generative AI involves identifying specific business needs, developing tailored AI models using techniques like GANs and VAEs, and integrating these models into existing workflows. Data quality and continuous model refinement are crucial for effective implementation. Businesses must also consider ethical implications and ensure transparency in AI decision-making. Generative AI's implementation aims to enhance efficiency, creativity, and innovation by leveraging autonomous generation and sophisticated learning algorithms to meet diverse business challenges.
https://www.leewayhertz.com/generative-ai-use-cases-and-applications/
Software Engineering and Project Management - Software Testing + Agile Method...Prakhyath Rai
Software Testing: A Strategic Approach to Software Testing, Strategic Issues, Test Strategies for Conventional Software, Test Strategies for Object -Oriented Software, Validation Testing, System Testing, The Art of Debugging.
Agile Methodology: Before Agile – Waterfall, Agile Development.
AUTOMATIC FRUIT RECOGNITION BASED ON DCNN FOR COMMERCIAL SOURCE TRACE SYSTEM
1. International Journal on Computational Science & Applications (IJCSA) Vol.8, No.2/3, June 2018
DOI:10.5121/ijcsa.2018.8301 1
AUTOMATIC FRUIT RECOGNITION BASED ON
DCNN FOR COMMERCIAL SOURCE TRACE SYSTEM
Israr Hussain, Qianhua He*
, Zhuliang Chen
School of Information and Electronic Engineering
South China University of Technology Guangzhou P. R. China,510641
ABSTRACT
Automatically fruit recognition by using machine vision is considered as challenging task due to similarities
between various types of fruits and external environmental changes e-g lighting. In this paper, fruit
recognition algorithm based on Deep Convolution Neural Network(DCNN) is proposed. Most of the
previous techniques have some limitations because they were examined and evaluated under limited
dataset, furthermore they have not considered external environmental changes. Another major contribution
in this paper is that we established fruit images database having 15 different categories comprising of
44406 images which were collected within a period of 6 months by keeping in view the limitations of
existing dataset under different real-world conditions. Images were directly used as input to DCNN for
training and recognition without extracting features, besides this DCNN learn optimal features from
images through adaptation process. The final decision was totally based on a fusion of all regional
classification using probability mechanism. Experimental results exhibit that the proposed approach have
efficient capability of automatically recognizing the fruit with a high accuracy of 99% and it can also
effectively meet real world application requirements.
KEYWORDS:
Fruit Recognition/Probability/ DCNN/ Machine vision
1. INTRODUCTION
Recognizing various kinds of fruits and vegetables is one of the most difficult task in supermarket
and fruit shops, since the cashier must point out the categories of a particular fruit to determine its
price. The use of barcodes has mostly overcome this problem for packed products but most of the
customers want to pick their commodities by themselves instead of the packed one and thus must
be weighed. This problem can be solved by issuing codes for every fruit, but here memorization
is another problem which lead to error in pricing. Another alternative solution is to issue the
cashier an inventory with pictures and codes, however flipping over the booklet is time
consuming [1]. Some alternative solutions were proposed to overcome this problem. Veggie
vision was the first supermarket produce recognition system consisting of an integrated scale and
image system with a user-friendly interface [2]. Hong et al. [3] employed morphological
examination to separate walnuts and hazelnuts into three groups. Baltazar at el. [4] first applied
data fusion to nondestructive image of fresh intact tomatoes, followed by three-class Bayesian
classifier. Pennington et al. [5] use a clustering algorithm for the classification of fruits and
vegetable. Pholpho et al. [6] used visible spectroscopy for classification of non-bruised and
bruised longan fruits, additionally it is combined with principal component analysis(PCA), Partial
Least Square Discriminant Analysis(PLS-DA) and Soft Independent Modeling of Class Analogy
(SIMCA) to developed classification models.
2. International Journal on Computational Science & Applications (IJCSA) Vol.8, No.2/3, June 2018
2
In recent years, Convolutional Neural Networks(CNN) have enjoyed great popularity as a means
for the image classification and categorization since Krizhevsky et al. [7] won the ImageNet
Large-Scale Visual Recognition Challenge(ILSVRC) 2012 competition. CNN as variant of the
standard deep neural network (DNN) is characterized by a special neural network architecture
consisting of alternating convolutional and pooling layers [8], it is also used to extract and
combine local features from a two-dimensional input. Compared to convolutional hand-crafted
feature extraction-based approaches, CNN is very effective since it is able to learn optimal
features from images through adaptive process. During CNN implementation phase for image
classification, researchers have to collect large-scale dataset as the ILSVERC contained more
than one million images [9 ,10] for a network which however is not a trial task. One simple way
to deal with this situation is to apply the CNN model that has been Pre-trained on a large-scale
image data [11,12] which is called transfer performing affine transformations to the raw images.
Convolutional Neural Networks (CNNs) have been established as a powerful class of models for
image recognition problem. Another Fruit recognition system based on convolutional neural
network approach is proposed by lei Hou in 2016[13], whereas in similar year another Fruit
Detection System using deep neural network is presented in [14].
Recent advancements in deep Convolutional Neural Networks (DCNNs) have led to impressive
progress in computer vision, especially in image classification. One of the most frequently used
deep learning methods in image processing is the convolutional neural networks. In previous
work, researchers mostly used manual method to determine the important features of each image
[16] (e.g. color, size, coordinate, textures) according to their requirements. We focused on
representation learning that automatically learned high level features only from users strokes in a
single image instead of manually tuning existing features.
In this paper, we proposed an efficient fruit classification system using deep convolutional neural
network model to extract image features and implement classification. Another contribution in
this paper is that we established fruit images database comprised of 15 categories. This paper was
aimed to apply Deep Convolutional Neural Network aided with the data expansion techniques to
a 15 class fruit images. Images can be directly used as input to the Deep Neural network without
any feature extraction. To improve the recognition performance, we proposed Deep
Convolutional neural network(DCNN) model with only five-layers, it is able to learn optimal
features from an large scale two dimensional input image. A series of experiments were carried
out using the proposed model on our own dataset of 44406 fruit images, results showed that our
classification system achieves an excellent accuracy rate of 99%.
The rest of the paper is organized as follow: Section 2 present proposed method, Section 3
Described the dataset, Section 4 comprised of experimental results, analysis, simulation and
confusion matrix whereas Section 5 based on performance comparison and last section represent
Conclusion.
2. PROPOSED METHOD & EXPLANATION
The methodology that involves in this scheme involves the following steps.
1. Select the fruit Images
2. Resize all fruit images from (320 ×258×3) to (150×150×3)
3. Converted all RGB images into Gray Scale
4. Transform our dataset from having shape (n, width, height) to (n, depth, width,
height)
3. International Journal on Computational Science & Applications (IJCSA) Vol.8, No.2/3, June 2018
3
5. Split dataset into training (30082), test (6804) & validation set (7520) using keras
train_test_split command
6. Convert our data type to float32 and normalize our data values from 0-255 to the
range [0, 1]
7. Preprocess class labels
8. Define our model architecture
9. Compile model with SGD optimizer and categorical-cross entropy, and learning
rate=0.0001
10. Fit and training data for 60 iterations and 20 iterations
11. Evaluate model on our test Dataset and calculate the confusion matrix
12. Classification has been done through Python 3.6
2.1 FLOW CHART OF FRUIT RECOGNITION SYSTEM
Flow Chart of fruit recognition system
2.2 PROPOSED DCNN ARCHITECTURE
In this paper we proposed a two-track deep neural network model architecture. The first track
consists of deep convolution neural network with max-pooling to enhance system ability, whereas
the second track comprised of fully connected layers. The network has four layers of hidden
neurons (three convolutional-pooling and one fully connected), apart from a final dense layer of
output neurons (the input is not considered as a layer). The input contains 150×150×3 neurons,
representing the RGB value for a 150×150×3 image. The first convolution-pooling layer uses a
local receptive field (convolutional kernel) of size 3×3 with a stride length of 1 pixel to extract 32
feature maps, followed by a max. pooling operation conducted in a 2×2 region, the second and
third convolution-pooling layers use the same 3×3 local receptive field (kernel) resulting 64 and
128 features maps respectively whereas other parameters remain unchanged. The fourth layer is
4. International Journal on Computational Science & Applications (IJCSA) Vol.8, No.2/3, June 2018
4
fully-connected layer with 64 ReLU (rectifier linear unit) neurons and the output layers have 15
SoftMax neurons that corresponding to the 15-various categories of fruits. The three
convolutional-pooling layers also use ReLU activation functions.
Four main operations are performed in the Conv.Net as shown in Figure.1
1. Convolution
2. Non-Linearity(ReLU)
3. Pooling & Sub Sampling
4. Classification (Fully Connected Layer)
Fig.1 Proposed Deep Neural Network Architecture
Convolution in case of Conv.Net is used to extract features from the input image. Convolution
preserves the spatial relation between pixels by learning image features using small squares of
input image. In our model we used 3×3 filter to compute convolved features. ReLU stand for
rectified linear unit is a non-linear operation. The purpose of ReLU is to introduce to non-linearity
in our Conv.Net because mostly Conv.Net has to learn non-linear real-world data. In our Network,
we used ReLU activation function because it trained the neural network several times faster
without a significant plenty to generalization accuracy. The biggest advantage of ReLU is indeed
non-saturation of its gradient, which greatly accelerates the convergence of stochastic gradient
descent compared to the sigmoid / tanh functions. Another nice property is that compared to tanh /
sigmoid neurons that involve expensive operations (exponentials, etc.), the ReLU can be
implemented by simply thresholding a matrix of activations at zero. Fully connected layer is a
traditional multi-layer perceptron that uses a SoftMax activation function in the output layer
(other classifiers like SVM can be used). The convolutional and max-pooling layers extract high
level features from image and used fully connected layer to classify the input images.
The network was trained with the SGD (stochastic gradient descent) algorithm with a cross-
entropy cost function. As we know Adaptive optimizers are great for getting viable results quickly,
but SGD + momentum is usually better for getting truly state of the art results. It just takes more
effort to tune the hyperparameters properly. The dropout used to overcome the overfitting
problems with dropout rates of 0.5 and 0.35 were set for the third convolutional-pooling layer and
fully- connected layer respectively. The pooling layer will simply perform down sampling along
with spatial dimensionality of the given input, further reducing the number of parameters within
that activation. The fully-connected layers then performed the same function as found in standard
ANNs and class scores are produced from the activations for classification purpose. Additionally,
ReLU is used between these layers to enhance performance.
5. International Journal on Computational Science & Applications (IJCSA) Vol.8, No.2/3, June 2018
5
3. DATASET
Datasets plays very important role in research as it is much expensive and harder to generate
because companies probably don’t feel freely giving away their investment, furthermore
universities don’t have many resources. Therefore, our major goal was to build a dataset that
contained many fruit images as feasible where each fruit is represented with many images as
possible. The images were captured at the resolution of 320×258×3 pixels by using HD Logitech
web camera. From experimental perspective, they were down sampled to 150×150×3 resolution
in this paper. The dataset comprising of 15 different categories and the number of images for each
class are listed in Table 1.
All images were stored in RGB color-space at 8 bits per channel. Images were collected at
various times in different days for the same category. These features increase the dataset
variability and represent more realistic scenario. The Images had large variations in quality and
lighting. Illumination is one of those variations in imagery. In fact, illumination can make two
images of same fruit less similar than two images of different kind of fruits. We used our own
intelligent weight machine and camera to captured all images. The fruit dataset was collected
under relatively unconstrained conditions. Images were also captured with different conditions i-e
room light on and off, moved the camera and intelligent weight machine near to the windows of
our lab, open windows, closed windows, open curtains, and closed curtains scenario. For a real
application in a supermarket, it might be necessary to cope with illumination variation, camera
capturing artifacts, specular reflection shading and shadows. We created all kind of real world
challenges for our dataset within our lab environment. Basic few conditions which we considered
during dataset collections are:
Pose Variations with different kind of fruits
Variable number of fruits/Vegetable
Used HD logistic webcam to captured to images
Same color but different Category fruits images
Different color same category fruit images
Different lighting conditions (i-e in fluorescent, natural light e-g within or without
sunshine as some of the fruits shops and supermarkets are without sunshine so it can
easily affect the recognition system.
To introduce a large amount of variations, fruit images have been captured by different fruit
orientation, variability on the number of fruits, illumination, and fruit size. Different external
environment variations contained in the dataset are represented in below table.1
3.1 EXISTING CATEGORIZATION CHALLENGES
Some of the existing fruit categorization challenges in our dataset are exhibit in below figures.
Different fruit images may have same color, shape and sometimes it is very tough for human eye
to clearly distinguish it in different lighting conditions. Fruit images were selected in such a way
that they cover a wide variety of fruit categories having same color. This could help to train a
strong classifier that can detect fruit images with a wide variety.
6. International Journal on Computational Science & Applications (IJCSA) Vol.8, No.2/3, June 2018
6
Figure2 Tomatoes (A) Persimmon (B) looks similar Figure.3 Apple (A) and Pear (B) look similar
Figure 5 demonstrate an example of Kiwi category with different illumination parameters. Our
dataset contained the samples of three different kind of kiwi fruit. The illuminations were
changed due to day light exposure, it has been generated through our artificial weight machine
near to the windows of our lab by opening and closing the windows, it’s not artificially tampered.
.
Figure.4 Variability in the number of elements’. kiwi category.
TABLE 1 SAMPLES OF 15 DIFFERENT CATEGORIES OF FRUITS
7. International Journal on Computational Science & Applications (IJCSA) Vol.8, No.2/3, June 2018
7
4. EXPERIMENTS RESULT & ANALYSIS
In this research analysis, we used Python 3.6, OpenCV to implement our solution. We have
developed this in keras with tensor flow back-end in spyder environment which was configured to
8. International Journal on Computational Science & Applications (IJCSA) Vol.8, No.2/3, June 2018
8
use GPU Quadro K4200 for training. We also used pyCUDA to implement some kernels. First,
we imported all keras libraries and open sources such as boken. Fruit identification in this work
has been accomplished through the training of the neural network by assigning proper feature
matrix and define the target class based on the input class of dataset. The database contained
44406 fruit images in which 30082 images used for training,6804 for testing &7520 for validation.
All the images down-sampled to a fixed resolution of 150×150×3(length=150, width=150, &
channel =3) from 320×258×3(length=320, width=258 &channel=3) and then randomly divided
into training, testing and validation according to above mentioned ratio. Label representation for
their respective categories is also made. Images and label are fed in to the CNN as input. Neural
network used for classification is shown in figure.1. The network was trained with the stochastic
gradient descent(SGD) algorithm with learning rate of 0.0001, momentum 0.9. Apparently if we
carefully tune the hyperparameters for SGD and Momentum we can obtain less test error
compared to Adam or RMS Prop. In my case I’ve found SGD optimizer work well by tuned
hyperparameters properly. It was trained for 20 & 60 epochs through the training set of 30082.In
the end, desired kind of fruit is determined through highest probability mechanism. In our
experiments we used three Convolutional & three Maxpooling2D layer with pool- size= (2,2). At
initial stage, the training loss was large which resulting in large learning rate to speed up the
training process but gradually the learning rate decreased with the loss, which help in avoiding
overfitting the best result. We used two drop layers with drop rate 0.5 and 0.3 to avoid overfitting
of model.. After several convolutional and max pooling layers, we used fully connected layer. Fig.
5 & 6 exhibit further investigation for visual proofs by visualizing high-dimensional features
space. This will give us an in depth look into their internal workings and help us in understanding
them better
Fig.5 Features activation map from Conv8 layer
Fig.6 Feature activation map from Con9 laver
9. International Journal on Computational Science & Applications (IJCSA) Vol.8, No.2/3, June 2018
9
4.1 SIMULATION RESULTS
Choosing a convolutional neural network architecture for real time object detection and
recognition is not an easy task because it is very difficult to estimate exact number of layers, kind
of layers and number of neurons to use in each layer. In this paper, we tested different kind of
CNN architecture to find the best suitable option. The CNN was first implemented without using
data expansion techniques. As shown in Fig.7&8, the accuracy and loss curves are set to 20 with
the training epoch during training. A large gap between training and validation accuracy occurred
after 7epochs, it indicating the presence of overfitting. The highest accuracy on test images was
found to be 82%, corresponding to a training accuracy 90%.
Fig.7 Model validation and training accuracy curve
Fig.8 Model validation and training loss curve
From the training and validation curve in fig.7 it seemed that the CNN model accuracy could be
further improved by increasing the training epochs. But when we increased the number of epochs
to 60, there is no increment had been observed in validation accuracy but the training accuracy
were increased to 95-96% as shown in figure 9. After repeated couple of experiments by applying
different overfitting techniques like added dropout layers, used data augmentation, by testing
different architectures that generalize well, varying hyper parameters, after applied these
techniques there is two significant improvement were achieved. First increased the accuracy till to
99% secondly the overfitting problem removed completely. The data augmentation is one way to
fight overfitting but it’s not enough, so we also considered other techniques such as
regularization, early stopping and all other above-mentioned techniques. Finally, we had
completely removed overfitting problem as showed in fig 12.
10. International Journal on Computational Science & Applications (IJCSA) Vol.8, No.2/3, June 2018
10
Fig.9 Validation and training accuracy curve
Fig.10 Validation & training loss curve
As shown in below figure11, the first test accuracy was greatly elevated to a level of 99% with 20
epochs and secondly the overfitting issue have been completely eliminated.
Fig.11 Model Accuracy Curves
Fig.12 Model Loss Curves
11. International Journal on Computational Science & Applications (IJCSA) Vol.8, No.2/3, June 2018
11
Table.2 demonstrate the classification report which presents the recognition probabilities by CNN
model for test images and corresponding recognition rates for each fruit category. The precision is
the ratio of tp /(tp+fp) where tp is the number of true positive and fp number of false positive. The
recall is the ratio of tp /(tp+fn) where tp number of positive and fn number of false. The recall is
intuitively the ability of the classifier to find all the positive samples. Support is the number of
occurrence of each class in y_train. F1-score is average of precision and recall. The support
represents the number of images for individual class used to test the model. We have also tested
our model for new real-world dataset and overall its performance is excellent.
Table 2 Classification Report Table
4.2 CONFUSION MATRIX
Confusion matrix allowed us to look at the particular misclassified examples yourself and
perform any further calculation as desired. In our confusion matrix X-axis is the predicted fruit
labels and Y-axis is the true labels. The diagonal represents the correct results. We can see most of
the fruit estimated correctly.
Figure 13 Confusion Matrix ……………………………………..
12. International Journal on Computational Science & Applications (IJCSA) Vol.8, No.2/3, June 2018
12
5. COMPARISON WITH OTHER TECHNIQUES
We compared our results with several recently proposed methods [13,15,17], based on deep
neural network. A fruit recognition algorithm based on convolution neural network(CNN)
combined with selective search algorithm is proposed by lei Hou 2016 [13]. The recognition rate
has been improved greatly up to 99.77%. But the database only contained 5330 images for 7
categories .4000 images used for training,1330 for testing. Although the recognition rate in this
method is quite good but less species has been used for fruit database in that experiment.
Furthermore, dataset is far from the real-world challenges and also not considering the external
environment change and other factors i-e light. This is the main reason it’s not more suitable for
real world applications.
Control system Development for fruit classification based on Convolutional neural network has
been suggested by Zaw Min Khaing in [15]. The model was evaluated under limited dataset
FIDS30 as fruit image dataset consist of 971 images for 30 different kind of fruits. The proposed
model had accuracy in the classification close to 94%. The model was developed using a limited
dataset. So, it can’t meet the actual application situation.
Fruit and vegetable classification system using image saliency and convolution neural network
has been proposed in [17]. In this method the model was evaluated under limited dataset consist
of 12173 images for 26 categories and have got an accuracy of 95.6%. Out of 12173 images, 80%
were used for training and 20% for validation. The number of training samples are small which is
not conducive to the extensibility of an algorithm.
Our proposed approach achieved almost similar accuracy with the above mention classification
system but we evaluated our model in much more complicated dataset which mostly meet all real-
world challenges. In our dataset many of fruits also contained their sub categories such as six
different kind of apple fruits, three different type of mango and kiwi fruits, three different kind
bananas fruits. Sub categories create miss matches between train and test set. Above mentioned
recognition systems validate on a database which did not contain sub categories of fruit. So, it
can’t meet the actual application situation. We also evaluated the performance of the model under
44406 image datasets which comprised of 15 different fruit categories. Deep learning algorithms
need a large amount of training samples to understand it perfectly. Hope our proposed model will
adapt the real-world application requirements. In our case we have enough dataset to trained a
CNN model.
6. CONCLUSION
In this paper, we proposed a novel fruit recognition classification system based on Deep
Convolution Neural Network(DCNN). On the other hand, we established our own experiments
database. The proposed method easily recognized fruits images with many challenges. In order to
improve the functionality and flexibility, we created all real-world challenges in our dataset.
Hence, the proposed method efficiently increased the recognition rate and it can meet the real
word application requirements. Experiments were conducted on our own dataset and the results
showed that our classification system has achieved an excellent accuracy rate of 99%. As a future
direction, we can also increase the number of classes of fruits and focus on Sub-class
categorization, fruit detection and localization in future work. We hope our method and database
will convey more than this paper and inspire the future studies. From future perspective, more
class and sub-class categorization should be included. This involve further experimenting with the
structure of the network. We can also apply our algorithm in other similar fields like shrimp color
monitoring.
13. International Journal on Computational Science & Applications (IJCSA) Vol.8, No.2/3, June 2018
13
ACKNOWLEDGMENT
This work was supported by National Natural Science Foundation of China (NSFC) (61571192).
The authors are expressing their sincere thanks to the South China University of Technology and
Chinese Scholarship Council(CSC) for their constant financial support and encouragement.
CONFLICT OF INTEREST
The authors declare no conflict of interest
ABBREVIATION
The following abbreviations were used in this paper.
DCNN Deep Convolutional Neural Network
SGD Stochastic Gradient Decent
ILSVRC Large-Scale Visual Recognition Challenge
ReLU Rectified Linear Unit
RGB Red Gray Blue
SVM Support Vector Machine
BOF Bag of Features
FC Layer Fully Connected Layer
GPU Graphic Processing Unit
REFERENCE
[1] Rocha, A.; Hauagge, D.C.; Wainer, J.; Goldenstein, S. Automatic fruit and vegetable classification
from images. Comput. Electron. Agric. 2010, 70, 96–104.
[2] Bolle, R.M.; Connell, J.H.; Haas, N.; Mohan, R.; Taubin, G. VeggieVision: A Produce Recognition
System. In Proceedings 3rd IEEE Workshop on Applications of Computer Vision, WACV'96,
Sarasota, FL, USA, 2–4 December 1996, pp. 244–251.
[3] Hong, S.G.; Maccaroni, M.; Figuli, P.J.; Pryor, B.M.; Belisario, A. Polyphasic Classification of
alternaria isolated from hazelnut and walnut fruit in Europe. Mycol. Res. 2006, 110, 1290–1300.
[4] Baltazar, A.; Aranda, J.I.; González-Aguilar, G. Bayesian classification of ripening stages of tomato
fruit using acoustic impact and colorimeter sensor data. Comput. Electron. Agric. 2008, 60, 113–121.
[5] Pennington, J.A.T.; Fisher, R.A. Classification of fruits and vegetables. J. Food Compos. Anal. 2009,
22, S23–S31.
[6] Pholpho, T.; Pathaveerat, S.; Sirisomboon, P. Classification of longan fruit bruising using visible
spectroscopy. J. Food Eng. 2011, 104, 169–172.
[7] Krizhevsky, I. Sutskever & G.E. Hinton, Imagenet classification with deep convolutional neural
networks, in Advances in Neural Information Processing Systems, 2012.
[8] Y.Lecun, L. Bottou, Y. Bengio & P. Haffner, Gradient-based learning applied to document
recognition, in Proceedings of the IEEE, 1998, 86(11): 2278-2324.
[9] Simonyan, A. Zisserman, Very deep convolutional networks for large-scale image recognition, arXiv
preprint arXiv:1409.1556, 2014.
14. International Journal on Computational Science & Applications (IJCSA) Vol.8, No.2/3, June 2018
14
[10] K.Yanai & Y. Kawano, Food image recognition using deep convolutional network with pre-training
and finetuning, in Multimedia & Expo Workshops (ICMEW), 2015 IEEE International Conference
on, 2015.
[11] M.Oquab, L. Bottou, I. Laptev, and J. Sivic, Learning and transferring mid-level image
representations using convolutional neural networks, in Proc. of IEEE Computer Vision and Pattern
Recognition, 2015
[12] A Fruit Detection System Using Deep Neural Networks (Inkyu Sa, Zongyuan Ge, Feras Dayoub, Ben
Upcroft, Tristan Perez, Chris McCool) Sensors (Basel) 2016 Aug; 16(8): 1222. Published online 2016
Aug 3. doi: 10.3390/s16081222
[13] L. Hou, Q. Wu, Q. Sun, H. Yang and P. Li, "Fruit recognition based on convolution neural network,"
2016 12th International Conference on Natural Computation, Fuzzy Systems and Knowledge
Discovery (ICNC-FSKD), Changsha, 2016, pp. 18-22
[14] Kaiming He et al. "Deep Residual Learning for Image Recognition." computer vision and pattern
recognition (2015), pp.770-778, 2015.
[15] Z.M. Khaing, Y. Naung and P. H. Htut, "Development of control system for fruit classification based
on convolutional neural network," 2018 IEEE Conference of Russian Young Researchers in Electrical
and Electronic Engineering (EIConRus), Moscow and St. Petersburg, Russia, 2018, pp. 1805-1807.
[16] Extracting Deep Features from a Single Image for Edit Propagation b Yuki Endo Satoshi Iizuka
Yoshihiro Kanamori Jun Mitaniy
[17] G. Zeng, "Fruit and vegetables classification system using image saliency and convolutional neural
network," 2017 IEEE 3rd Information Technology and Mechatronics Engineering Conference
(ITOEC), Chongqing, 2017, pp. 613-6