This paper presents the application of Wavelet Transform and Genetic Algorithm in a novel
steganography scheme. We employ a genetic algorithm based mapping function to embed data in Discrete Wavelet
Transform coefficients in 4x4 blocks on the cover image. The optimal pixel adjustment process is applied after
embedding the message. We utilize the frequency domain to improve the robustness of steganography and, we
implement Genetic Algorithm and Optimal Pixel Adjustment Process to obtain an optimal mapping function to
reduce the difference error between the cover and the stego-image, therefore improving the hiding capacity with
low distortions. Our Simulation results reveal that the novel scheme outperforms adaptive steganography technique
based on wavelet transform in terms of peak signal to noise ratio and capacity, 39.94 dB and 50% respectively.
Shallow vs. Deep Image Representations: A Comparative Study with Enhancements...CSCJournals
The traditional approach for solving the object recognition problem requires image representations to be first extracted and then fed to a learning model such as an SVM. These representations are handcrafted and heavily engineered by running the object image through a sequence of pipeline steps which requires a good prior knowledge of the problem domain in order to engineer these representations. Moreover, since the classification is done in a separate step, the resultant handcrafted representations are not tuned by the learning model which prevents it from learning complex representations that might would give it more discriminative power. However, in end-to-end deep learning models, image representations along with the classification decision boundary are all learnt directly from the raw data requiring no prior knowledge of the problem domain. These models deeply learn the object image representation hierarchically in multiple layers corresponding to multiple levels of abstraction resulting in representations that are more discriminative and give better results on challenging benchmarks. In contrast to the traditional handcrafted representations, the performance of deep representations improves with the introduction of more data, and more learning layers (more depth) and they perform well on large-scale machine learning problems. The purpose of this study is six fold: (1) review the literature of the pipeline processes used in the previous state-of-the-art codebook model approach for tackling the problem of generic object recognition, (2) Introduce several enhancements in the local feature extraction and normalization steps of the recognition pipeline, (3) compare the enhancements proposed to different encoding methods and contrast them to previous results, (4) experiment with current state-of-the-art deep model architectures used for object recognition, (5) compare between deep representations extracted from the deep learning model and shallow representations handcrafted through the recognition pipeline, and finally, (6) improve the results further by combining multiple different deep learning models into an ensemble and taking the maximum posterior probability.
Faster Training Algorithms in Neural Network Based Approach For Handwritten T...CSCJournals
Handwritten text and character recognition is a challenging task compared to recognition of handwritten numeral and computer printed text due to its large variety in nature. As practical pattern recognition problems uses bulk data and there is a one step self sufficient deterministic theory to resolve recognition problems by calculating inverse of Hessian Matrix and multiplication the inverse matrix it with first order local gradient vector. But in practical cases when neural network is large the inversing operation of the Hessian Matrix is not manageable and another condition must be satisfied the Hessian Matrix must be positive definite which may not be satishfied. In these cases some repetitive recursive models are taken. In several research work in past decade it was experienced that Neural Network based approach provides most reliable performance in handwritten character and text recognition but recognition performance depends upon some important factors like no of training samples, reliable features and no of features per character, training time, variety of handwriting etc. Important features from different types of handwriting are collected and are fed to the neural network for training. It is true that more no of features increases test efficiency but it takes longer time to converge the error curve. To reduce this training time effectively proper train algorithm should be chosen so that the system provides best train and test efficiency in least possible time that is to provide the system fastest intelligence. We have used several second order conjugate gradient algorithms for training of neural network. We have found that Scaled Conjugate Gradient Algorithm , a second order training algorithm as the fastest for training of neural network for our application. Training using SCG takes minimum time with excellent test efficiency. A scanned handwritten text is taken as input and character level segmentation is done. Some important and reliable features from each character are extracted and used as input to a neural network for training. When the error level reaches into a satisfactory level (10 -12 ) weights are accepted for testing a test script. Finally a lexicon matching algorithm solves the minor misclassification problems.
SAR Image Classification by Multilayer Back Propagation Neural NetworkIJMTST Journal
A novel descriptive feature extraction method of Discrete Fourier transform and neural network classifier for classification of Synthetic Aperture Radar (SAR) images is proposed. The classification process has the following stages (1) Image Segmentation using statistical Region Merging (SRM) (2) Polar transform and Feature extraction using Discrete Fourier Transform (3) Neural Network classification using back propagation. This is generally the first step in image analysis. Segmentation subdivides an image into its constituent parts or objects. The level to which this subdivision is carried depends on the problem being solved. The image segmentation in this study is performed using Statistical Region Merging proposed Richard Nock and Frank Nielsen. The key idea of the Statistical Region Merging model is to formulate image segmentation as an inference problem. Here the merging procedure is based on the theorem. Feature vectors as the input for the neural network. Polar transform is applied to segmented SAR image. The rotation problem under the Cartesian coordinates becomes the translation problem under the polar coordinates.
PERFORMANCE EVALUATION OF FUZZY LOGIC AND BACK PROPAGATION NEURAL NETWORK FOR...ijesajournal
ABSTRACT
Fuzzy c-mean is one of the efficient tools used in character recognition. Back propagation neural network is another powerful that may be used in such field. A comparison between fuzzy c-mean and BP neural network classifiers are presented in this research to obtain the performance of both classifiers. The comparison was based on recognition efficiency; this efficiency was evaluated as the ratio of the number of assigned characters with unknown one to the number of character set related to that character. The fuzzy C-mean and BP neural network algorithms were tested on a set of hand written and machine printed dataset named Chars74K dataset using Matlab (2016 b) programming language and the result was that neural network classifier gave 82% recognition efficiency while fuzzy c –mean gave 78%. Neural network classifier is more superior than fuzzy C-mean in recognition due to the limitations of processing time of fuzzy C-mean that requires smaller image size and eventually this will cause less efficiency.
Shallow vs. Deep Image Representations: A Comparative Study with Enhancements...CSCJournals
The traditional approach for solving the object recognition problem requires image representations to be first extracted and then fed to a learning model such as an SVM. These representations are handcrafted and heavily engineered by running the object image through a sequence of pipeline steps which requires a good prior knowledge of the problem domain in order to engineer these representations. Moreover, since the classification is done in a separate step, the resultant handcrafted representations are not tuned by the learning model which prevents it from learning complex representations that might would give it more discriminative power. However, in end-to-end deep learning models, image representations along with the classification decision boundary are all learnt directly from the raw data requiring no prior knowledge of the problem domain. These models deeply learn the object image representation hierarchically in multiple layers corresponding to multiple levels of abstraction resulting in representations that are more discriminative and give better results on challenging benchmarks. In contrast to the traditional handcrafted representations, the performance of deep representations improves with the introduction of more data, and more learning layers (more depth) and they perform well on large-scale machine learning problems. The purpose of this study is six fold: (1) review the literature of the pipeline processes used in the previous state-of-the-art codebook model approach for tackling the problem of generic object recognition, (2) Introduce several enhancements in the local feature extraction and normalization steps of the recognition pipeline, (3) compare the enhancements proposed to different encoding methods and contrast them to previous results, (4) experiment with current state-of-the-art deep model architectures used for object recognition, (5) compare between deep representations extracted from the deep learning model and shallow representations handcrafted through the recognition pipeline, and finally, (6) improve the results further by combining multiple different deep learning models into an ensemble and taking the maximum posterior probability.
Faster Training Algorithms in Neural Network Based Approach For Handwritten T...CSCJournals
Handwritten text and character recognition is a challenging task compared to recognition of handwritten numeral and computer printed text due to its large variety in nature. As practical pattern recognition problems uses bulk data and there is a one step self sufficient deterministic theory to resolve recognition problems by calculating inverse of Hessian Matrix and multiplication the inverse matrix it with first order local gradient vector. But in practical cases when neural network is large the inversing operation of the Hessian Matrix is not manageable and another condition must be satisfied the Hessian Matrix must be positive definite which may not be satishfied. In these cases some repetitive recursive models are taken. In several research work in past decade it was experienced that Neural Network based approach provides most reliable performance in handwritten character and text recognition but recognition performance depends upon some important factors like no of training samples, reliable features and no of features per character, training time, variety of handwriting etc. Important features from different types of handwriting are collected and are fed to the neural network for training. It is true that more no of features increases test efficiency but it takes longer time to converge the error curve. To reduce this training time effectively proper train algorithm should be chosen so that the system provides best train and test efficiency in least possible time that is to provide the system fastest intelligence. We have used several second order conjugate gradient algorithms for training of neural network. We have found that Scaled Conjugate Gradient Algorithm , a second order training algorithm as the fastest for training of neural network for our application. Training using SCG takes minimum time with excellent test efficiency. A scanned handwritten text is taken as input and character level segmentation is done. Some important and reliable features from each character are extracted and used as input to a neural network for training. When the error level reaches into a satisfactory level (10 -12 ) weights are accepted for testing a test script. Finally a lexicon matching algorithm solves the minor misclassification problems.
SAR Image Classification by Multilayer Back Propagation Neural NetworkIJMTST Journal
A novel descriptive feature extraction method of Discrete Fourier transform and neural network classifier for classification of Synthetic Aperture Radar (SAR) images is proposed. The classification process has the following stages (1) Image Segmentation using statistical Region Merging (SRM) (2) Polar transform and Feature extraction using Discrete Fourier Transform (3) Neural Network classification using back propagation. This is generally the first step in image analysis. Segmentation subdivides an image into its constituent parts or objects. The level to which this subdivision is carried depends on the problem being solved. The image segmentation in this study is performed using Statistical Region Merging proposed Richard Nock and Frank Nielsen. The key idea of the Statistical Region Merging model is to formulate image segmentation as an inference problem. Here the merging procedure is based on the theorem. Feature vectors as the input for the neural network. Polar transform is applied to segmented SAR image. The rotation problem under the Cartesian coordinates becomes the translation problem under the polar coordinates.
PERFORMANCE EVALUATION OF FUZZY LOGIC AND BACK PROPAGATION NEURAL NETWORK FOR...ijesajournal
ABSTRACT
Fuzzy c-mean is one of the efficient tools used in character recognition. Back propagation neural network is another powerful that may be used in such field. A comparison between fuzzy c-mean and BP neural network classifiers are presented in this research to obtain the performance of both classifiers. The comparison was based on recognition efficiency; this efficiency was evaluated as the ratio of the number of assigned characters with unknown one to the number of character set related to that character. The fuzzy C-mean and BP neural network algorithms were tested on a set of hand written and machine printed dataset named Chars74K dataset using Matlab (2016 b) programming language and the result was that neural network classifier gave 82% recognition efficiency while fuzzy c –mean gave 78%. Neural network classifier is more superior than fuzzy C-mean in recognition due to the limitations of processing time of fuzzy C-mean that requires smaller image size and eventually this will cause less efficiency.
Review and comparison of tasks scheduling in cloud computingijfcstjournal
Recently, there has been a dramatic increase in the popularity of cloud computing systems that rent
computing resources on-demand, bill on a pay-as-you-go basis, and multiplex many users on the same
physical infrastructure. It is a virtual pool of resources which are provided to users via Internet. It gives
users virtually unlimited pay-per-use computing resources without the burden of managing the underlying
infrastructure. One of the goals is to use the resources efficiently and gain maximum profit. Scheduling is a
critical problem in Cloud computing, because a cloud provider has to serve many users in Cloud
computing system. So scheduling is the major issue in establishing Cloud computing systems. The
scheduling algorithms should order the jobs in a way where balance between improving the performance
and quality of service and at the same time maintaining the efficiency and fairness among the jobs. This
paper introduces and explores some of the methods provided for in cloud computing has been scheduled.
Finally the waiting time and time to implement some of the proposed algorithm is evaluated
Multilayer extreme learning machine for hand movement prediction based on ele...journalBEEI
Brain computer interface (BCI) technology connects humans with machines via electroencephalography (EEG). The mechanism of BCI is pattern recognition, which proceeds by feature extraction and classification. Various feature extraction and classification methods can differentiate human motor movements, especially those of the hand. Combinations of these methods can greatly improve the accuracy of the results. This article explores the performances of nine feature-extraction types computed by a multilayer extreme learning machine (ML-ELM). The proposed method was tested on different numbers of EEG channels and different ML-ELM structures. Moreover, the performance of ML-ELM was compared with those of ELM, Support Vector Machine and Naive Bayes in classifying real and imaginary hand movements in offline mode. The ML-ELM with discrete wavelet transform (DWT) as feature extraction outperformed the other classification methods with highest accuracy 0.98. So, the authors also found that the structures influenced the accuracy of ML-ELM for different task, feature extraction used and channel used.
Multilinear Kernel Mapping for Feature Dimension Reduction in Content Based M...ijma
In the process of content-based multimedia retrieval, multimedia information is processed in order to
obtain descriptive features. Descriptive representation of features, results in a huge feature count, which
results in processing overhead. To reduce this descriptive feature overhead, various approaches have been
used to dimensional reduction, among which PCA and LDA are the most used methods. However, these
methods do not reflect the significance of feature content in terms of inter-relation among all dataset
features. To achieve a dimension reduction based on histogram transformation, features with low
significance can be eliminated. In this paper, we propose a feature dimensional reduction approaches to
achieve the dimension reduction approach based on a multi-linear kernel (MLK) modeling. A benchmark
dataset for the experimental work is taken and the proposed work is observed to be improved in analysis in
comparison to the conventional system.
Image fusion is a sub field of image processing in which more than one images are fused to create an image where all the objects are in focus. The process of image fusion is performed for multi-sensor and multi-focus images of the same scene. Multi-sensor images of the same scene are captured by different sensors whereas multi-focus images are captured by the same sensor. In multi-focus images, the objects in the scene which are closer to the camera are in focus and the farther objects get blurred. Contrary to it, when the farther objects are focused then closer objects get blurred in the image. To achieve an image where all the objects are in focus, the process of images fusion is performed either in spatial domain or in transformed domain. In recent times, the applications of image processing have grown immensely. Usually due to limited depth of field of optical lenses especially with greater focal length, it becomes impossible to obtain an image where all the objects are in focus. Thus, it plays an important role to perform other tasks of image processing such as image segmentation, edge detection, stereo matching and image enhancement. Hence, a novel feature-level multi-focus image fusion technique has been proposed which fuses multi-focus images. Thus, the results of extensive experimentation performed to highlight the efficiency and utility of the proposed technique is presented. The proposed work further explores comparison between fuzzy based image fusion and neuro fuzzy fusion technique along with quality evaluation indices.
Mr image compression based on selection of mother wavelet and lifting based w...ijma
Magnetic Resonance (MR) image is a medical image technique required enormous data to be stored and
transmitted for high quality diagnostic application. Various algorithms have been proposed to improve the
performance of the compression scheme. In this paper we extended the commonly used algorithms to image
compression and compared its performance. For an image compression technique, we have linked different
wavelet techniques using traditional mother wavelets and lifting based Cohen-Daubechies-Feauveau
wavelets with the low-pass filters of the length 9 and 7 (CDF 9/7) wavelet transform with Set Partition in
Hierarchical Trees (SPIHT) algorithm. A novel image quality index with highlighting shape of histogram
of the image targeted is introduced to assess image compression quality. The index will be used in place of
existing traditional Universal Image Quality Index (UIQI) “in one go”. It offers extra information about
the distortion between an original image and a compressed image in comparisons with UIQI. The proposed
index is designed based on modelling image compression as combinations of four major factors: loss of
correlation, luminance distortion, contrast distortion and shape distortion. This index is easy to calculate
and applicable in various image processing applications. One of our contributions is to demonstrate the
choice of mother wavelet is very important for achieving superior wavelet compression performances based
on proposed image quality indexes. Experimental results show that the proposed image quality index plays
a significantly role in the quality evaluation of image compression on the open sources “BrainWeb:
Simulated Brain Database (SBD) ”.
A systematic image compression in the combination of linear vector quantisati...eSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
A Literature Survey: Neural Networks for object detectionvivatechijri
Humans have a great capability to distinguish objects by their vision. But, for machines object
detection is an issue. Thus, Neural Networks have been introduced in the field of computer science. Neural
Networks are also called as ‘Artificial Neural Networks’ [13]. Artificial Neural Networks are computational
models of the brain which helps in object detection and recognition. This paper describes and demonstrates the
different types of Neural Networks such as ANN, KNN, FASTER R-CNN, 3D-CNN, RNN etc. with their accuracies.
From the study of various research papers, the accuracies of different Neural Networks are discussed and
compared and it can be concluded that in the given test cases, the ANN gives the best accuracy for the object
detection.
A simple framework for contrastive learning of visual representationsDevansh16
Link: https://machine-learning-made-simple.medium.com/learnings-from-simclr-a-framework-contrastive-learning-for-visual-representations-6c145a5d8e99
If you'd like to discuss something, text me on LinkedIn, IG, or Twitter. To support me, please use my referral link to Robinhood. It's completely free, and we both get a free stock. Not using it is literally losing out on free money.
Check out my other articles on Medium. : https://rb.gy/zn1aiu
My YouTube: https://rb.gy/88iwdd
Reach out to me on LinkedIn. Let's connect: https://rb.gy/m5ok2y
My Instagram: https://rb.gy/gmvuy9
My Twitter: https://twitter.com/Machine01776819
My Substack: https://devanshacc.substack.com/
Live conversations at twitch here: https://rb.gy/zlhk9y
Get a free stock on Robinhood: https://join.robinhood.com/fnud75
This paper presents SimCLR: a simple framework for contrastive learning of visual representations. We simplify recently proposed contrastive self-supervised learning algorithms without requiring specialized architectures or a memory bank. In order to understand what enables the contrastive prediction tasks to learn useful representations, we systematically study the major components of our framework. We show that (1) composition of data augmentations plays a critical role in defining effective predictive tasks, (2) introducing a learnable nonlinear transformation between the representation and the contrastive loss substantially improves the quality of the learned representations, and (3) contrastive learning benefits from larger batch sizes and more training steps compared to supervised learning. By combining these findings, we are able to considerably outperform previous methods for self-supervised and semi-supervised learning on ImageNet. A linear classifier trained on self-supervised representations learned by SimCLR achieves 76.5% top-1 accuracy, which is a 7% relative improvement over previous state-of-the-art, matching the performance of a supervised ResNet-50. When fine-tuned on only 1% of the labels, we achieve 85.8% top-5 accuracy, outperforming AlexNet with 100X fewer labels.
Comments: ICML'2020. Code and pretrained models at this https URL
Subjects: Machine Learning (cs.LG); Computer Vision and Pattern Recognition (cs.CV); Machine Learning (stat.ML)
Cite as: arXiv:2002.05709 [cs.LG]
(or arXiv:2002.05709v3 [cs.LG] for this version)
Submission history
From: Ting Chen [view email]
[v1] Thu, 13 Feb 2020 18:50:45 UTC (5,093 KB)
[v2] Mon, 30 Mar 2020 15:32:51 UTC (5,047 KB)
[v3] Wed, 1 Jul 2020 00:09:08 UTC (5,829 KB)
AN ENHANCEMENT FOR THE CONSISTENT DEPTH ESTIMATION OF MONOCULAR VIDEOS USING ...mlaij
Depth estimation has made great progress in the last few years due to its applications in robotics science
and computer vision. Various methods have been implemented and enhanced to estimate the depth without
flickers and missing holes. Despite this progress, it is still one of the main challenges for researchers,
especially for the video applications which have more complexity of the neural network which af ects the
run time. Moreover to use such input like monocular video for depth estimation is considered an attractive
idea, particularly for hand-held devices such as mobile phones, they are very popular for capturing
pictures and videos, in addition to having a limited amount of RAM. Here in this work, we focus on
enhancing the existing consistent depth estimation for monocular videos approach to be with less usage of
RAM and with using less number of parameters without having a significant reduction in the quality of the
depth estimation.
The comparison of the text classification methods to be used for the analysis...ijcsit
Text classification is used for the purpose of preventing the leakage of the data which is highly important
within the institution through unallowed ways. The results obtained from the text classification process
should be integrated into the DLP architecture immediately. The data flowing through the net requires
instant control and the flow of the sensitive data should be prevented. The use of the machinery learning
methods is required to perform the text classification which will be integrated into the DLP architecture.
The experimental results of the comparison of text classification methods to be used in the interface written
on the ICAP protocol have been prepared in the networked architecture developed for the DLP system.
Also, the choice of the text classification method to be used in the instant control of the sensitive data has
been carried out. The DLP text classification architecture developed helps decide the classification method
through the examination of the data in motion. The method to be chosen for the text classification is applied
to the ICAP protocol, and the analysis of the sensitive data and confidentiality are provided.
Audio Steganography Coding Using the Discreet Wavelet TransformsCSCJournals
The performance of audio steganography compression system using discreet wavelet transform (DWT) is investigated. Audio steganography coding is the technology of transforming stego-speech into efficiently encoded version that can be decoded in the receiver side to produce a close representation of the initial signal (non compressed). Experimental results prove the efficiency of the used compression technique since the compressed stego-speech are perceptually intelligible and indistinguishable from the equivalent initial signal, while being able to recover the initial stego-speech with slight degradation in the quality .
Understanding Relational Violence Enacted by MenBill Warters
Slides from a presentation given by Bill Warters on October 16th at the Peace and Justice Studies Association annual conference held at James Madison University.
Here's the session description from the program:
Understanding Relational Violence Enacted by Men: Perceptions, Motivations and the Skills of Transformation
This session will invite the participants into the perceptual world of men who have been repeatedly violent to a partner or spouse. Drawing on 49 in-depth interviews we'll explore an inductive model I developed that illustrates the journey these men must travel to move from being a batterer to being a just and peaceable companion. We'll look at the motivations and satisfactions of using violence as reported by men and consider the pernicious role that male socialization patterns can play. As you will see, gendered worldviews and military training contribute greatly to problems in this area. Finally we'll explore the necessary skills men may need in order to become their best, most peaceable selves.
Review and comparison of tasks scheduling in cloud computingijfcstjournal
Recently, there has been a dramatic increase in the popularity of cloud computing systems that rent
computing resources on-demand, bill on a pay-as-you-go basis, and multiplex many users on the same
physical infrastructure. It is a virtual pool of resources which are provided to users via Internet. It gives
users virtually unlimited pay-per-use computing resources without the burden of managing the underlying
infrastructure. One of the goals is to use the resources efficiently and gain maximum profit. Scheduling is a
critical problem in Cloud computing, because a cloud provider has to serve many users in Cloud
computing system. So scheduling is the major issue in establishing Cloud computing systems. The
scheduling algorithms should order the jobs in a way where balance between improving the performance
and quality of service and at the same time maintaining the efficiency and fairness among the jobs. This
paper introduces and explores some of the methods provided for in cloud computing has been scheduled.
Finally the waiting time and time to implement some of the proposed algorithm is evaluated
Multilayer extreme learning machine for hand movement prediction based on ele...journalBEEI
Brain computer interface (BCI) technology connects humans with machines via electroencephalography (EEG). The mechanism of BCI is pattern recognition, which proceeds by feature extraction and classification. Various feature extraction and classification methods can differentiate human motor movements, especially those of the hand. Combinations of these methods can greatly improve the accuracy of the results. This article explores the performances of nine feature-extraction types computed by a multilayer extreme learning machine (ML-ELM). The proposed method was tested on different numbers of EEG channels and different ML-ELM structures. Moreover, the performance of ML-ELM was compared with those of ELM, Support Vector Machine and Naive Bayes in classifying real and imaginary hand movements in offline mode. The ML-ELM with discrete wavelet transform (DWT) as feature extraction outperformed the other classification methods with highest accuracy 0.98. So, the authors also found that the structures influenced the accuracy of ML-ELM for different task, feature extraction used and channel used.
Multilinear Kernel Mapping for Feature Dimension Reduction in Content Based M...ijma
In the process of content-based multimedia retrieval, multimedia information is processed in order to
obtain descriptive features. Descriptive representation of features, results in a huge feature count, which
results in processing overhead. To reduce this descriptive feature overhead, various approaches have been
used to dimensional reduction, among which PCA and LDA are the most used methods. However, these
methods do not reflect the significance of feature content in terms of inter-relation among all dataset
features. To achieve a dimension reduction based on histogram transformation, features with low
significance can be eliminated. In this paper, we propose a feature dimensional reduction approaches to
achieve the dimension reduction approach based on a multi-linear kernel (MLK) modeling. A benchmark
dataset for the experimental work is taken and the proposed work is observed to be improved in analysis in
comparison to the conventional system.
Image fusion is a sub field of image processing in which more than one images are fused to create an image where all the objects are in focus. The process of image fusion is performed for multi-sensor and multi-focus images of the same scene. Multi-sensor images of the same scene are captured by different sensors whereas multi-focus images are captured by the same sensor. In multi-focus images, the objects in the scene which are closer to the camera are in focus and the farther objects get blurred. Contrary to it, when the farther objects are focused then closer objects get blurred in the image. To achieve an image where all the objects are in focus, the process of images fusion is performed either in spatial domain or in transformed domain. In recent times, the applications of image processing have grown immensely. Usually due to limited depth of field of optical lenses especially with greater focal length, it becomes impossible to obtain an image where all the objects are in focus. Thus, it plays an important role to perform other tasks of image processing such as image segmentation, edge detection, stereo matching and image enhancement. Hence, a novel feature-level multi-focus image fusion technique has been proposed which fuses multi-focus images. Thus, the results of extensive experimentation performed to highlight the efficiency and utility of the proposed technique is presented. The proposed work further explores comparison between fuzzy based image fusion and neuro fuzzy fusion technique along with quality evaluation indices.
Mr image compression based on selection of mother wavelet and lifting based w...ijma
Magnetic Resonance (MR) image is a medical image technique required enormous data to be stored and
transmitted for high quality diagnostic application. Various algorithms have been proposed to improve the
performance of the compression scheme. In this paper we extended the commonly used algorithms to image
compression and compared its performance. For an image compression technique, we have linked different
wavelet techniques using traditional mother wavelets and lifting based Cohen-Daubechies-Feauveau
wavelets with the low-pass filters of the length 9 and 7 (CDF 9/7) wavelet transform with Set Partition in
Hierarchical Trees (SPIHT) algorithm. A novel image quality index with highlighting shape of histogram
of the image targeted is introduced to assess image compression quality. The index will be used in place of
existing traditional Universal Image Quality Index (UIQI) “in one go”. It offers extra information about
the distortion between an original image and a compressed image in comparisons with UIQI. The proposed
index is designed based on modelling image compression as combinations of four major factors: loss of
correlation, luminance distortion, contrast distortion and shape distortion. This index is easy to calculate
and applicable in various image processing applications. One of our contributions is to demonstrate the
choice of mother wavelet is very important for achieving superior wavelet compression performances based
on proposed image quality indexes. Experimental results show that the proposed image quality index plays
a significantly role in the quality evaluation of image compression on the open sources “BrainWeb:
Simulated Brain Database (SBD) ”.
A systematic image compression in the combination of linear vector quantisati...eSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
A Literature Survey: Neural Networks for object detectionvivatechijri
Humans have a great capability to distinguish objects by their vision. But, for machines object
detection is an issue. Thus, Neural Networks have been introduced in the field of computer science. Neural
Networks are also called as ‘Artificial Neural Networks’ [13]. Artificial Neural Networks are computational
models of the brain which helps in object detection and recognition. This paper describes and demonstrates the
different types of Neural Networks such as ANN, KNN, FASTER R-CNN, 3D-CNN, RNN etc. with their accuracies.
From the study of various research papers, the accuracies of different Neural Networks are discussed and
compared and it can be concluded that in the given test cases, the ANN gives the best accuracy for the object
detection.
A simple framework for contrastive learning of visual representationsDevansh16
Link: https://machine-learning-made-simple.medium.com/learnings-from-simclr-a-framework-contrastive-learning-for-visual-representations-6c145a5d8e99
If you'd like to discuss something, text me on LinkedIn, IG, or Twitter. To support me, please use my referral link to Robinhood. It's completely free, and we both get a free stock. Not using it is literally losing out on free money.
Check out my other articles on Medium. : https://rb.gy/zn1aiu
My YouTube: https://rb.gy/88iwdd
Reach out to me on LinkedIn. Let's connect: https://rb.gy/m5ok2y
My Instagram: https://rb.gy/gmvuy9
My Twitter: https://twitter.com/Machine01776819
My Substack: https://devanshacc.substack.com/
Live conversations at twitch here: https://rb.gy/zlhk9y
Get a free stock on Robinhood: https://join.robinhood.com/fnud75
This paper presents SimCLR: a simple framework for contrastive learning of visual representations. We simplify recently proposed contrastive self-supervised learning algorithms without requiring specialized architectures or a memory bank. In order to understand what enables the contrastive prediction tasks to learn useful representations, we systematically study the major components of our framework. We show that (1) composition of data augmentations plays a critical role in defining effective predictive tasks, (2) introducing a learnable nonlinear transformation between the representation and the contrastive loss substantially improves the quality of the learned representations, and (3) contrastive learning benefits from larger batch sizes and more training steps compared to supervised learning. By combining these findings, we are able to considerably outperform previous methods for self-supervised and semi-supervised learning on ImageNet. A linear classifier trained on self-supervised representations learned by SimCLR achieves 76.5% top-1 accuracy, which is a 7% relative improvement over previous state-of-the-art, matching the performance of a supervised ResNet-50. When fine-tuned on only 1% of the labels, we achieve 85.8% top-5 accuracy, outperforming AlexNet with 100X fewer labels.
Comments: ICML'2020. Code and pretrained models at this https URL
Subjects: Machine Learning (cs.LG); Computer Vision and Pattern Recognition (cs.CV); Machine Learning (stat.ML)
Cite as: arXiv:2002.05709 [cs.LG]
(or arXiv:2002.05709v3 [cs.LG] for this version)
Submission history
From: Ting Chen [view email]
[v1] Thu, 13 Feb 2020 18:50:45 UTC (5,093 KB)
[v2] Mon, 30 Mar 2020 15:32:51 UTC (5,047 KB)
[v3] Wed, 1 Jul 2020 00:09:08 UTC (5,829 KB)
AN ENHANCEMENT FOR THE CONSISTENT DEPTH ESTIMATION OF MONOCULAR VIDEOS USING ...mlaij
Depth estimation has made great progress in the last few years due to its applications in robotics science
and computer vision. Various methods have been implemented and enhanced to estimate the depth without
flickers and missing holes. Despite this progress, it is still one of the main challenges for researchers,
especially for the video applications which have more complexity of the neural network which af ects the
run time. Moreover to use such input like monocular video for depth estimation is considered an attractive
idea, particularly for hand-held devices such as mobile phones, they are very popular for capturing
pictures and videos, in addition to having a limited amount of RAM. Here in this work, we focus on
enhancing the existing consistent depth estimation for monocular videos approach to be with less usage of
RAM and with using less number of parameters without having a significant reduction in the quality of the
depth estimation.
The comparison of the text classification methods to be used for the analysis...ijcsit
Text classification is used for the purpose of preventing the leakage of the data which is highly important
within the institution through unallowed ways. The results obtained from the text classification process
should be integrated into the DLP architecture immediately. The data flowing through the net requires
instant control and the flow of the sensitive data should be prevented. The use of the machinery learning
methods is required to perform the text classification which will be integrated into the DLP architecture.
The experimental results of the comparison of text classification methods to be used in the interface written
on the ICAP protocol have been prepared in the networked architecture developed for the DLP system.
Also, the choice of the text classification method to be used in the instant control of the sensitive data has
been carried out. The DLP text classification architecture developed helps decide the classification method
through the examination of the data in motion. The method to be chosen for the text classification is applied
to the ICAP protocol, and the analysis of the sensitive data and confidentiality are provided.
Audio Steganography Coding Using the Discreet Wavelet TransformsCSCJournals
The performance of audio steganography compression system using discreet wavelet transform (DWT) is investigated. Audio steganography coding is the technology of transforming stego-speech into efficiently encoded version that can be decoded in the receiver side to produce a close representation of the initial signal (non compressed). Experimental results prove the efficiency of the used compression technique since the compressed stego-speech are perceptually intelligible and indistinguishable from the equivalent initial signal, while being able to recover the initial stego-speech with slight degradation in the quality .
Understanding Relational Violence Enacted by MenBill Warters
Slides from a presentation given by Bill Warters on October 16th at the Peace and Justice Studies Association annual conference held at James Madison University.
Here's the session description from the program:
Understanding Relational Violence Enacted by Men: Perceptions, Motivations and the Skills of Transformation
This session will invite the participants into the perceptual world of men who have been repeatedly violent to a partner or spouse. Drawing on 49 in-depth interviews we'll explore an inductive model I developed that illustrates the journey these men must travel to move from being a batterer to being a just and peaceable companion. We'll look at the motivations and satisfactions of using violence as reported by men and consider the pernicious role that male socialization patterns can play. As you will see, gendered worldviews and military training contribute greatly to problems in this area. Finally we'll explore the necessary skills men may need in order to become their best, most peaceable selves.
Flash Series! Ecommerce Navigation: The Importance of Getting It RightHanapin Marketing
Millions of people shop online these days. With projections saying that ecommerce sales will exceed 550 billion dollars in 2016, not only do you need to have a PPC strategy in place, you need to make sure you site is user-friendly. And the most important element of you ecommerce website is the site navigation.
In this presentation, ecommerce expert and co-founder of Edgacent, Linda Bustos, dives deep into why ecommerce navigation is the most important area of your site to get right. Learn how to constructively improve your navigation through CAMA (Catalog and Merchandising Architecture) to develop strategic navigation requirements.
Review Paper on Recovery of Data during Software FaultAM Publications
In this paper here we will discuss different types of technique that are used for recovery of data during
software fault. The major objective of this paper to specify recovery technique that are used during software and
hardware fault.
A Low Power Hybrid Partition SRAM based TCAM with a Parity BitAM Publications
Ternary Content Addressable memory provides high speed search operation. When compared to RAM, TCAM suffers
certain limitation like low bit density, slow access time, high cost per bit. In this paper, we introduced a parity bit to boost the
searching speed HP SRAM based TCAM with less power and delay. A novel memory architecture called HP SRAM based TCAM,
which emulates TCAM functionality with SRAM memory. The architecture of HP SRAM based TCAM with a parity bit was
verified by VHDL in ModelSim.
Mastering the Art of Ecommerce Reporting With ExcelHanapin Marketing
If you work at an ecommerce company or work with an ecommerce site, you know how important it is to get accurate reporting, in order to manage your marketing programs.
In this presentation, experts from Hanapin Marketing walk you through a number of reports and calculations you can use in excel to make sure you have the more relevant and valuable information for your ecommerce accounts.
This is the final presentation for project named "Audio Steganography" for btech final year computer science students, and without the screenshots of the progress this can b used as a synopsis presentation.
Programmatic Advertising: How To Join In On the FunHanapin Marketing
Carrie Albright, Associate Director of Services at Hanapin Marketing, discusses how to get into Programmatic Advertising, vendors you can hire, the definitions of different programmatic strategies, and steps to building your strategic plan.
Originally presented at Hero Conf London in October 2016.
Nowadays, digital watermarking has many
applications such as broadcast monitoring, owner identification,
proof of ownership, transaction tracking. Embedding a hidden
stream of bits in a file is called Digital Watermarking. This paper
introduces a LSB information hiding algorithm which can lift the
wavelet transform image. LSB based Steganography embeds the
hiding text message in least significant bit of the pixels. The
proposed method has good invisibility, robustness for a lot of
hidden attacks. As we think about the capacity lead us to think
about improved approach which can be achieved through
hardware implementation system by using Field Programmable
Gate Array (FPGA). In this paper hardware implementation of
digital watermarking system is proposed. MATLAB is used to
convert images into pixel-format files and to observe simulation
results. To implement this paper XPS & VB are needed. In XPS,
first select hardware & software components then by adding
source and header files & converting into bit streams and
download into FPGA, to obtain Stego image.
Neural network based image compression with lifting scheme and rlceSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
Neural network based image compression with lifting scheme and rlceSAT Journals
Abstract Image compression is a process that helps in fast data transfer and effective memory utilization. In effect, the objective is to reduce data redundancy of the image while retaining high image quality. This paper proposes an approach for Wavelet based Image Compression using MLFF Neural Network with Error Back Propagation (EBP) training algorithm for second level approximation component and modified RLC is applied on second level Horizontal and Vertical components with threshold to discard insignificant coefficients. All other sub-bands (i.e. Detail components of 1st level and Diagonal component of 2nd level) that do not affect the quality of image (both subjective and objective) are neglected. With the proposed method in this paper CR (27.899), PSNR (70.16 dB) and minimum MSE (0.0063) of still image obtained are better when compared with SOFM, EZW and SPIHT. Keywords: Image compression, wavelet, MLFFNN, EBP
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
An Efficient Frame Embedding Using Haar Wavelet Coefficients And Orthogonal C...IJERA Editor
Digital media, applications, copyright defense, and multimedia security have become a vital aspect of our daily life. Digital watermarking is a technology used for the copyright security of digital applications. In this work we have dealt with a process able to mark digital pictures with a visible and semi invisible hided information, called watermark. This process may be the basis of a complete copyright protection system. Watermarking is implemented here using Haar Wavelet Coefficients and Principal Component analysis. Experimental results show high imperceptibility where there is no noticeable difference between the watermarked video frames and the original frame in case of invisible watermarking, vice-versa for semi visible implementation. The watermark is embedded in lower frequency band of Wavelet Transformed cover image. The combination of the two transform algorithm has been found to improve performance of the watermark algorithm. The robustness of the watermarking scheme is analyzed by means of two distinct performance measures viz. Peak Signal to Noise Ratio (PSNR) and Normalized Coefficient (NC).
Comparison Between Levenberg-Marquardt And Scaled Conjugate Gradient Training...CSCJournals
The Internet paved way for information sharing all over the world decades ago and its popularity for distribution of data has spread like a wildfire ever since. Data in the form of images, sounds, animations and videos is gaining users’ preference in comparison to plain text all across the globe. Despite unprecedented progress in the fields of data storage, computing speed and data transmission speed, the demands of available data and its size (due to the increase in both, quality and quantity) continue to overpower the supply of resources. One of the reasons for this may be how the uncompressed data is compressed in order to send it across the network. This paper compares the two most widely used training algorithms for multilayer perceptron (MLP) image compression – the Levenberg-Marquardt algorithm and the Scaled Conjugate Gradient algorithm. We test the performance of the two training algorithms by compressing the standard test image (Lena or Lenna) in terms of accuracy and speed. Based on our results, we conclude that both algorithms were comparable in terms of speed and accuracy. However, the Levenberg- Marquardt algorithm has shown slightly better performance in terms of accuracy (as found in the average training accuracy and mean squared error), whereas the Scaled Conjugate Gradient algorithm faired better in terms of speed (as found in the average training iteration) on a simple MLP structure (2 hidden layers).
International Journal of Engineering and Science Invention (IJESI)inventionjournals
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online.
Improvement in Traditional Set Partitioning in Hierarchical Trees (SPIHT) Alg...AM Publications
In this paper, an improved SPIHT algorithm based on Huffman coding and discrete wavelet transform for
image compression has been developed. The traditional SPIHT algorithm application is limited in terms of PSNR and
Compression Ratio. The improved SPIHT algorithm gives better performance as compared with traditional SPIHT
algorithm; here we used additional Huffman encoder along with SPIHT encoder. Huffman coding is an entropy
encoding algorithm uses a specific method for choosing the representation for each symbol, resulting in a prefix code.
The input gray scale image is decomposed by 'bior4.4' wavelet and wavelet coefficients are indexed and scanned by
DWT, then applied to encode the coefficient by SPIHT encoder followed by Huffman encoder which gives the
compressed image. At receiver side the decoding process is applied by Huffman decoder followed by SPIHT decoder.
The Reconstructed gray scale image looks similar to the applied gray scale image.
Semantic Image Retrieval Using Relevance Feedback dannyijwest
This paper presents optimized interactive content-based image retrieval framework based on AdaBoost
learning method. As we know relevance feedback (RF) is online process, so we have optimized the learning
process by considering the most positive image selection on each feedback iteration. To learn the system we
have used AdaBoost. The main significances of our system are to address the small training sample and to
reduce retrieval time. Experiments are conducted on 1000 semantic colour images from Corel database to
demonstrate the effectiveness of the proposed framework. These experiments employed large image
database and combined RCWFs and DT-CWT texture descriptors to represent content of the images.
Abstract: Primarily due to the progresses in super resolution imagery, the methods of segment-based image analysis for generating and updating geographical information are becoming more and more important. This work presents a image segmentation based on colour features with K-means clustering. The entire work is divided into two stages. First enhancement of color separation of satellite image using de correlation stretching is carried out and then the regions are grouped into a set of five classes using K-means clustering algorithm. At first, the spatial data is concentrated focused around every pixel, and at that point two separating procedures are added to smother the impact of pseudoedges. What's more, the spatial data weight is built and grouped with k-means bunching, and the regularization quality in every district is controlled by the bunching focus esteem. The exploratory results, on both reenacted and genuine datasets, demonstrate that the proposed methodology can adequately lessen the pseudoedges of the aggregate variety regularization in the level.
DEVELOPMENT OF TODDLER FAMILY CADRE TRAINING BASED ON ANDROID APPLICATIONS IN...AM Publications
Toddler family cadre is a community members work voluntarily in fostering and providing information to parents of toddlers about how to properly care for children. Toddler Family cadre desperately need training to increase their skills. There are still a few Toddler family cadres who get training so that the knowledge and skills of parents and other family members in developing toddlers' growth through physical stimulation, motoric intelligence, emotional and social economy as well as possible are still lacking. The purpose of this study is to develop an Android- assisted Toddler family cadre training model in Demak. This research is research in tian research and development. The research location was in Demak Regency. Toddler family cadres became the object of this research. Development of Toddler family cadre training models assisted by Android in Demak is feasible to be used as an effort to improve Toddler Family cadres' capabilities.
TESTING OF COMPOSITE ON DROP-WEIGHT IMPACT TESTING AND DAMAGE IDENTIFICATION ...AM Publications
In recent years the use of composite materials in structural components has become increasingly common in a wide range of engineering applications. Composite materials offer numerous advantages over more conventional materials because of their superior specific properties, but a serious obstacle to a more widespread use of these materials is their high sensitivity to localized impact loading. This paper presents an experimental study to assess the impact response of drop weight impact tests on fiber reinforced polymer composites with deferent load and damage identification of composite using Non-destructive testing techniques ultrasonic testing (UT) C scan. In the study includes checking the strength of the specimen, plotting of graphs between the height and the impact energy obtained and tabulating the results after conducting the various functional tests.
THE USE OF FRACTAL GEOMETRY IN TILING MOTIF DESIGNAM Publications
In this paper I will present the use of fractal geometry to design tile motifs. A fractal is a geometric figure that combines the several characteristics among others: its parts have the same form as the whole, fragmented, and formation by iteration. The concept of fractals has been spread over all fields of sciences, technology, and art. This paper aims to provide an algorithm to creating motifs of tile algorithm for create the tile motif consists of base, iteration, coloration and duplication. In order to help the reader better understand the algorithm, I will present some script using Matlab. We describe a mathematically based algorithm that can fill a spatial region with sequence of randomly placed which may be transformed copies of one motif or several motifs. By using this algorithm, I can produce thousand variety of aesthetically pleasing tile motifs, of which we show a number of examples.
TWO-DIMENSIONAL INVERSION FINITE ELEMENT MODELING OF MAGNETOTELLURIC DATA: CA...AM Publications
Two-dimensional resistivity analysis of magnetotelluric data has been done at “Z” geothermal area which is located in southern part of Indonesia. The objective is to understand subsurface structure beneath reasearch area based on 2-D modeling of magnetotelluric data. The inversion finite element method were used for numerical simulations which requires discretization on the boundary of the modeling domain. The modeling results of magnetotelluric data shows relativity structure dissemination: 0-10 ohm.m in a thickness of 1 km (Clay Cap), 10-100 ohm.m with 1-2 km depth respectively (reservoir zone), and on a scale of 100-1000 ohm.m in a depth of 2-3 km (heat source zone). The result of relativity structure can be used to delineate an area with geothermal prospect around 12 km2.
USING THE GENETIC ALGORITHM TO OPTIMIZE LASER WELDING PARAMETERS FOR MARTENSI...AM Publications
To achieve the pre-set welding size, this paper presents the optimization of the constrained overlap laser welding input parameters for AISI 416 and AISI 440FSe stainless, thickness 0.5 mm. In this study, the proposed optimization algorithm is the Genetic Algorithm (GA). After training 10 times for 30 NP (population size), each training repeated 200 times, the results achieved as expected. The error is compared with the result of the affirmation experiment not exceeding 5%.
ANALYSIS AND DESIGN E-MARKETPLACE FOR MICRO, SMALL AND MEDIUM ENTERPRISESAM Publications
The Ministry of Cooperatives and Small and Medium Enterprises launched in 2018 the number of Micro, Small and Medium Enterprises (MSMEs) in Indonesia as many as 58.97 million people. It is predicted that the number of MSMEs players in 2019 will amount to 59.2 million. This shows that the Indonesian people have made changes in the field of family economics which initially as consumptive are now productive. The community prefers to carry out activities that can increase family income. Future MSMEs remain the mainstay of the national economy. In accordance with the government roadmap, in 2020 e-commerce transactions are predicted to reach Rp1,300 trillion or equivalent to USD130 billion. According to data from the Central Statistics Agency (BPS), the contribution of MSMEs to Indonesia's Gross Domestic Product (GDP) reached 61.41%, with the number of MSMEs reaching almost 60 million units. However, only around 8% or 3.79 million of the 59.2 million MSMEs players have used online platforms to market their products. Based on the above problems, researchers conducted research on the analysis and display of E-Marketplace for MSMEs in Indonesia. The type of research used is action research. The object of research is MSMEs which are under the Office of Industry and Trade of Sragen Regency. The method of data collection is by techniques: (1) interview, (2) documentation (3) observation, (4) literature study. The researcher uses the waterfall method in developing the system. The research team has successfully analyzed the E-Market place according to the results of data collection. The research team has succeeded in designing the E-Marketplace for MSMEs. E-Marketplace designed can be used by admin, MSME and user. Admin is in charge of managing E-Marketplace and has full access rights. MSMEs can register online and manage their products in E-Marketplace. Users or buyers can search data in E-Marketplace as desired. To make transactions, users can interact directly with MSMEs according to the data provided in E-Marketplace. E-Marketplace can be used for marketing together MSMEs products. This e-marketplace can be accessed at www.umkmonline.com
REMOTE SENSING AND GEOGRAPHIC INFORMATION SYSTEMS AM Publications
Remote sensing technology's increasing accessibility helps us observe research and learn about our globe in ways we could only imagine a generation ago. Guides to profound knowledge of historical, conceptual and practical uses of remote sensing which is increasing GIS technology. This paper will go briefly through remote sensing benefits, history, technology and the GIS and remote sensing integration and their applications. Remote sensing (RS) is used in mapping the predicted and actual species and dominates the ecosystem canopy.
EVALUATE THE STRAIN ENERGY ERROR FOR THE LASER WELD BY THE H-REFINEMENT OF TH...AM Publications
Currently, the finite element method (FEM) is still one of the useful tools in numerical simulation for technical problems. With this method, a continuum model presented by a certain number of elements with a simple approximation field causes the presence of discretization error in solutions. This paper considers the butt weld by laser which subjected the tension for AISI 1018 steel highness 8 mm. The aim of the study is to use the h-refinement of the FEM in estimation the strain energy error for the laser weld mentioned. The results show that the stability of the h-refinement shown by the value of the relative error of the strain energy is quite small, specifically; FEM is less than 5.7% and extra is no more than 3.7%.
HMM APPLICATION IN ISOLATED WORD SPEECH RECOGNITIONAM Publications
Speech recognition is always being an all-time trendy topic for discussion and also for researches and we see a major application in our life. This paper provides the work done on the application of Hidden Markov model to implement isolated word speech recognition on MATLAB and to develop and train the system for set of self-selective words for specific user (user dependent) to get maximum efficiency in word recognition system. Which uses the forward and Baum-welch algorithm and fitting Gaussian of the Baum-welch algorithm for all the iteration perform. We use a sample of 7 alphabets which are recorded in 15 different ways giving total of 105 word to use for training with each word with 15 variations. This system can be used in real world in system security using voice security system and mainly for children and impaired people.
PEDESTRIAN DETECTION IN LOW RESOLUTION VIDEOS USING A MULTI-FRAME HOG-BASED D...AM Publications
Detecting pedestrians in low resolution videos is a challenging task, due to the small size of pedestrians in the images and the limited information. In practical outdoor surveillance scenarios the pedestrian size is usually small. Existing state-of-the-art pedestrian detection methods that use histogram of oriented gradient (HOG) features have poor performance in this problem domain. To compensate for the lack of information in a single frame, we propose a novel detection method that recognizes pedestrians in a short sequence of frames. Namely, we take the single-frame HOG-based detector and extend it to multiple frames. Our detector is applied to regions containing potential moving objects. In the case of video taken from a moving camera on an aerial platform, video stabilization is first performed to register the frames. A classifier is then applied to features extracted from spatio-temporal volumes surrounding the potential moving objects. On challenging stationary and aerial video datasets, our detection accuracy outperforms several state-of-the-art algorithms.
The aim of this paper is to help the blind people to identify and catch the public transport vehicles with the help of Light Fidelity technology. It is a Navigation aid. When the bus arrives at the bus stand, transmitter in the bus transmits the light signals and receiver in the stick, receives the light signals and a sound signal is generated through the speaker present in the stick. The sound message contains the bus number and the destination of the bus. In addition to this, if the person is absconded or lost, details of the location will be sent to his/her family members by pressing a button. This is made possible with the help of Global System for Mobile (GSM). Finally, presence of water can be detected along the blind person’s path, with the help of water sensors.
EFFECT OF SILICON - RUBBER (SR) SHEETS AS AN ALTERNATIVE FILTER ON HIGH AND L...AM Publications
A digital radiography delivers a radiation dose to patients; therefore it poses potential risk to the patients. One effort to reduce dose is carried out using a radiation filter, e.g. Silicone Rubber (SR) sheet. The purpose of this research was to determine the impact of the SR sheet on the high contrast objects (HCO) and the low contrast objects (LCO). The dose reduction was determined from attenuation x-rays before and after using the SR sheet. Assessment of HCO and LCO was observed from CDR TOR phantom at tube voltage of 48 kVp and tube current of 8 mAs. The physical parameter to assess image quality was the Signal to Noise Ratio (SNR) value in LCO. The maximum x-ray attenuation using the SR sheet is 48.82%. The visibility of the HCO remains the same, namely 16 objects; however the LCO slighly decreases from 14 objects to 13 objects after using the SR sheet. The SNR value decreases with an average value of 15.17%.Therefore, the SR sheet as a alternative filter has no effect on the HCO and has realtively little effect on the LCO. Thus, the SR sheet potentially is used for radiation protection in patients, especially on examinations that do not require low contrast resolution.
UTILIZATION OF IMMUNIZATION SERVICES AMONG CHILDREN UNDER FIVE YEARS OF AGE I...AM Publications
Immunization is the key strategy to curb communicable diseases which are the number one killer of children under five. Immunization prevents mortalities of approximating three million children under five annually. This study aimed to assess utilization of immunization services among children under five of age in Kirinyaga County, Kenya.
REPRESENTATION OF THE BLOCK DATA ENCRYPTION ALGORITHM IN AN ANALYTICAL FORM F...AM Publications
The article presents the study of cryptographic transformations of the Kuznyechik algorithm in relation to differential analysis and the translation of their representations into a more convenient form for cryptanalysis. A simplification of the type of transformations of the algorithm to algebraic the form, in which cryptanalysis software will be more effective. Since the description of the algorithm in the analytical form allows for 16 cycles of execution of the shift register with linear feedback, each of which will be carried out 16 operations of multiplication and 15 operations of addition, reduced to 16 multiplying and 15 the operations of addition. The result is an algebraic form of a linear transformation (from a shift register with linear feedback to the multiplication of the matrix in a finite field). In the future, the algebraic type of transformation can be used to effectively carry out differential cryptanalysis.
Optical character recognition (OCR) is process of classification of optical patterns contained in a digital image. The process of OCR Recognition involves several steps including pre-processing, segmentation, feature extraction, classification. Pre-processing is for done the basic operation on input image like noise reduction which remove the noisy signal from image. Segmentation stage for segment the given image into line by line and segment each character from segmented line. Future extraction calculates the characteristics of character. A Radial Basis Function Neural Network (RBFNN) is used to classification contains the database and does the comparison.
Surveillance refers to the task of observing a scene, often for lengthy periods in search of particular objects or particular behaviour. This task has many applications, foremost among them is security (monitoring for undesirable behaviour such as theft or vandalism), but increasing numbers of others in areas such as agriculture also exist. Historically, closed circuit TV (CCTV) surveillance has been mundane and labour Intensive, involving personnel scanning multiple screens, but the advent of reasonably priced fast hardware means that automatic surveillance is becoming a realistic task to attempt in real time. Several attempts at this are underway.
SIMULATION OF ATMOSPHERIC POLLUTANTS DISPERSION IN AN URBAN ENVIRONMENTAM Publications
Interest in air pollution investigation of urban environment due to existence of industrial and commercial activities along with vehicular emission and existence of buildings and streets which setup natural barrier for pollutant dispersion in the urban environment has increased. The air pollution modelling is a multidisciplinary subject when the entire cities are taken under consideration where urban planning and geometries are complex which needs a large software packages to be developed like Operational Street Pollution Model (OSPM), California Line Source model (CALINE series) etc. On overviewing various works it can be summarized that the air pollutant dispersion in urban street canyons and all linked phenomenon such as wind flow, pollutant concentrations, temperature distribution etc. generally depend on wind speed and direction, building heights and density, road width, source and intensity of air pollution, meteorological variables like temperature, humidity etc. A unique and surprising case is observed every time on numerous combinations of these factors. The main aim of this study is to simulate the atmospheric pollutant dispersion for given pollutant like carbon monoxide, sulphur dioxide and nitrogen dioxide and given atmospheric conditions like wind speed and direction. Computational Fluid Dynamics (CFD) simulation for analysing the atmospheric pollutant dispersion is done after natural airflow analysis. Volume rendering is done for variables such as phase 2 volume fraction and velocity with resolution as 250 pixels per inch and transparency as 20%. It can be observed that all the three pollutant namely nitrogen dioxide, sulphur dioxide and carbon monoxide the phase 2 volume fraction changes from 0 to 1. The wind velocity changes from 3.395×10-13 m/s to 1.692×102 m/s. The dispersion of pollutants follow the sequence Sulphur dioxide>Carbon monoxide>Nitrogen dioxide.
PREPARATION AND EVALUATION OF WOOL KERATIN BASED CHITOSAN NANOFIBERS FOR AIR ...AM Publications
In this article, we have extracted keratin from deccani wool waste and prepared the wool keratin based Chitosan nanofibers by electrospinning technique. The prepared nanofibers mat were prepared with different weight percent ratio like 1wt.%, 3wt.% and 5wt.% with respect to polymer i.e Chitosan. The physicochemical and filtration properties of wool keratin based Chitosan nanofibers were studied. Wool keratin based Chitosan nanofibers were characterized by Fourier transform infrared spectroscopy (FTIR), X-ray diffraction (XRD), differential scanning calorimetry (DSC) and scanning electron microscopy (FESEM). The filtration efficiency of keratin Chitosan nanofibers were investigated through DOP test and heavy metal removal capacity of evaluated through Atomic absorption spectroscopy. FTIR results were showed that Keratin gets compatible with Chitosan. XRD patterns revealed keratin was in crystalline nature and increase the crystalline nature of Chitosan nanofibers. FESEM images showed that uniform nanofibers generation with average fiber diameter 80nm. Nanofibers filtration efficiency against a particulate matter in air was obtained more than 99.53% and excellent property of removal of heavy metal.
ANALYSIS ON LOAD BALANCING ALGORITHMS IMPLEMENTATION ON CLOUD COMPUTING ENVIR...AM Publications
Cloud computing means storing and accessing data and programs over the Internet instead of your computer's hard drive. The cloud is just a metaphor for the Internet. The elements involved in cloud computing are clients, data center and distributed server. One of the main problems in cloud computing is load balancing. Balancing the load means to distribute the workload among several nodes evenly so that no single node will be overloaded. Load can be of any type that is it can be CPU load, memory capacity or network load. In this paper we presented an architecture of load balancing and algorithm which will further improve the load balancing problem by minimizing the response time. In this paper, we have proposed the enhanced version of existing regulated load balancing approach for cloud computing by comping the Randomization and greedy load balancing algorithm. To check the performance of proposed approach, we have used the cloud analyst simulator (Cloud Analyst). Through simulation analysis, it has been found that proposed improved version of regulated load balancing approach has shown better performance in terms of cost, response time and data processing time.
A MODEL BASED APPROACH FOR IMPLEMENTING WLAN SECURITY AM Publications
This paper presents various security features and configurations commonly implemented in WLANs and their aggregated security levels and then proposes a model that enables implementation and evaluation of WLAN security
Overview of the fundamental roles in Hydropower generation and the components involved in wider Electrical Engineering.
This paper presents the design and construction of hydroelectric dams from the hydrologist’s survey of the valley before construction, all aspects and involved disciplines, fluid dynamics, structural engineering, generation and mains frequency regulation to the very transmission of power through the network in the United Kingdom.
Author: Robbie Edward Sayers
Collaborators and co editors: Charlie Sims and Connor Healey.
(C) 2024 Robbie E. Sayers
Cosmetic shop management system project report.pdfKamal Acharya
Buying new cosmetic products is difficult. It can even be scary for those who have sensitive skin and are prone to skin trouble. The information needed to alleviate this problem is on the back of each product, but it's thought to interpret those ingredient lists unless you have a background in chemistry.
Instead of buying and hoping for the best, we can use data science to help us predict which products may be good fits for us. It includes various function programs to do the above mentioned tasks.
Data file handling has been effectively used in the program.
The automated cosmetic shop management system should deal with the automation of general workflow and administration process of the shop. The main processes of the system focus on customer's request where the system is able to search the most appropriate products and deliver it to the customers. It should help the employees to quickly identify the list of cosmetic product that have reached the minimum quantity and also keep a track of expired date for each cosmetic product. It should help the employees to find the rack number in which the product is placed.It is also Faster and more efficient way.
About
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
• Remote control: Parallel or serial interface.
• Compatible with MAFI CCR system.
• Compatible with IDM8000 CCR.
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
• Easy in configuration using DIP switches.
Technical Specifications
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
Key Features
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
• Remote control: Parallel or serial interface
• Compatible with MAFI CCR system
• Copatiable with IDM8000 CCR
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
Application
• Remote control: Parallel or serial interface.
• Compatible with MAFI CCR system.
• Compatible with IDM8000 CCR.
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
• Easy in configuration using DIP switches.
Final project report on grocery store management system..pdfKamal Acharya
In today’s fast-changing business environment, it’s extremely important to be able to respond to client needs in the most effective and timely manner. If your customers wish to see your business online and have instant access to your products or services.
Online Grocery Store is an e-commerce website, which retails various grocery products. This project allows viewing various products available enables registered users to purchase desired products instantly using Paytm, UPI payment processor (Instant Pay) and also can place order by using Cash on Delivery (Pay Later) option. This project provides an easy access to Administrators and Managers to view orders placed using Pay Later and Instant Pay options.
In order to develop an e-commerce website, a number of Technologies must be studied and understood. These include multi-tiered architecture, server and client-side scripting techniques, implementation technologies, programming language (such as PHP, HTML, CSS, JavaScript) and MySQL relational databases. This is a project with the objective to develop a basic website where a consumer is provided with a shopping cart website and also to know about the technologies used to develop such a website.
This document will discuss each of the underlying technologies to create and implement an e- commerce website.
Hierarchical Digital Twin of a Naval Power SystemKerry Sado
A hierarchical digital twin of a Naval DC power system has been developed and experimentally verified. Similar to other state-of-the-art digital twins, this technology creates a digital replica of the physical system executed in real-time or faster, which can modify hardware controls. However, its advantage stems from distributing computational efforts by utilizing a hierarchical structure composed of lower-level digital twin blocks and a higher-level system digital twin. Each digital twin block is associated with a physical subsystem of the hardware and communicates with a singular system digital twin, which creates a system-level response. By extracting information from each level of the hierarchy, power system controls of the hardware were reconfigured autonomously. This hierarchical digital twin development offers several advantages over other digital twins, particularly in the field of naval power systems. The hierarchical structure allows for greater computational efficiency and scalability while the ability to autonomously reconfigure hardware controls offers increased flexibility and responsiveness. The hierarchical decomposition and models utilized were well aligned with the physical twin, as indicated by the maximum deviations between the developed digital twin hierarchy and the hardware.
Hybrid optimization of pumped hydro system and solar- Engr. Abdul-Azeez.pdffxintegritypublishin
Advancements in technology unveil a myriad of electrical and electronic breakthroughs geared towards efficiently harnessing limited resources to meet human energy demands. The optimization of hybrid solar PV panels and pumped hydro energy supply systems plays a pivotal role in utilizing natural resources effectively. This initiative not only benefits humanity but also fosters environmental sustainability. The study investigated the design optimization of these hybrid systems, focusing on understanding solar radiation patterns, identifying geographical influences on solar radiation, formulating a mathematical model for system optimization, and determining the optimal configuration of PV panels and pumped hydro storage. Through a comparative analysis approach and eight weeks of data collection, the study addressed key research questions related to solar radiation patterns and optimal system design. The findings highlighted regions with heightened solar radiation levels, showcasing substantial potential for power generation and emphasizing the system's efficiency. Optimizing system design significantly boosted power generation, promoted renewable energy utilization, and enhanced energy storage capacity. The study underscored the benefits of optimizing hybrid solar PV panels and pumped hydro energy supply systems for sustainable energy usage. Optimizing the design of solar PV panels and pumped hydro energy supply systems as examined across diverse climatic conditions in a developing country, not only enhances power generation but also improves the integration of renewable energy sources and boosts energy storage capacities, particularly beneficial for less economically prosperous regions. Additionally, the study provides valuable insights for advancing energy research in economically viable areas. Recommendations included conducting site-specific assessments, utilizing advanced modeling tools, implementing regular maintenance protocols, and enhancing communication among system components.
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...Amil Baba Dawood bangali
Contact with Dawood Bhai Just call on +92322-6382012 and we'll help you. We'll solve all your problems within 12 to 24 hours and with 101% guarantee and with astrology systematic. If you want to take any personal or professional advice then also you can call us on +92322-6382012 , ONLINE LOVE PROBLEM & Other all types of Daily Life Problem's.Then CALL or WHATSAPP us on +92322-6382012 and Get all these problems solutions here by Amil Baba DAWOOD BANGALI
#vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore#blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #blackmagicforlove #blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #Amilbabainuk #amilbabainspain #amilbabaindubai #Amilbabainnorway #amilbabainkrachi #amilbabainlahore #amilbabaingujranwalan #amilbabainislamabad