IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
Number Plate Recognition for Indian Vehiclesmonjuri10
This paper presents Automatic Number Plate
extraction, character segmentation and recognition for
Indian vehicles. In India, number plate models are not
followed strictly. Characters on plate are in different
Indian languages, as well as in English. Due to variations
in the representation of number plates, vehicle number
plate extraction, character segmentation and recognition
are crucial. We present the number plate extraction,
character segmentation and recognition work, with english
characters. Number plate extraction is done using Sobel
filter, morphological operations and connected component
analysis. Character segmentation is done by using
connected component and vertical projection analysis.
Character recognition is carried out using Support Vector
machine (SVM). The segmentation accuracy is 80% and
recognition rate is 79.84 %.
Optical character recognition (ocr) pptDeijee Kalita
The document discusses optical character recognition (OCR), which is the process of converting scanned images of printed or handwritten text into machine-encoded text. It provides a brief history of OCR, explaining some of the early developments. It also outlines the typical steps involved, including pre-processing, character recognition, and post-processing. Examples of applications of OCR technology are given.
Gabor filter is a powerful way to enhance biometric images like fingerprint images in order to extract correct features from these images, Gabor filter used in extracting features directly asin iris images, and sometimes Gabor filter has been used for texture analysis. In fingerprint images The even symmetric Gabor filter is contextual filter or multi-resolution filter will be used to enhance fingerprint imageby filling small gaps (low-pass effect) in the direction of the ridge (black regions) and to increase the discrimination between ridge and valley (black and white regions) in the direction, orthogonal to the ridge, the proposed method in applying Gabor filter on fingerprint images depending on translated fingerprint image into binary image after applying some simple enhancing methods to partially overcome time consuming problem of the Gabor filter.
This document describes a deep learning approach for detecting diabetic retinopathy using OCT images. It discusses the proposed system which will use OCT images and apply classification algorithms to identify the level of infection. The model will be trained on datasets of infected images to accurately detect regions of infection and the condition level. Image processing techniques like median filtering and edge detection will be used along with statistical data extraction and supervised training to identify clusters and classify images. Results will be compared to evaluate the machine learning models. The system aims to automate diabetic retinopathy detection to improve efficiency over conventional methods.
YOLOv5 BASED WEB APPLICATION FOR INDIAN CURRENCY NOTE DETECTIONIRJET Journal
This document presents a web application designed using YOLOv5 and Flask for detecting Indian currency notes to aid visually impaired people. The researchers trained a YOLOv5 model on a dataset of Indian currency note images. They evaluated the model's performance using metrics like precision, recall, and mean average precision (mAP). They then built a web app with front-end components for uploading images and back-end components using Flask and YOLOv5 for detecting notes in images. The app detects notes with over 90% probability and outputs the label and an audio file of the label in English and Hindi. Testing showed the model and app could accurately detect currency notes in single and multiple denomination images on both laptop and
Computer vision is a field of artificial intelligence that uses digital images and deep learning to teach machines to interpret and understand visual input. Early experiments in computer vision in the 1950s used neural networks to detect edges and classify simple shapes, while the 1970s saw the first commercial application in optical character recognition. Today, computer vision can perform tasks like facial recognition, object detection in images and video, and image segmentation, classification, and analysis that rival and exceed human visual abilities. Computer vision works by acquiring an image, processing it through machine learning models, and understanding what is depicted to take appropriate actions.
OCR Presentation (Optical Character Recognition)Neeraj Neupane
Optical Character Recognition (OCR) is a technology that converts non-digital text into editable formats. It works by recognizing printed or written characters using computer vision techniques. The document describes the architecture and objectives of an OCR system, including converting documents to text, speeding up processing, and embedding in applications. It outlines common OCR methods such as grayscaling, binarization, noise removal, sharpening, segmentation, feature extraction, and recognition to identify characters. Diagrams show the system architecture and workflow. Screenshots demonstrate the developed OCR system in use. The conclusion discusses automatic data entry and future areas like recognizing handwriting.
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
Number Plate Recognition for Indian Vehiclesmonjuri10
This paper presents Automatic Number Plate
extraction, character segmentation and recognition for
Indian vehicles. In India, number plate models are not
followed strictly. Characters on plate are in different
Indian languages, as well as in English. Due to variations
in the representation of number plates, vehicle number
plate extraction, character segmentation and recognition
are crucial. We present the number plate extraction,
character segmentation and recognition work, with english
characters. Number plate extraction is done using Sobel
filter, morphological operations and connected component
analysis. Character segmentation is done by using
connected component and vertical projection analysis.
Character recognition is carried out using Support Vector
machine (SVM). The segmentation accuracy is 80% and
recognition rate is 79.84 %.
Optical character recognition (ocr) pptDeijee Kalita
The document discusses optical character recognition (OCR), which is the process of converting scanned images of printed or handwritten text into machine-encoded text. It provides a brief history of OCR, explaining some of the early developments. It also outlines the typical steps involved, including pre-processing, character recognition, and post-processing. Examples of applications of OCR technology are given.
Gabor filter is a powerful way to enhance biometric images like fingerprint images in order to extract correct features from these images, Gabor filter used in extracting features directly asin iris images, and sometimes Gabor filter has been used for texture analysis. In fingerprint images The even symmetric Gabor filter is contextual filter or multi-resolution filter will be used to enhance fingerprint imageby filling small gaps (low-pass effect) in the direction of the ridge (black regions) and to increase the discrimination between ridge and valley (black and white regions) in the direction, orthogonal to the ridge, the proposed method in applying Gabor filter on fingerprint images depending on translated fingerprint image into binary image after applying some simple enhancing methods to partially overcome time consuming problem of the Gabor filter.
This document describes a deep learning approach for detecting diabetic retinopathy using OCT images. It discusses the proposed system which will use OCT images and apply classification algorithms to identify the level of infection. The model will be trained on datasets of infected images to accurately detect regions of infection and the condition level. Image processing techniques like median filtering and edge detection will be used along with statistical data extraction and supervised training to identify clusters and classify images. Results will be compared to evaluate the machine learning models. The system aims to automate diabetic retinopathy detection to improve efficiency over conventional methods.
YOLOv5 BASED WEB APPLICATION FOR INDIAN CURRENCY NOTE DETECTIONIRJET Journal
This document presents a web application designed using YOLOv5 and Flask for detecting Indian currency notes to aid visually impaired people. The researchers trained a YOLOv5 model on a dataset of Indian currency note images. They evaluated the model's performance using metrics like precision, recall, and mean average precision (mAP). They then built a web app with front-end components for uploading images and back-end components using Flask and YOLOv5 for detecting notes in images. The app detects notes with over 90% probability and outputs the label and an audio file of the label in English and Hindi. Testing showed the model and app could accurately detect currency notes in single and multiple denomination images on both laptop and
Computer vision is a field of artificial intelligence that uses digital images and deep learning to teach machines to interpret and understand visual input. Early experiments in computer vision in the 1950s used neural networks to detect edges and classify simple shapes, while the 1970s saw the first commercial application in optical character recognition. Today, computer vision can perform tasks like facial recognition, object detection in images and video, and image segmentation, classification, and analysis that rival and exceed human visual abilities. Computer vision works by acquiring an image, processing it through machine learning models, and understanding what is depicted to take appropriate actions.
OCR Presentation (Optical Character Recognition)Neeraj Neupane
Optical Character Recognition (OCR) is a technology that converts non-digital text into editable formats. It works by recognizing printed or written characters using computer vision techniques. The document describes the architecture and objectives of an OCR system, including converting documents to text, speeding up processing, and embedding in applications. It outlines common OCR methods such as grayscaling, binarization, noise removal, sharpening, segmentation, feature extraction, and recognition to identify characters. Diagrams show the system architecture and workflow. Screenshots demonstrate the developed OCR system in use. The conclusion discusses automatic data entry and future areas like recognizing handwriting.
Graphical Password Authentication using Cued click point technique with zero ...NurrulHafizza
1) The document discusses a student's final year project proposal on implementing a graphical password using cued recall techniques combined with zero knowledge protocol for authentication.
2) Most users tend to use weak passwords that are easy to remember but vulnerable to attacks, while strong passwords are difficult for users to remember. The project aims to design and test a graphical password system that is more secure than text passwords.
3) The methodology will include use case and flowcharts showing the registration and login processes that require users to select and click points on images for authentication. Literature on different graphical password techniques is also reviewed to support the proposed approach.
Diabetes is a disease which is rapidly increasing all over the world. It occurs when pancreas does not produce sufficient insulin, or body can not sufficiently use insulin it produces. Diabetes person has increase blood glucose in the body. One of the major problem diabetic patients suffers from is the Diabetic Retinopathy (DR) and blindness. Since the number of diabetes patients is continuously increasing, it increases the data as well.
Handwriting Recognition Using Deep Learning and Computer VersionNaiyan Noor
This document presents a method for handwriting recognition using deep learning and computer vision. It discusses preprocessing images by removing noise and converting to grayscale. Thresholding is used to separate darker text pixels from lighter background pixels. The image is then segmented into individual lines and words. Python libraries like TensorFlow, Spyder and Jupyter Notebook are used. The goal is to build a system that can recognize text in images and display the text to users. Future work may include recognizing cursive text and additional languages.
Diabetes prediction using machine learningdataalcott
This document discusses a proposed system to classify and predict diabetes using machine learning and deep learning algorithms. The objectives are to classify the PIMA Indian diabetes dataset and design an interactive application where users can input data to get a prediction. The proposed system uses support vector machine (SVM) for machine learning and neural networks for deep learning. It aims to improve accuracy over existing systems by using deep learning techniques. The methodology involves collecting a dataset, preprocessing, splitting for training and testing, applying algorithms, and evaluating results.
Automatic License Plate Recognition using OpenCVEditor IJCATR
Automatic License Plate Recognition system is a real time embedded system which automatically recognizes the license plate of vehicles. There are many applications ranging from complex security systems to common areas and from parking admission to urban traffic control. Automatic license plate recognition (ALPR) has complex characteristics due to diverse effects such as of light and speed. Most of the ALPR systems are built using proprietary tools like Matlab. This paper presents an alternative method of implementing ALPR systems using Free Software including Python and the Open Computer Vision Library.
deep learning applications in medical image analysis brain tumorVenkat Projects
The tremendous success of machine learning algorithms at image recognition tasks in recent years intersects with a time of dramatically increased use of electronic medical records and diagnostic imaging. This review introduces the machine learning algorithms as applied to medical image analysis, focusing on convolutional neural networks, and emphasizing clinical aspects of the _eld. The advantage of machine learning in an era of medical big data is that signi_cant hierarchal relationships within the data can be discovered algorithmically without laborious hand-crafting of features. We cover key research areas and applications of medical image classi_cation, localization, detection, segmentation, and registration. We conclude by discussing research obstacles, emerging trends, and possible future directions.
Computer vision is a field of artificial intelligence that uses digital image processing techniques to analyze visual content and understand scenes. The goal is to extract meaningful information from digital images and emulate some of the capabilities of human vision such as object recognition. Computer vision has applications in security, human behavior analysis, optical character recognition, special effects in movies, face filters in social media, autonomous vehicles, and scene formation. OpenCV is a popular open source library for computer vision that is written in C++ but can be used with other languages like C#, Java, and Python.
This document discusses Internet of Things (IoT) security. It begins by defining IoT and describing common IoT applications in consumer, commercial, industrial, and infrastructure sectors. It then defines IoT security and explains that security is an important area due to the rapid growth of connected devices. The document outlines four layers of IoT security: device, communication, cloud, and lifecycle management. It identifies some of the main security issues like default passwords, unpatched systems, and access to APIs and data. Finally, it discusses best practices for IoT security including authentication, encryption, privacy controls, and firmware updates.
LICENSE NUMBER PLATE RECOGNITION SYSTEM USING ANDROID APPAditya Mishra
The document outlines the development of a number plate recognition system using optical character recognition, including analyzing existing approaches, designing the system architecture, specifying functional and non-functional requirements, and testing the system. It also provides integrated summaries of several research papers on topics like automatic number plate recognition, optical character recognition techniques, and license plate recognition using OCR and template matching.
Presentation on the New Technology based on the recognition of letters that would be available on Soft and Hard copy both and allow all the format in Soft Copy. Optical character Recognition based on the recognition of letters with all the existing languages.
License Plate Recognition System using Python and OpenCVVishal Polley
License plate recognition (LPR) is a type of technology, mainly software, that enables computer systems to read automatically the registration number (license number) of vehicles from digital pictures.
ICCES 2017 - Crowd Density Estimation Method using Regression AnalysisAhmed Gad
The oral presentation of the paper titled "Crowd Density Estimation Method using Multiple Feature Categories and Multiple Regression Models".
This paper was accepted for publication and oral presentation in the 12th IEEE International Conference on Computer Engineering and Systems (ICCES 2017) held from 19 to 20 December 2017 in Cairo, Egypt.
The paper proposed a new method to estimate the number of people within crowded scenes using regression analysis. The two challenges in crowd density estimation using regression analysis are perspective distortion and non-linearity. This paper solves the perspective distortion using perspective normalization which is the best way to deal with that problem based on recent works.
The second challenge is solved by creating a new combination of features collected from multiple already existing categories including segmented region, texture, edge, and keypoints. This paper created a feature vector of length 164.
Five regression models are used which are GPR, RF, RPF, LASSO, and KNN.
Based on the experimental results, our proposed method gives better results than previous works.
----------------------------------
أحمد فوزي جاد Ahmed Fawzy Gad
قسم تكنولوجيا المعلومات Information Technology (IT) Department
كلية الحاسبات والمعلومات Faculty of Computers and Information (FCI)
جامعة المنوفية, مصر Menoufia University, Egypt
Teaching Assistant/Demonstrator
ahmed.fawzy@ci.menofia.edu.eg
---------------------------------
Find me on:
Blog
(Arabic) https://aiage-ar.blogspot.com.eg/
(English) https://aiage.blogspot.com.eg/
YouTube
https://www.youtube.com/AhmedGadFCIT
Google Plus
https://plus.google.com/u/0/+AhmedGadIT
SlideShare
https://www.slideshare.net/AhmedGadFCIT
LinkedIn
https://www.linkedin.com/in/ahmedfgad
reddit
https://www.reddit.com/user/AhmedGadFCIT
ResearchGate
https://www.researchgate.net/profile/Ahmed_Gad13
Academia
https://menofia.academia.edu/Gad
Google Scholar
https://scholar.google.com.eg/citations?user=r07tjocAAAAJ&hl=en
Mendelay
https://www.mendeley.com/profiles/ahmed-gad12
ORCID
https://orcid.org/0000-0003-1978-8574
StackOverFlow
http://stackoverflow.com/users/5426539/ahmed-gad
Twitter
https://twitter.com/ahmedfgad
Facebook
https://www.facebook.com/ahmed.f.gadd
Pinterest
https://www.pinterest.com/ahmedfgad
This is a small presentation on my project , diabetes prediction using R language.The method used is knn(K nearest neighbour). it the basic Machine learning algorithm.
Digital image processing the statistical and structural approaches and the graph based approach for image recognition with algorithms and examples and applications where graph matching is used in pattern recognition.
The document provides an overview of computer vision including:
- It defines computer vision as using observed image data to infer something about the world.
- It briefly discusses the history of computer vision from early projects in 1966 to David Marr establishing the foundations of modern computer vision in the 1970s.
- It lists several related fields that computer vision draws from including artificial intelligence, information engineering, neurobiology, solid-state physics, and signal processing.
- It provides examples of applications of computer vision such as self-driving vehicles, facial recognition, augmented reality, and uses in smartphones, the web, VR/AR, medical imaging, and insurance.
The document discusses the COCO dataset which contains 2.5 million labeled instances in 328,000 images across 91 object categories. It aims to address challenges in detecting non-iconic views of objects, contextual reasoning between objects, and precise 2D localization of objects. The dataset was collected from images featuring objects for ages 4-8 that received over 5,000 votes each. Images underwent category labeling, instance spotting, and instance segmentation annotation processes involving over 70,000 hours of work. Statistics on the dataset are provided and compared to other datasets. The document concludes by outlining plans to add annotations for stuff and human keypoints while removing some images.
This document discusses using deep learning and convolutional neural networks to detect diabetic retinopathy through analyzing fundus images. It proposes a CNN model trained on a public Kaggle dataset to classify images based on the severity of retinopathy. The CNN architecture would automatically diagnose retinopathy without user input. The document outlines modules for an app, including uploading images, displaying results, and providing doctor referrals. It aims to address the growing problem of vision loss from diabetic retinopathy worldwide.
The document discusses machine learning, including an introduction defining it as algorithms and data that allow computers to learn without human intervention. It lists common machine learning algorithms like neural networks and decision trees. The three main types of machine learning are supervised, unsupervised, and reinforcement learning. Examples of machine learning uses include traffic prediction, virtual assistants, and bioinformatics. Popular programming languages for machine learning are Python, Java, C/C++, R, and JavaScript. The key difference between machine learning and artificial intelligence is that machine learning allows machines to learn from data without being programmed. Advantages include speed, accuracy, automation, and security, while disadvantages include difficulty identifying errors and requiring large amounts of data and space.
The document describes how to detect lines in an image using the Hough transform. It explains that the Hough transform represents lines in a polar coordinate system and works by plotting the curves for each edge point and finding the intersections, which indicate collinear points that make up a line. It then outlines the steps to apply this technique: 1) load an image, 2) optionally convert to grayscale and blur, 3) perform edge detection using Canny, and 4) detect lines using Hough transform by finding intersections above a threshold.
Graphical Password Authentication using Cued click point technique with zero ...NurrulHafizza
1) The document discusses a student's final year project proposal on implementing a graphical password using cued recall techniques combined with zero knowledge protocol for authentication.
2) Most users tend to use weak passwords that are easy to remember but vulnerable to attacks, while strong passwords are difficult for users to remember. The project aims to design and test a graphical password system that is more secure than text passwords.
3) The methodology will include use case and flowcharts showing the registration and login processes that require users to select and click points on images for authentication. Literature on different graphical password techniques is also reviewed to support the proposed approach.
Diabetes is a disease which is rapidly increasing all over the world. It occurs when pancreas does not produce sufficient insulin, or body can not sufficiently use insulin it produces. Diabetes person has increase blood glucose in the body. One of the major problem diabetic patients suffers from is the Diabetic Retinopathy (DR) and blindness. Since the number of diabetes patients is continuously increasing, it increases the data as well.
Handwriting Recognition Using Deep Learning and Computer VersionNaiyan Noor
This document presents a method for handwriting recognition using deep learning and computer vision. It discusses preprocessing images by removing noise and converting to grayscale. Thresholding is used to separate darker text pixels from lighter background pixels. The image is then segmented into individual lines and words. Python libraries like TensorFlow, Spyder and Jupyter Notebook are used. The goal is to build a system that can recognize text in images and display the text to users. Future work may include recognizing cursive text and additional languages.
Diabetes prediction using machine learningdataalcott
This document discusses a proposed system to classify and predict diabetes using machine learning and deep learning algorithms. The objectives are to classify the PIMA Indian diabetes dataset and design an interactive application where users can input data to get a prediction. The proposed system uses support vector machine (SVM) for machine learning and neural networks for deep learning. It aims to improve accuracy over existing systems by using deep learning techniques. The methodology involves collecting a dataset, preprocessing, splitting for training and testing, applying algorithms, and evaluating results.
Automatic License Plate Recognition using OpenCVEditor IJCATR
Automatic License Plate Recognition system is a real time embedded system which automatically recognizes the license plate of vehicles. There are many applications ranging from complex security systems to common areas and from parking admission to urban traffic control. Automatic license plate recognition (ALPR) has complex characteristics due to diverse effects such as of light and speed. Most of the ALPR systems are built using proprietary tools like Matlab. This paper presents an alternative method of implementing ALPR systems using Free Software including Python and the Open Computer Vision Library.
deep learning applications in medical image analysis brain tumorVenkat Projects
The tremendous success of machine learning algorithms at image recognition tasks in recent years intersects with a time of dramatically increased use of electronic medical records and diagnostic imaging. This review introduces the machine learning algorithms as applied to medical image analysis, focusing on convolutional neural networks, and emphasizing clinical aspects of the _eld. The advantage of machine learning in an era of medical big data is that signi_cant hierarchal relationships within the data can be discovered algorithmically without laborious hand-crafting of features. We cover key research areas and applications of medical image classi_cation, localization, detection, segmentation, and registration. We conclude by discussing research obstacles, emerging trends, and possible future directions.
Computer vision is a field of artificial intelligence that uses digital image processing techniques to analyze visual content and understand scenes. The goal is to extract meaningful information from digital images and emulate some of the capabilities of human vision such as object recognition. Computer vision has applications in security, human behavior analysis, optical character recognition, special effects in movies, face filters in social media, autonomous vehicles, and scene formation. OpenCV is a popular open source library for computer vision that is written in C++ but can be used with other languages like C#, Java, and Python.
This document discusses Internet of Things (IoT) security. It begins by defining IoT and describing common IoT applications in consumer, commercial, industrial, and infrastructure sectors. It then defines IoT security and explains that security is an important area due to the rapid growth of connected devices. The document outlines four layers of IoT security: device, communication, cloud, and lifecycle management. It identifies some of the main security issues like default passwords, unpatched systems, and access to APIs and data. Finally, it discusses best practices for IoT security including authentication, encryption, privacy controls, and firmware updates.
LICENSE NUMBER PLATE RECOGNITION SYSTEM USING ANDROID APPAditya Mishra
The document outlines the development of a number plate recognition system using optical character recognition, including analyzing existing approaches, designing the system architecture, specifying functional and non-functional requirements, and testing the system. It also provides integrated summaries of several research papers on topics like automatic number plate recognition, optical character recognition techniques, and license plate recognition using OCR and template matching.
Presentation on the New Technology based on the recognition of letters that would be available on Soft and Hard copy both and allow all the format in Soft Copy. Optical character Recognition based on the recognition of letters with all the existing languages.
License Plate Recognition System using Python and OpenCVVishal Polley
License plate recognition (LPR) is a type of technology, mainly software, that enables computer systems to read automatically the registration number (license number) of vehicles from digital pictures.
ICCES 2017 - Crowd Density Estimation Method using Regression AnalysisAhmed Gad
The oral presentation of the paper titled "Crowd Density Estimation Method using Multiple Feature Categories and Multiple Regression Models".
This paper was accepted for publication and oral presentation in the 12th IEEE International Conference on Computer Engineering and Systems (ICCES 2017) held from 19 to 20 December 2017 in Cairo, Egypt.
The paper proposed a new method to estimate the number of people within crowded scenes using regression analysis. The two challenges in crowd density estimation using regression analysis are perspective distortion and non-linearity. This paper solves the perspective distortion using perspective normalization which is the best way to deal with that problem based on recent works.
The second challenge is solved by creating a new combination of features collected from multiple already existing categories including segmented region, texture, edge, and keypoints. This paper created a feature vector of length 164.
Five regression models are used which are GPR, RF, RPF, LASSO, and KNN.
Based on the experimental results, our proposed method gives better results than previous works.
----------------------------------
أحمد فوزي جاد Ahmed Fawzy Gad
قسم تكنولوجيا المعلومات Information Technology (IT) Department
كلية الحاسبات والمعلومات Faculty of Computers and Information (FCI)
جامعة المنوفية, مصر Menoufia University, Egypt
Teaching Assistant/Demonstrator
ahmed.fawzy@ci.menofia.edu.eg
---------------------------------
Find me on:
Blog
(Arabic) https://aiage-ar.blogspot.com.eg/
(English) https://aiage.blogspot.com.eg/
YouTube
https://www.youtube.com/AhmedGadFCIT
Google Plus
https://plus.google.com/u/0/+AhmedGadIT
SlideShare
https://www.slideshare.net/AhmedGadFCIT
LinkedIn
https://www.linkedin.com/in/ahmedfgad
reddit
https://www.reddit.com/user/AhmedGadFCIT
ResearchGate
https://www.researchgate.net/profile/Ahmed_Gad13
Academia
https://menofia.academia.edu/Gad
Google Scholar
https://scholar.google.com.eg/citations?user=r07tjocAAAAJ&hl=en
Mendelay
https://www.mendeley.com/profiles/ahmed-gad12
ORCID
https://orcid.org/0000-0003-1978-8574
StackOverFlow
http://stackoverflow.com/users/5426539/ahmed-gad
Twitter
https://twitter.com/ahmedfgad
Facebook
https://www.facebook.com/ahmed.f.gadd
Pinterest
https://www.pinterest.com/ahmedfgad
This is a small presentation on my project , diabetes prediction using R language.The method used is knn(K nearest neighbour). it the basic Machine learning algorithm.
Digital image processing the statistical and structural approaches and the graph based approach for image recognition with algorithms and examples and applications where graph matching is used in pattern recognition.
The document provides an overview of computer vision including:
- It defines computer vision as using observed image data to infer something about the world.
- It briefly discusses the history of computer vision from early projects in 1966 to David Marr establishing the foundations of modern computer vision in the 1970s.
- It lists several related fields that computer vision draws from including artificial intelligence, information engineering, neurobiology, solid-state physics, and signal processing.
- It provides examples of applications of computer vision such as self-driving vehicles, facial recognition, augmented reality, and uses in smartphones, the web, VR/AR, medical imaging, and insurance.
The document discusses the COCO dataset which contains 2.5 million labeled instances in 328,000 images across 91 object categories. It aims to address challenges in detecting non-iconic views of objects, contextual reasoning between objects, and precise 2D localization of objects. The dataset was collected from images featuring objects for ages 4-8 that received over 5,000 votes each. Images underwent category labeling, instance spotting, and instance segmentation annotation processes involving over 70,000 hours of work. Statistics on the dataset are provided and compared to other datasets. The document concludes by outlining plans to add annotations for stuff and human keypoints while removing some images.
This document discusses using deep learning and convolutional neural networks to detect diabetic retinopathy through analyzing fundus images. It proposes a CNN model trained on a public Kaggle dataset to classify images based on the severity of retinopathy. The CNN architecture would automatically diagnose retinopathy without user input. The document outlines modules for an app, including uploading images, displaying results, and providing doctor referrals. It aims to address the growing problem of vision loss from diabetic retinopathy worldwide.
The document discusses machine learning, including an introduction defining it as algorithms and data that allow computers to learn without human intervention. It lists common machine learning algorithms like neural networks and decision trees. The three main types of machine learning are supervised, unsupervised, and reinforcement learning. Examples of machine learning uses include traffic prediction, virtual assistants, and bioinformatics. Popular programming languages for machine learning are Python, Java, C/C++, R, and JavaScript. The key difference between machine learning and artificial intelligence is that machine learning allows machines to learn from data without being programmed. Advantages include speed, accuracy, automation, and security, while disadvantages include difficulty identifying errors and requiring large amounts of data and space.
The document describes how to detect lines in an image using the Hough transform. It explains that the Hough transform represents lines in a polar coordinate system and works by plotting the curves for each edge point and finding the intersections, which indicate collinear points that make up a line. It then outlines the steps to apply this technique: 1) load an image, 2) optionally convert to grayscale and blur, 3) perform edge detection using Canny, and 4) detect lines using Hough transform by finding intersections above a threshold.
Komputerin Aparat Təminatı
1 - Xarici və daxili yaddaş qurğuları
2 - Mikroprosessorlar
3 - Müasir anakartlar və onların giriş\çıxış portları
4 - Monitorların xarakteristikaları və iş prinsipi
5 - Çap qurğuları və skanerlərin iş prinsipi
6 - Registr yaddaşı və keş yaddaşın iş prinsipləri
7 - Statik və dinamik yaddaş qurğuları
8 - Hesablama sistemlərinin interfeysləri
9 - Çoxprosessorlu (çoxnüvəli) sistemlər
10 - Müasir hesablama sistemlərinin arkitekturası və strukturu
Ходжали́нская резня́ — массовое убийство жителей азербайджанского города Ходжалы армянскими вооружёнными формированиями, которое в ряде источников характеризуется как самое крупное и жестокое кровопролитие за время Карабахской войны. В азербайджанских источниках эти события именуются Ходжалинской трагедией (азерб. Xocalı faciəsi).
Чёрный январь (азерб. Qara Yanvar) или также Кровавый январь (азерб. Qanlı Yanvar) — подавление политической оппозиции подразделениями Советской Армии в ночь на 20 января 1990 года в столице Азербайджанской ССР — городе Баку, закончившееся гибелью более сотни мирных жителей, в основном азербайджанцев. Акциям протеста азербайджанской оппозиции предшествовало насилие против армянского населения Баку.
2. KVANTLAŞDIRMA (GÖRÜNTÜ EMALI)
• Təsvir emalında adını çəkdiyimiz kvantizasiya, bir neçə müxtəlif
dəyərləri vahid bir kvant dəyərinə sıxaraq əldə edilən itkili bir sıxılma
üsuludur. Şəkilin (verilən) içərisində diskret işarələrin sayı azaldıqda,
şəkil daha sıxılabilən olur. Məsələn, rəqəmsal bir görüntünü göstərmək
üçün tələb olunan rənglərin sayının azaldılması, fayl ölçüsünü
azaltmağa imkan verir.
3. RƏNG KVANTLAŞDIRMASI
• Kompüter qrafikasında rəng kvantlanması və ya rəngli görüntü
kvantlaşdırılması rəng sahəsində tətbiq olunan anlayışdır;
• yeni görüntünün orijinal görüntüyə mümkün qədər vizual bənzər olması
niyyəti ilə bir görüntüdə istifadə olunan fərqli rənglərin sayını azaldan bir
prosesdir.
• Bitmaplarda rəng kvantlaşdırmasını həyata keçirən kompüter alqoritmləri
1970-ci illərdən bəri öyrənilir.
• Rəng kvantasiyası, adətən yaddaş məhdudiyyətlərinə görə məhdud sayda
rəng nümayiş etdirə bilən cihazlarda çox rəngli görüntülərin görünməsi üçün
vacibdir və müəyyən növ şəkillərin səmərəli sıxılmasını təmin edir.
4. RƏNG KVANTLAŞDIRMASI
ALQORİTMLƏRİ
• 1979-cu ildə Paul Heckbert tərəfindən
icad edilən rəng miqdarına görə ən
populyar alqoritm median kəsilmiş
alqoritmdir. Bu sxem üzrə bir çox versiya
istifadə olunur. Bu vaxtdan əvvəl rənglərin
ölçülməsi ən çox populyasiya alqoritmi və
ya populyasiya üsulu ilə aparılmışdır, bu
da bərabər ölçülü silsilələrin histoqramını
qurur və ən çox nöqtəni əhatə edən
silsilələrə rəng verir. Daha müasir bir
populyar metod, əvvəlcə Gervautz və
Purgathofer tərəfindən səkkizbucaq
istifadə edərək hazırlanan və Xerox PARC
tədqiqatçısı Dan Bloomberg tərəfindən
təkmilləşdirilmiş alqoritmin istifadəsi
çoxluq təşkil edir.
5. ŞƏKİL SEQMENTASİYASİ
• Kompüter görmə qabiliyyətində (computer vision), görüntü
seqmentasiyası rəqəmsal bir görüntünün çox seqmentə bölünməsi
prosesidir (görüntü obyektləri olaraq da tanınan piksel dəstləri).
Segmentasiyanin məqsədi görüntünün təsvirini kompüter üçün anlamlı
və təhlil etmək üçün daha asan bir verilən formasına çevirmək və / və ya
dəyişdirməkdir. Şəkil seqmentasiyası adətən şəkillərdəki obyektlərin və
sərhədlərin (xətlər, əyrilər və s.) müəyyən edilməsi üçün istifadə olunur.
Daha doğrusu, görüntü seqmentləşdirilməsi, eyni etiketli piksellərin
müəyyən xüsusiyyətləri bölüşdürdüyü bir görüntüdə hər pikselə etiket
verilməsi prosesidir.
6. SEQMENTASİYANİN TƏTBİQ SAHƏLƏRİ
• Məzmuna əsaslanan görüntü alınması
• Maşın görməsi
• Kompüter tomoqrafiyasından və maqnetik
rezonans görüntüsündən həcmli görüntülər
daxil olmaqla tibbi görüntüləmə.
• Şişləri və digər patologiyaları tapmaq
• Toxuma həcmini ölçmək
• Diaqnoz, anatomik quruluşun öyrənilməsi
• Cərrahiyyə planlaşdırılması
• Virtual cərrahiyyə simulyasiyası
• Əməliyyatdaxili naviqasiya
• Obyekt aşkarlanması
• Piyada aşkarlanması
• Üz aşkarlanması
• Əyləc işığının aşkarlanması
• Peyk şəkillərində obyektləri (yollar, meşələr,
bitkilər və s.) Yerləşdirin
• Tanınma prosesləri
• Üzün tanınması
• Barmaq izinin tanınması
• İrsin tanınması
• Trafikə nəzarət sistemləri
• Video nəzarət
• Video obyektin birgə seqmentləşdirilməsi və
fəaliyyət lokalizasiyası
7. THRESHOLDING
(EŞİK - ПОРОГОВ)
• Şəkil seqmentləşdirilməsinin ən sadə üsulu eşik metodu adlanır. Bu üsul,
boz miqyaslı bir görüntüyü ikili (binar) görüntüyə çevirmək üçün eşik
dəyərinə əsaslanır.
• Bu metodun açarı eşik əmsalını seçməkdir. Maksimum entropiya
metodu, balanslaşdırılmış histoqram həddi, Otsu metodu və
klasterləşdirmə də daxil olmaqla bir neçə populyar metod istifadə
olunur.
• Son zamanlarda, kompüter tomoqrafiya (CT) şəkillərini thresholding
etmək üçün metodlar hazırlanmışdır. Əsas ideya odur ki, Otsu
metodundan fərqli olaraq, eşiklər (yenidən qurulmuş) görüntü yerinə
radioqraflardan alınır.
8. MATLABDA TƏSVİRİN
KVANTLAŞDIRILMASI
• İmquantize funksiyası – Verilmiş miqdar səviyyələri və çıxış
dəyərlərindən istifadə edərək görüntünü kvantlaşdırır
• quant_A = imquantize(A,levels) A görüntüsünü levels adnlanan
parametrdə verilmiş N tutumlu massivdəki kvantlaşdırma dəyərlərini
istifadə edərək kvantlaşdırır. Alınan quant_A şəkli A şəkli ilə eyni ölçüyə
malikdir və I və N+I aralığında N+I sayda diskret qiymətlər saxlayır.
9. NÜMÜNƏ: ŞƏKLI IKI ƏRƏFƏDƏN ISTIFADƏ
EDƏRƏK ÜÇ SEQMENTƏ BÖLMƏK
• Şəkli yaddaşdan oxuyuruq və ekrana çıxarırıq
I = rgb2gray(imread(“sekil.png”));
imshow(I)
axis off
title(“Original Sekil”)
10. İki hədd (eşik) səviyyəsi hesablayırıq.
thresh = multithresh(I,2);
imquantize istifadə edərək görüntüyü üç səviyyəyə bölürük.
seg_I = imquantize(I,thresh);
label2rgb istifadə edərək seqmentli görüntüyü rəngli
görüntüyə çeviririk və ekranda göstəririk.
RGB = label2rgb(seg_I); figure;
imshow(RGB)
axis off
title('RGB Seqmentli Görüntü')
11. BOZ RƏNG ÇALARLARI SƏVIYYƏSINI
256-DAN 8-Ə ENDIRMƏK
• Bir görüntüdə diskret səviyyələrin sayını 256-
dan 8-ə qədər azaldırıq. Bu nümunədə səkkiz
çıxış səviyyəsinin hər birinə dəyərlər təyin
etmək üçün iki fərqli metoddan istifadə edirik.
I = imread(qepik.png’);
imshow(I)
axis off
title(‘Qepikler')
12. • Multithresh-dən yeddi eşik əldə
edərək görüntüyü səkkiz
səviyyəyə bölün.
thresh = multithresh(I,7);
valuesMax vektorunu elə qururuq ki, hər kvantlaşdırma
intervalında maksimum dəyər çıxış görüntüsünün səkkiz
səviyyəsinə də təyin edilsin.
valuesMax = 1x8 uint8 sətir vektoru
65 88 119 149 169 189 215 255
[quant8_I_max, index] = imquantize(I,thresh,valuesMax);
valuesMin = [min(I(:)) thresh]
valuesMin = 1x8 uint8 row vector 23 65 88 119 149 169 189
215
quant8_I_min = valuesMin(index);