Performance analysis of automated brain tumor detection from MR imaging and CT scan using basic image processing techniques based on various hard and soft computing has been performed in our work. Moreover, we applied six traditional classifiers to detect brain tumor in the images. Then we applied CNN for brain tumor detection to include deep learning method in our work. We compared the result of the traditional one having the best accuracy (SVM) with the result of CNN. Furthermore, our work presents a generic method of tumor detection and extraction of its various features.
details about brain tumor
literature survey on many reference papers related to brain tumor detection using various techniques
our proposed novel methodology for brain tumor detection
Talk @ ACM SF Bayarea Chapter on Deep Learning for medical imaging space.
The talk covers use cases, special challenges and solutions for Deep Learning for Medical Image Analysis using Tensorflow+Keras. You will learn about:
- Use cases for Deep Learning in Medical Image Analysis
- Different DNN architectures used for Medical Image Analysis
- Special purpose compute / accelerators for Deep Learning (in the Cloud / On-prem)
- How to parallelize your models for faster training of models and serving for inferenceing.
- Optimization techniques to get the best performance from your cluster (like Kubernetes/ Apache Mesos / Spark)
- How to build an efficient Data Pipeline for Medical Image Analysis using Deep Learning
- Resources to jump start your journey - like public data sets, common models used in Medical Image Analysis
BRAIN TUMOR MRI IMAGE SEGMENTATION AND DETECTION IN IMAGE PROCESSINGDharshika Shreeganesh
Image processing is an active research area in which medical image processing is a highly challenging field. Medical imaging
techniques are used to image the inner portions of the human body for medical diagnosis. Brain tumor is a serious life altering
disease condition. Image segmentation plays a significant role in image processing as it helps in the extraction of suspicious regions
from the medical images. In this paper we have proposed segmentation of brain MRI image using K-means clustering algorithm
followed by morphological filtering which avoids the misclustered regions that can inevitably be formed after segmentation of the brain MRI image for detection of tumor location.
Performance analysis of automated brain tumor detection from MR imaging and CT scan using basic image processing techniques based on various hard and soft computing has been performed in our work. Moreover, we applied six traditional classifiers to detect brain tumor in the images. Then we applied CNN for brain tumor detection to include deep learning method in our work. We compared the result of the traditional one having the best accuracy (SVM) with the result of CNN. Furthermore, our work presents a generic method of tumor detection and extraction of its various features.
details about brain tumor
literature survey on many reference papers related to brain tumor detection using various techniques
our proposed novel methodology for brain tumor detection
Talk @ ACM SF Bayarea Chapter on Deep Learning for medical imaging space.
The talk covers use cases, special challenges and solutions for Deep Learning for Medical Image Analysis using Tensorflow+Keras. You will learn about:
- Use cases for Deep Learning in Medical Image Analysis
- Different DNN architectures used for Medical Image Analysis
- Special purpose compute / accelerators for Deep Learning (in the Cloud / On-prem)
- How to parallelize your models for faster training of models and serving for inferenceing.
- Optimization techniques to get the best performance from your cluster (like Kubernetes/ Apache Mesos / Spark)
- How to build an efficient Data Pipeline for Medical Image Analysis using Deep Learning
- Resources to jump start your journey - like public data sets, common models used in Medical Image Analysis
BRAIN TUMOR MRI IMAGE SEGMENTATION AND DETECTION IN IMAGE PROCESSINGDharshika Shreeganesh
Image processing is an active research area in which medical image processing is a highly challenging field. Medical imaging
techniques are used to image the inner portions of the human body for medical diagnosis. Brain tumor is a serious life altering
disease condition. Image segmentation plays a significant role in image processing as it helps in the extraction of suspicious regions
from the medical images. In this paper we have proposed segmentation of brain MRI image using K-means clustering algorithm
followed by morphological filtering which avoids the misclustered regions that can inevitably be formed after segmentation of the brain MRI image for detection of tumor location.
Brain tumor detection with the mri image and 54900 image Performance analysis of automated brain tumor detection from MR imaging and CT scan using basic image processing techniques based on various hard and soft .This repo is of segmentation and morphological operations which are the basic concepts of image processing. Detection and extraction of tumor from MRI scan images of the brain is done using python.
Adaptive K-Means Clustering Algorithm for MR Breast Image Segmentation
3D Brain Tumor Segmentation Scheme using K-mean Clustering and Connected Component Labeling Algorithms
Volume Identification and Estimation of MRI Brain Tumor
MRI Breast cancer diagnosis hybrid approach using adaptive Ant-based segmentation and Multilayer Perceptron NN classifier
Intro to Deep Learning for Medical Image Analysis, with Dan Lee from Dentuit AISeth Grimes
Dan Lee from Dentuit AI presented an Intro to Deep Learning for Medical Image Analysis at the Maryland AI meetup (https://www.meetup.com/Maryland-AI), May 27, 2020. Visit https://www.youtube.com/watch?v=xl8i7CGDQi0 for video.
Brain Tumor Detection Using Image ProcessingSinbad Konick
The process of brain tumor detection using various filters and finding out the best possible approach. Processing the image and using other filters and find out the result.
image classification is a common problem in Artificial Intelligence , we used CIFR10 data set and tried a lot of methods to reach a high test accuracy like neural networks and Transfer learning techniques .
you can view the source code and the papers we read on github : https://github.com/Asma-Hawari/Machine-Learning-Project-
Brain Tumor Segmentation using Enhanced U-Net Model with Empirical AnalysisMD Abdullah Al Nasim
Cancer of the brain is deadly and requires careful surgical segmentation. The brain tumors were segmented using U-Net using a Convolutional Neural Network (CNN). When looking for overlaps of necrotic, edematous, growing, and healthy tissue, it might be hard to get relevant information from the images. The 2D U-Net network was improved and trained with the BraTS datasets to find these four areas. U-Net can set up many encoder and decoder routes that can be used to get information from images that can be used in different ways. To reduce computational time, we use image segmentation to exclude insignificant background details. Experiments on the BraTS datasets show that our proposed model for segmenting brain tumors from MRI (MRI) works well. In this study, we demonstrate that the BraTS datasets for 2017, 2018, 2019, and 2020 do not significantly differ from the BraTS 2019 dataset's attained dice scores of 0.8717 (necrotic), 0.9506 (edema), and 0.9427 (enhancing).
Introduction to image processing and pattern recognitionSaibee Alam
this power point presentation provides a brief introduction to image processing and pattern recognition and its related research papers including conclusion
The next phase of Smart Network Convergence could be putting Deep Learning systems on the Internet. Deep Learning and Blockchain Technology might be combined in the smart networks of the future for automated identification (deep learning) and automated transaction (blockchain). Large scale future-class problems might be addressed with Blockchain Deep Learning nets as an advanced computational infrastructure, challenges such as million-member genome banks, energy storage markets, global financial risk assessment, real-time voting, and asteroid mining.
Blockchain Deep Learning nets and Smart Networks more generally are computing networks with intelligence built in such that identification and transfer is performed by the network itself through sophisticated protocols that automatically identify (deep learning), and validate, confirm, and route transactions (blockchain) within the network.
In this project, we propose methods for semantic segmentation with the deep learning state-of-the-art models. Moreover,
we want to filterize the segmentation to the specific object in specific application. Instead of concentrating on unnecessary objects we
can focus on special ones and make it more specialize and effecient for special purposes. Furtheromore, In this project, we leverage
models that are suitable for face segmentation. The models that are used in this project are Mask-RCNN and DeepLabv3. The
experimental results clearly indicate that how illustrated approach are efficient and robust in the segmentation task to the previous work
in the field of segmentation. These models are reached to 74.4 and 86.6 precision of Mean of Intersection over Union. The visual
Results of the models are shown in Appendix part.
A fast-paced introduction to Deep Learning concepts, such as activation functions, cost functions, back propagation, and then a quick dive into CNNs. Basic knowledge of vectors, matrices, and derivatives is helpful in order to derive the maximum benefit from this session.
Decomposing image generation into layout priction and conditional synthesisNaeem Shehzad
in this presentation you can learn how to decompose an image into layout and find the predictions. In this presentation , I mention all the data in very convenient way , I hope you can take it easy.
Thank you.
The presentation is made on CNN's which is explained using the image classification problem, the presentation was prepared in perspective of understanding computer vision and its applications. I tried to explain the CNN in the most simple way possible as for my understanding. This presentation helps the beginners of CNN to have a brief idea about the architecture and different layers in the architecture of CNN with the example. Please do refer the references in the last slide for a better idea on working of CNN. In this presentation, I have also discussed the different types of CNN(not all) and the applications of Computer Vision.
Brain tumor detection with the mri image and 54900 image Performance analysis of automated brain tumor detection from MR imaging and CT scan using basic image processing techniques based on various hard and soft .This repo is of segmentation and morphological operations which are the basic concepts of image processing. Detection and extraction of tumor from MRI scan images of the brain is done using python.
Adaptive K-Means Clustering Algorithm for MR Breast Image Segmentation
3D Brain Tumor Segmentation Scheme using K-mean Clustering and Connected Component Labeling Algorithms
Volume Identification and Estimation of MRI Brain Tumor
MRI Breast cancer diagnosis hybrid approach using adaptive Ant-based segmentation and Multilayer Perceptron NN classifier
Intro to Deep Learning for Medical Image Analysis, with Dan Lee from Dentuit AISeth Grimes
Dan Lee from Dentuit AI presented an Intro to Deep Learning for Medical Image Analysis at the Maryland AI meetup (https://www.meetup.com/Maryland-AI), May 27, 2020. Visit https://www.youtube.com/watch?v=xl8i7CGDQi0 for video.
Brain Tumor Detection Using Image ProcessingSinbad Konick
The process of brain tumor detection using various filters and finding out the best possible approach. Processing the image and using other filters and find out the result.
image classification is a common problem in Artificial Intelligence , we used CIFR10 data set and tried a lot of methods to reach a high test accuracy like neural networks and Transfer learning techniques .
you can view the source code and the papers we read on github : https://github.com/Asma-Hawari/Machine-Learning-Project-
Brain Tumor Segmentation using Enhanced U-Net Model with Empirical AnalysisMD Abdullah Al Nasim
Cancer of the brain is deadly and requires careful surgical segmentation. The brain tumors were segmented using U-Net using a Convolutional Neural Network (CNN). When looking for overlaps of necrotic, edematous, growing, and healthy tissue, it might be hard to get relevant information from the images. The 2D U-Net network was improved and trained with the BraTS datasets to find these four areas. U-Net can set up many encoder and decoder routes that can be used to get information from images that can be used in different ways. To reduce computational time, we use image segmentation to exclude insignificant background details. Experiments on the BraTS datasets show that our proposed model for segmenting brain tumors from MRI (MRI) works well. In this study, we demonstrate that the BraTS datasets for 2017, 2018, 2019, and 2020 do not significantly differ from the BraTS 2019 dataset's attained dice scores of 0.8717 (necrotic), 0.9506 (edema), and 0.9427 (enhancing).
Introduction to image processing and pattern recognitionSaibee Alam
this power point presentation provides a brief introduction to image processing and pattern recognition and its related research papers including conclusion
The next phase of Smart Network Convergence could be putting Deep Learning systems on the Internet. Deep Learning and Blockchain Technology might be combined in the smart networks of the future for automated identification (deep learning) and automated transaction (blockchain). Large scale future-class problems might be addressed with Blockchain Deep Learning nets as an advanced computational infrastructure, challenges such as million-member genome banks, energy storage markets, global financial risk assessment, real-time voting, and asteroid mining.
Blockchain Deep Learning nets and Smart Networks more generally are computing networks with intelligence built in such that identification and transfer is performed by the network itself through sophisticated protocols that automatically identify (deep learning), and validate, confirm, and route transactions (blockchain) within the network.
In this project, we propose methods for semantic segmentation with the deep learning state-of-the-art models. Moreover,
we want to filterize the segmentation to the specific object in specific application. Instead of concentrating on unnecessary objects we
can focus on special ones and make it more specialize and effecient for special purposes. Furtheromore, In this project, we leverage
models that are suitable for face segmentation. The models that are used in this project are Mask-RCNN and DeepLabv3. The
experimental results clearly indicate that how illustrated approach are efficient and robust in the segmentation task to the previous work
in the field of segmentation. These models are reached to 74.4 and 86.6 precision of Mean of Intersection over Union. The visual
Results of the models are shown in Appendix part.
A fast-paced introduction to Deep Learning concepts, such as activation functions, cost functions, back propagation, and then a quick dive into CNNs. Basic knowledge of vectors, matrices, and derivatives is helpful in order to derive the maximum benefit from this session.
Decomposing image generation into layout priction and conditional synthesisNaeem Shehzad
in this presentation you can learn how to decompose an image into layout and find the predictions. In this presentation , I mention all the data in very convenient way , I hope you can take it easy.
Thank you.
The presentation is made on CNN's which is explained using the image classification problem, the presentation was prepared in perspective of understanding computer vision and its applications. I tried to explain the CNN in the most simple way possible as for my understanding. This presentation helps the beginners of CNN to have a brief idea about the architecture and different layers in the architecture of CNN with the example. Please do refer the references in the last slide for a better idea on working of CNN. In this presentation, I have also discussed the different types of CNN(not all) and the applications of Computer Vision.
• They are relatively expensive to produce compared to other battery technologies.
• They have a limited lifespan, typically around 2-3 years, and their capacity gradually decreases over time.
• Lithium-ion batteries can be sensitive to high temperatures and overcharging, which can cause them to overheat, swell, or catch fire.
• They require special care and handling to prevent damage, such as avoiding deep discharge and extreme temperatures.
• The production of lithium-ion batteries relies on the mining and processing of materials such as lithium, cobalt, and nickel, which can have significant environmental impacts.
• Recycling of lithium-ion batteries can be challenging and costly, leading to concerns about e-waste and sustainability.
• They are relatively expensive to produce compared to other battery technologies.
• They have a limited lifespan, typically around 2-3 years, and their capacity gradually decreases over time.
• Lithium-ion batteries can be sensitive to high temperatures and overcharging, which can cause them to overheat, swell, or catch fire.
• They require special care and handling to prevent damage, such as avoiding deep discharge and extreme temperatures.
• The production of lithium-ion batteries relies on the mining and processing of materials such as lithium, cobalt, and nickel, which can have significant environmental impacts.
• Recycling of lithium-ion batteries can be challenging and costly, leading to concerns about e-waste and sustainability.
• They are relatively expensive to produce compared to other battery technologies.
• They have a limited lifespan, typically around 2-3 years, and their capacity gradually decreases over time.
• Lithium-ion batteries can be sensitive to high temperatures and overcharging, which can cause them to overheat, swell, or catch fire.
• They require special care and handling to prevent damage, such as avoiding deep discharge and extreme temperatures.
• The production of lithium-ion batteries relies on the mining and processing of materials such as lithium, cobalt, and nickel, which can have significant environmental impacts.
• Recycling of lithium-ion batteries can be challenging and costly, leading to concerns about e-waste and sustainability.
Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,Statistical process control in machining,
https://telecombcn-dl.github.io/2018-dlcv/
Deep learning technologies are at the core of the current revolution in artificial intelligence for multimedia data analysis. The convergence of large-scale annotated datasets and affordable GPU hardware has allowed the training of neural networks for data analysis tasks which were previously addressed with hand-crafted features. Architectures such as convolutional neural networks, recurrent neural networks and Q-nets for reinforcement learning have shaped a brand new scenario in signal processing. This course will cover the basic principles and applications of deep learning to computer vision problems, such as image classification, object detection or image captioning.
Finding the best solution for Image ProcessingTech Triveni
What is beyond using Tensorflow, GPU or TPU to process images seamlessly? Do we have a silver bullet for image processing? Over the years, image processing has picked up a different level of attraction. Everyone can think about its ease of usability because it has become a reality now. We have started seeing how Residual Neural Network architecture is being used for different cases and not only that, how Residual Neural network is being tweaked to solve different problems. Along with tweaking the ResNet, preprocessing is also being improved to support different architecture for this matter.
Everyone has almost become cyborg already with mobile phones in our hands and apparently until human beings bring the AI/ML to the phones completely they are not taking any rest. We are going to see the development of different architecture and algorithms around running AI/ML on low configuration devices.
In this session, we are going to talk about different research papers submitted for these matters and some implementations for the same as well.
PR-330: How To Train Your ViT? Data, Augmentation, and Regularization in Visi...Jinwon Lee
안녕하세요 TensorFlow Korea 논문 읽기 모임 PR-12의 330번째 논문 리뷰입니다.
오늘은 무려 5만개의 학습된 ViT model을 제공하는 구글스러운 논문을 리뷰해보았습니다. ViT가 CNN을 조금씩 대체해가고 있는데요, ViT는 CNN과 달리 inductive bias가 적은 관계로
좋은 성능을 위해서는 굉장히 많은 data가 필요하거나, augmentation과 regularization을 많이 써줘야 합니다.
그런데 이렇게 다양한 경우 즉 다양한 data, 다양한 model size, 다양한 augmentation 방법, 다양한 regularization, 다양한 data size 등등에 따른 ViT의 성능과 속도 등의 비교 분석 실험이 지금까지는 없었죠.
이 논문에서는 그 어려운 걸(?) 해냈습니다. 그리고 수많은 ViT를 이용해 실험을 하면서 몇가지 중요한 finding들을 찾았습니다.
요약하면 다음과 같습니다.
1. augmentation과 regularization을 잘 쓰면 1/10의 data로도 전체 data 다 쓴거랑 대부분 비슷한 성능을 낼 수 있다. 그런데 항상 그런건 아니다.
반대로 말하면 data가 10배 있으면 augmentation이나 regularization안 쓰고도 좋은 성능을 낼 수 있다.
2. downstream task 학습할 때 scratch부터 학습하는거랑 large dataset으로 pre-trained한 걸 이용해서 transfer learning하는 건 후자가 좋다.
3. transfer learning 할 때도 pre-trained model 중에 data 많이 써서 학습한게 더 좋다.
4. augmentation/regularization은 data가 많으면 별 도움이 안되고 둘 중에는 augmenation이 더 좋다.
5. pre-trained model이 많을 때 model을 고르는 방법은 그냥 upstream에서 제일 잘됐던 걸 고르면 얼추 잘된다.
6. 속도를 빠르게 하고 싶을 때는 model을 작은거 쓰지말고 patch size를 키워라. 그래야 성능이 별로 안떨어진다.
입니다.
흥미로운 결과들이 많으니 자세한 내용은 아래 영상을 참고해주세요!
감사합니다!
영상링크: https://youtu.be/A3RrAIx-KCc
논문링크: https://arxiv.org/abs/2106.10270
Semantic Segmentation on Satellite ImageryRAHUL BHOJWANI
This is an Image Semantic Segmentation project targeted on Satellite Imagery. The goal was to detect the pixel-wise segmentation map for various objects in Satellite Imagery including buildings, water bodies, roads etc. The data for this was taken from the Kaggle competition <https://www.kaggle.com/c/dstl-satellite-imagery-feature-detection>.
We implemented FCN, U-Net and Segnet Deep learning architectures for this task.
How to Split Bills in the Odoo 17 POS ModuleCeline George
Bills have a main role in point of sale procedure. It will help to track sales, handling payments and giving receipts to customers. Bill splitting also has an important role in POS. For example, If some friends come together for dinner and if they want to divide the bill then it is possible by POS bill splitting. This slide will show how to split bills in odoo 17 POS.
How to Create Map Views in the Odoo 17 ERPCeline George
The map views are useful for providing a geographical representation of data. They allow users to visualize and analyze the data in a more intuitive manner.
Synthetic Fiber Construction in lab .pptxPavel ( NSTU)
Synthetic fiber production is a fascinating and complex field that blends chemistry, engineering, and environmental science. By understanding these aspects, students can gain a comprehensive view of synthetic fiber production, its impact on society and the environment, and the potential for future innovations. Synthetic fibers play a crucial role in modern society, impacting various aspects of daily life, industry, and the environment. ynthetic fibers are integral to modern life, offering a range of benefits from cost-effectiveness and versatility to innovative applications and performance characteristics. While they pose environmental challenges, ongoing research and development aim to create more sustainable and eco-friendly alternatives. Understanding the importance of synthetic fibers helps in appreciating their role in the economy, industry, and daily life, while also emphasizing the need for sustainable practices and innovation.
The Art Pastor's Guide to Sabbath | Steve ThomasonSteve Thomason
What is the purpose of the Sabbath Law in the Torah. It is interesting to compare how the context of the law shifts from Exodus to Deuteronomy. Who gets to rest, and why?
The French Revolution, which began in 1789, was a period of radical social and political upheaval in France. It marked the decline of absolute monarchies, the rise of secular and democratic republics, and the eventual rise of Napoleon Bonaparte. This revolutionary period is crucial in understanding the transition from feudalism to modernity in Europe.
For more information, visit-www.vavaclasses.com
Unit 8 - Information and Communication Technology (Paper I).pdfThiyagu K
This slides describes the basic concepts of ICT, basics of Email, Emerging Technology and Digital Initiatives in Education. This presentations aligns with the UGC Paper I syllabus.
The Roman Empire A Historical Colossus.pdfkaushalkr1407
The Roman Empire, a vast and enduring power, stands as one of history's most remarkable civilizations, leaving an indelible imprint on the world. It emerged from the Roman Republic, transitioning into an imperial powerhouse under the leadership of Augustus Caesar in 27 BCE. This transformation marked the beginning of an era defined by unprecedented territorial expansion, architectural marvels, and profound cultural influence.
The empire's roots lie in the city of Rome, founded, according to legend, by Romulus in 753 BCE. Over centuries, Rome evolved from a small settlement to a formidable republic, characterized by a complex political system with elected officials and checks on power. However, internal strife, class conflicts, and military ambitions paved the way for the end of the Republic. Julius Caesar’s dictatorship and subsequent assassination in 44 BCE created a power vacuum, leading to a civil war. Octavian, later Augustus, emerged victorious, heralding the Roman Empire’s birth.
Under Augustus, the empire experienced the Pax Romana, a 200-year period of relative peace and stability. Augustus reformed the military, established efficient administrative systems, and initiated grand construction projects. The empire's borders expanded, encompassing territories from Britain to Egypt and from Spain to the Euphrates. Roman legions, renowned for their discipline and engineering prowess, secured and maintained these vast territories, building roads, fortifications, and cities that facilitated control and integration.
The Roman Empire’s society was hierarchical, with a rigid class system. At the top were the patricians, wealthy elites who held significant political power. Below them were the plebeians, free citizens with limited political influence, and the vast numbers of slaves who formed the backbone of the economy. The family unit was central, governed by the paterfamilias, the male head who held absolute authority.
Culturally, the Romans were eclectic, absorbing and adapting elements from the civilizations they encountered, particularly the Greeks. Roman art, literature, and philosophy reflected this synthesis, creating a rich cultural tapestry. Latin, the Roman language, became the lingua franca of the Western world, influencing numerous modern languages.
Roman architecture and engineering achievements were monumental. They perfected the arch, vault, and dome, constructing enduring structures like the Colosseum, Pantheon, and aqueducts. These engineering marvels not only showcased Roman ingenuity but also served practical purposes, from public entertainment to water supply.
Model Attribute Check Company Auto PropertyCeline George
In Odoo, the multi-company feature allows you to manage multiple companies within a single Odoo database instance. Each company can have its own configurations while still sharing common resources such as products, customers, and suppliers.
How to Make a Field invisible in Odoo 17Celine George
It is possible to hide or invisible some fields in odoo. Commonly using “invisible” attribute in the field definition to invisible the fields. This slide will show how to make a field invisible in odoo 17.
2. Why CNN??
The Convolutional Neural Network (CNN or
ConvNet) is a subtype of Neural Networks that
is mainly used for applications in image and
speech recognition.
Its built-in convolutional layer reduces the high
dimensionality of images without losing its
information. That is why CNNs are especially
suited for this use case.
6. VGG Net
● It is a classical CNN architecture developed to increase the depth of CNN’s to
increase the models performance.
● VGG stands for Visual Geometry Group is a deep CNN model with multiple
layers,has about 16 or 19 layers.
● It is a convolutional neural network model proposed by A. Zisserman and K.
Simonyan from the University of Oxford
● The VGG16 model achieves almost 92.7% top-5 test accuracy in
ImageNet.(ImageNet is a dataset consisting of more than 14 million images
belonging to nearly 1000 classes.
7. VGG Architecture:
● It has 16 layers(13 are convolutional layers
and 3 fully connected layers).
● Input: takes an image input size of 224 by
224.
● Convolutional Layers: uses a 3 by 3 filter
and a stride size of 1 and it is followed by
RELU unit which is rectified linear unit
activation function.
● Hidden Layers: all the hidden layers in
VGG use RELU.
● Fully Connected Layers: There are 3 fully
connected layers,the first two have 4096
channels each, and the third has 1000
channels, 1 for each class.
8. ● There are a few convolution layers in this architecture
followed by a pooling layer that reduces the height and the
width (reducing volume).
● VGG16 focuses on 3x3 filter convolution layers with stride 1
and always utilizes the same padding and MaxPool layer of a
2x2 filter with stride 2.
● If we look at the number of filters that we can use, around 64
filters are available that we can double to about 128 and then
to 256 filters. In the last layers, we can use 512 filters.
9. VGG 19:
● VGG19 model (also VGGNet-19) is the same as
the VGG16 except that it has 19 layers.
● The “16” and “19” stand for the number of weight
layers in the model (convolutional layers).
● This means that VGG19 has three more
convolutional layers than VGG16.
10. VGG 16 vs VGG 19
VGG 16 VGG 19
● 16 Layers ● 19 Layers
● Has less Weight. ● Has more Weight
● The size of the “VGG-16” network in
terms of fully connected nodes is 533
MB
● The size of the “VGG-16” network in
terms of fully connected nodes is 574 MB
11. Advantage of VGG 19 over VGG 16
● The main advantage of VGG19 over VGG16 is that it has more layers, which enables it to learn more
complex representations of the data.
● VGG19 is more accurate than VGG16
In Conclusion:
● VGG16 and VGG19 are both convolutional neural networks developed by the Visual Geometry Group (VGG)
at the University of Oxford, both are trained for image classification tasks.
● The main difference between them is the number of layers, VGG16 is a 16-layer CNN, while VGG19 is a 19-
layer CNN,
● VGG19 is more accurate than VGG16
12. DATA
AUGMENTATION
Artificially increasing the training
set by creating modified copies of
a dataset using existing data.
Includes making minor changes
to the dataset or using deep
learning to generate new data
points.
13. AUGMENTED DATA
It is driven from original data with
some minor changes to increase the
size and diversity of the training
set.
It is generated artificially without
using the original dataset. It often
uses DNNs (Deep Neural
Networks) and GANs (Generative
Adversarial Networks) to
generate synthetic data.
SYNTHETIC DATA
14. WHY SHOULD WE USE DATA AUGMENTATION ??
➢ To prevent models from overfitting.
➢ The initial training set is too small.
➢ To improve the model accuracy.
➢ To Reduce the operational cost of labeling and cleaning the raw dataset.
➢ Increases generalization ability of the models.
➢ Helps to resolve class imbalance issues in classification.
15. LIMITATIONS OF DATA AUGMENTATION
➢ The biases in the original dataset persist in the augmented data.
➢ Quality assurance for data augmentation is expensive.
➢ Research and development are required to build a system with advanced
applications. For example, generating high-resolution images using
GANs can be challenging.
➢ Finding an effective data augmentation approach can be challenging.
17. AUDIO DATA
AUGMENTATION
➢ Noise injection: add
gaussian or random noise
➢ Shifting: shift audio left (fast
forward) or right with
random seconds.
➢ Changing the speed:
stretches times series by a
fixed rate.
➢ Changing the pitch:
randomly change the pitch of
the audio.
18. TEXT DATA
AUGMENTATION
➢ Word or sentence shuffling
➢ Word replacement
➢ Syntax-tree manipulation
➢ Random word insertion
➢ Random word deletion
19. IMAGE AUGMENTATION
➢ Geometric transformations : randomly flip, crop, rotate, stretch,
and zoom images.
➢ Color space transformations : randomly change RGB color
channels, contrast, saturation and brightness.
➢ Kernel filters: randomly change the sharpness or blurring of the
image.
➢ Random erasing: delete some part of the initial image.
➢ Mixing images: blending and mixing multiple images.
21. Adversarial Training
based Augmentation
The objective is to transform the
images to deceive a deep-learning
model.
The method learns to generate
masks which when applied to the
input image, generated different
augmented images.
22. GAN based
Augmentation Synthesize images for data
augmentation
Generator is to generate fake
images from the latent space and
the goal of the discriminator is to
distinguish the synthetic fake
images from the real images
23. Neural Style Transfer
based Augmentation Deep Neural Networks are
trained to extract the content(high
level features) from one image
and style(low level features) from
another image and compose the
augmented image using the
extracted content and style.
24. Data Augmentation
in Medical field
Points to remember
➢ Image quality
➢ Tumor location and size
➢ Class imbalance
➢ Validation and evaluation
27. 2. Data Preparation and Preprocessing
➢ Convert the image to grayscale, and blur it slightly
➢ Threshold the image, then perform a series of erosions and dilations to
remove any small regions of noise
➢ Crop new image out of the original image using the four extreme points
(left, right, top, bottom)
30. 3. Load the data
➢ Read the image.
➢ Crop the part of the image representing only the brain.
➢ Resize the image
➢ Apply normalization because we want pixel values to be scaled to the
range 0-1.
➢ Append the image to X and its label to y
➢ Shuffle X and y
33. 4. Split the data
Split X and y into training, validation (development) and validation sets.
➢ 70% of the data for training.
➢ 15% of the data for validation.
➢ 15% of the data for testing.
34. 5. Build the model
➢ Load the VGG16 model, pretrained on ImageNet
➢ Freeze the layers in the base model so that they are not trainable
➢ Create a new model that includes the VGG16 base model and additional layers for
classification