OpenPOWER AI virtual University's - focus on bringing together industry, government and academic expertise to connect and help shape the AI future .
https://www.youtube.com/channel/UCYLtbUp0AH0ZAv5mNut1Kcg
Artificial Neural Networks have been very successfully used in several machine learning applications. They are often the building blocks when building deep learning systems. We discuss the hypothesis, training with backpropagation, update methods, regularization techniques.
Приложение на изкуствен интелект при анализа на медийно съдържание в интернет...PlovDev Conference
Data aggregation / transformation / refinement
Train corpora
Tokenization process
Attributes and instances
Vector space modeling / Bag of words
Feature selection
Machine learning algorithms Classifiers – binary / multi-class, multi-label / problem transformation methods
Learning evaluation
Quality Management workflow / “Human in the loop” supervised learning
Artificial Neural Networks have been very successfully used in several machine learning applications. They are often the building blocks when building deep learning systems. We discuss the hypothesis, training with backpropagation, update methods, regularization techniques.
Приложение на изкуствен интелект при анализа на медийно съдържание в интернет...PlovDev Conference
Data aggregation / transformation / refinement
Train corpora
Tokenization process
Attributes and instances
Vector space modeling / Bag of words
Feature selection
Machine learning algorithms Classifiers – binary / multi-class, multi-label / problem transformation methods
Learning evaluation
Quality Management workflow / “Human in the loop” supervised learning
Mathematical Functions
Types of functions
Activation function
Laws of activation function
Types of Activation functions
Limitations of activation function
Piotr Mirowski - Review Autoencoders (Deep Learning) - CIUUK14Daniel Lewis
Piotr Mirowski (of Microsoft Bing London) presented Review of Auto-Encoders to the Computational Intelligence Unconference 2014, with our Deep Learning stream. These are his slides. Original link here: https://piotrmirowski.files.wordpress.com/2014/08/piotrmirowski_ciunconf_2014_reviewautoencoders.pptx
He also has Matlab-based tutorial on auto-encoders available here:
https://github.com/piotrmirowski/Tutorial_AutoEncoders/
Artificial neural network for machine learninggrinu
An Artificial Neurol Network (ANN) is a computational model. It is based on the structure and functions of biological neural networks. It works like the way human brain processes information. ANN includes a large number of connected processing units that work together to process information. They also generate meaningful results from it.
Neural networks Self Organizing Map by Engr. Edgar Carrillo IIEdgar Carrillo
This presentation talks about neural networks and self organizing maps. In this presentation,Engr. Edgar Caburatan Carrillo II also discusses its applications.
Deep Learning Interview Questions And Answers | AI & Deep Learning Interview ...Simplilearn
This Deep Learning interview questions and answers presentation will help you prepare for Deep Learning interviews. This presentation is ideal for both beginners as well as professionals who are appearing for Deep Learning, Machine Learning or Data Science interviews. Learn what are the most important Deep Learning interview questions and answers and know what will set you apart in the interview process.
Some of the important Deep Learning interview questions are listed below:
1. What is Deep Learning?
2. What is a Neural Network?
3. What is a Multilayer Perceptron (MLP)?
4. What is Data Normalization and why do we need it?
5. What is a Boltzmann Machine?
6. What is the role of Activation Functions in neural network?
7. What is a cost function?
8. What is Gradient Descent?
9. What do you understand by Backpropagation?
10. What is the difference between Feedforward Neural Network and Recurrent Neural Network?
11. What are some applications of Recurrent Neural Network?
12. What are Softmax and ReLU functions?
13. What are hyperparameters?
14. What will happen if learning rate is set too low or too high?
15. What is Dropout and Batch Normalization?
16. What is the difference between Batch Gradient Descent and Stochastic Gradient Descent?
17. Explain Overfitting and Underfitting and how to combat them.
18. How are weights initialized in a network?
19. What are the different layers in CNN?
20. What is Pooling in CNN and how does it work?
Simplilearn’s Deep Learning course will transform you into an expert in deep learning techniques using TensorFlow, the open-source software library designed to conduct machine learning & deep neural network research. With our deep learning course, you’ll master deep learning and TensorFlow concepts, learn to implement algorithms, build artificial neural networks and traverse layers of data abstraction to understand the power of data and prepare you for your new role as deep learning scientist.
Why Deep Learning?
It is one of the most popular software platforms used for deep learning and contains powerful tools to help you build and implement artificial neural networks.
Advancements in deep learning are being seen in smartphone applications, creating efficiencies in the power grid, driving advancements in healthcare, improving agricultural yields, and helping us find solutions to climate change.
There is booming demand for skilled deep learning engineers across a wide range of industries, making this deep learning course with TensorFlow training well-suited for professionals at the intermediate to advanced level of experience. We recommend this deep learning online course particularly for the following professionals:
1. Software engineers
2. Data scientists
3. Data analysts
4. Statisticians with an interest in deep learning
Learn more at: https//www.simplilearn.com
Mathematical Functions
Types of functions
Activation function
Laws of activation function
Types of Activation functions
Limitations of activation function
Piotr Mirowski - Review Autoencoders (Deep Learning) - CIUUK14Daniel Lewis
Piotr Mirowski (of Microsoft Bing London) presented Review of Auto-Encoders to the Computational Intelligence Unconference 2014, with our Deep Learning stream. These are his slides. Original link here: https://piotrmirowski.files.wordpress.com/2014/08/piotrmirowski_ciunconf_2014_reviewautoencoders.pptx
He also has Matlab-based tutorial on auto-encoders available here:
https://github.com/piotrmirowski/Tutorial_AutoEncoders/
Artificial neural network for machine learninggrinu
An Artificial Neurol Network (ANN) is a computational model. It is based on the structure and functions of biological neural networks. It works like the way human brain processes information. ANN includes a large number of connected processing units that work together to process information. They also generate meaningful results from it.
Neural networks Self Organizing Map by Engr. Edgar Carrillo IIEdgar Carrillo
This presentation talks about neural networks and self organizing maps. In this presentation,Engr. Edgar Caburatan Carrillo II also discusses its applications.
Deep Learning Interview Questions And Answers | AI & Deep Learning Interview ...Simplilearn
This Deep Learning interview questions and answers presentation will help you prepare for Deep Learning interviews. This presentation is ideal for both beginners as well as professionals who are appearing for Deep Learning, Machine Learning or Data Science interviews. Learn what are the most important Deep Learning interview questions and answers and know what will set you apart in the interview process.
Some of the important Deep Learning interview questions are listed below:
1. What is Deep Learning?
2. What is a Neural Network?
3. What is a Multilayer Perceptron (MLP)?
4. What is Data Normalization and why do we need it?
5. What is a Boltzmann Machine?
6. What is the role of Activation Functions in neural network?
7. What is a cost function?
8. What is Gradient Descent?
9. What do you understand by Backpropagation?
10. What is the difference between Feedforward Neural Network and Recurrent Neural Network?
11. What are some applications of Recurrent Neural Network?
12. What are Softmax and ReLU functions?
13. What are hyperparameters?
14. What will happen if learning rate is set too low or too high?
15. What is Dropout and Batch Normalization?
16. What is the difference between Batch Gradient Descent and Stochastic Gradient Descent?
17. Explain Overfitting and Underfitting and how to combat them.
18. How are weights initialized in a network?
19. What are the different layers in CNN?
20. What is Pooling in CNN and how does it work?
Simplilearn’s Deep Learning course will transform you into an expert in deep learning techniques using TensorFlow, the open-source software library designed to conduct machine learning & deep neural network research. With our deep learning course, you’ll master deep learning and TensorFlow concepts, learn to implement algorithms, build artificial neural networks and traverse layers of data abstraction to understand the power of data and prepare you for your new role as deep learning scientist.
Why Deep Learning?
It is one of the most popular software platforms used for deep learning and contains powerful tools to help you build and implement artificial neural networks.
Advancements in deep learning are being seen in smartphone applications, creating efficiencies in the power grid, driving advancements in healthcare, improving agricultural yields, and helping us find solutions to climate change.
There is booming demand for skilled deep learning engineers across a wide range of industries, making this deep learning course with TensorFlow training well-suited for professionals at the intermediate to advanced level of experience. We recommend this deep learning online course particularly for the following professionals:
1. Software engineers
2. Data scientists
3. Data analysts
4. Statisticians with an interest in deep learning
Learn more at: https//www.simplilearn.com
This is a slide deck from a presentation, that my colleague Shirin Glander (https://www.slideshare.net/ShirinGlander/) and I did together. As we created our respective parts of the presentation on our own, it is quite easy to figure out who did which part of the presentation as the two slide decks look quite different ... :)
For the sake of simplicity and completeness, I just copied the two slide decks together. As I did the "surrounding" part, I added Shirin's part at the place when she took over and then added my concluding slides at the end. Well, I'm sure, you will figure it out easily ... ;)
The presentation was intended to be an introduction to deep learning (DL) for people who are new to the topic. It starts with some DL success stories as motivation. Then a quick classification and a bit of history follows before the "how" part starts.
The first part of the "how" is some theory of DL, to demystify the topic and explain and connect some of the most important terms on the one hand, but also to give an idea of the broadness of the topic on the other hand.
After that the second part dives deeper into the question how to actually implement DL networks. This part starts with coding it all on your own and then moves on to less coding step by step, depending on where you want to start.
The presentation ends with some pitfalls and challenges that you should have in mind if you want to dive deeper into DL - plus the invitation to become part of it.
As always the voice track of the presentation is missing. I hope that the slides are of some use for you, though.
This is a slide deck from a presentation, that my colleague Uwe Friedrichsen (https://www.slideshare.net/ufried/) and I did together. As we created our respective parts of the presentation on our own, it is quite easy to figure out who did which part of the presentation as the two slide decks look quite different ... :)
For the sake of simplicity and completeness, Uwe copied the two slide decks together. As he did the "surrounding" part, he added my part at the place where I took over and then added concluding slides at the end. Well, I'm sure, you will figure it out easily ... ;)
The presentation was intended to be an introduction to deep learning (DL) for people who are new to the topic. It starts with some DL success stories as motivation. Then a quick classification and a bit of history follows before the "how" part starts.
The first part of the "how" is some theory of DL, to demystify the topic and explain and connect some of the most important terms on the one hand, but also to give an idea of the broadness of the topic on the other hand.
After that the second part dives deeper into the question how to actually implement DL networks. This part starts with coding it all on your own and then moves on to less coding step by step, depending on where you want to start.
The presentation ends with some pitfalls and challenges that you should have in mind if you want to dive deeper into DL - plus the invitation to become part of it.
As always the voice track of the presentation is missing. I hope that the slides are of some use for you, though.
CSSC × GDSC: Intro to Machine Learning!
Aaron Shah and Manav Bhojak on October 5, 2023
🤖 Join us for an exciting ML Workshop! 🚀 Dive into the world of Machine Learning, where we'll unravel the mysteries of CNNs, RNNs, Transformers, and more. 🤯
Get ready to embark on a journey of discovery! We'll begin with an easy-to-follow introduction to the fascinating realm of ML. 📚
🛠️ In our hands-on session, we'll walk you through setting up your environment. No tech hurdles here! 🌐
🔍 Then, we'll get down to the nitty-gritty, guiding you through our starter code for a thrilling hands-on example. Together, we'll explore the power of ML in action! 💡
The Libre-SOC Project aims to create an entirely Libre-Licensed, transparently-developed fully auditable Hybrid 3D CPU-GPU-VPU, using the Supercomputer-class OpenPOWER ISA as the foundation.
Our first test ASIC is a 180nm "Fixed-Point" Power ISA v3.0B processor, 5.1mm x 5.9mm, as a proof-of-concept for the team, whose primary expertise is in Software Engineering. Software Engineering training brings a radically different approach to Hardware development: extensive unit tests, source code revision control, automated development tools are normal. Libre Project Management brings even more: bug trackers, mailing lists, auditable IRC logs and a wiki are standard fare for Libre Projects that are simply not normal Industry-Standard practice.
This talk therefore goes through the workflow, from the original HDL through to the GDS-II layout, showing how we were able to keep track of the development that led to the IMEC 180nm tape-out in July 2021. In particular, by following a parallel development process involving "Real" and "Symbolic" Cell Libraries, developed by Chips4Makers, will be shown how our developers did not need to sign a Foundry NDA, but were still able to work side-by-side with a University that did. With this parallel development process, the University upheld their NDA obligations, and Libre-SOC were simultaneously able to honour its Transparency Objectives.
Workload Transformation and Innovations in POWER Architecture Ganesan Narayanasamy
IT Industry is going through two major transformations. One is adaption of AI and tight integration of the same in the commercial applications and enterprise workflow. Two the transformation in software architecture through the concepts like microservices and the cloud native architecture. These transformation alongside the aggressive adaption of IoT/mobile and 5G in all our day today activities is making the world operate in more real time manner which opens-up a new challenge to improve the hardware architecture to adapt to these requirements. These above two major transformation pushes the boundary of the entire systems stack making the designer rethink hardware. This talk presents you a picture of how the enterprise Industry leading POWER architecture is transforming to fulfill the performance demands of these newer generation workloads with primary focus on the AI acceleration on the chip.
July 16th 2021 , Friday for our newest workshop with DoMS, IIT Roorkee, Concept to Solutions using OpenPOWER Stack. It's time to discover advances in #DeepLearning tools and techniques from the world's leading innovators across industries, research, and public speakers.
Register here:
https://lnkd.in/ggxMq2N
This presentation covers two uses cases using OpenPOWER Systems
1. Diabetic Retinopathy using AI on NVIDIA Jetson Nano: The objective is to classify the diabetic level solely on retina image in a remote area with minimum doctor's inference. The model uses VGG16 network architecture and gets trained from scratch on POWER9. The model was deployed on the Jetson Nano board.
1. Classifying Covid positivity using lung X-ray images: The idea is to build ML models to detect positive cases using X-ray images. The model was trained on POWER9, and the application was developed using Python.
IBM Bayesian Optimization Accelerator (BOA) is a do-it-yourself toolkit to apply state-of-the-art Bayesian inferencing techniques and obtain optimal solutions for complex, real-world design simulations without requiring deep machine learning skills. This talk will describe IBM BOA, its differentiation and ease of use, and how researchers can take advantage of it for optimizing any arbitrary HPC simulation.
This presentation covers various partners and collaborators who are currently working with OpenPOWER foundation ,Use cases of OpenPOWER systems in multiple Industries , OpenPOWER Workgroups and OpenCAPI features .
The IBM POWER10 processor represents the 10th generation of the POWER family of enterprise computing engines. Its performance is a result of both powerful processing cores and high-bandwidth intra- and inter-chip interconnect. POWER10 systems can be configured with up to 16 processor chips and 1920 simultaneous threads of execution. Cross-system memory sharing, through the new Memory Inception technology, and 2 Petabytes of addressing space support an expansive memory system. The POWER10 processing core has been significantly enhanced over its POWER9 predecessor, including a doubling of vector units and the addition of an all-new matrix math engine. Throughput gains from POWER9 to POWER10 average 30% at the core level and three-fold at the socket level. Those gains can reach ten- or twenty-fold at the socket level for matrix-intensive computations.
Everything is changing from Health Care to the Automotive markets without forgetting Financial markets or any type of engineering everything has stopped being created as an individual or best-case scenario a team effort to something that is being developed and perfectioned by using AI and hundreds of computers.And even AI is something that we no longer can run in a single computer, no matter how powerful it is. What drives everything today is HPC or High-Performance Computing heavily linked to AI In this session we will discuss about AI, HPC computing, IBM Power architecture and how it can help develop better Healthcare, better Automobiles, better financials and better everything that we run on them
Macromolecular crystallography is an experimental technique allowing to explore 3D atomic structure of proteins, used by academics for research in biology and by pharmaceutical companies in rational drug design. While up to now development of the technique was limited by scientific instruments performance, recently computing performance becomes a key limitation. In my presentation I will present a computing challenge to handle 18 GB/s data stream coming from the new X-ray detector. I will show PSI experiences in applying conventional hardware for the task and why this attempt failed. I will then present how IC 922 server with OpenCAPI enabled FPGA boards allowed to build a sustainable and scalable solution for high speed data acquisition. Finally, I will give a perspective, how the advancement in hardware development will enable better science by users of the Swiss Light Source.
AI in healthcare and Automobile Industry using OpenPOWER/IBM POWER9 systemsGanesan Narayanasamy
As the adoption of AI technologies increases and matures, the focus will shift from exploration to time to market, productivity and integration with existing workflows. Governing Enterprise data, scaling AI model development, selecting a complete, collaborative hybrid platform and tools for rapid solution deployments are key focus areas for growing data scientist teams tasked to respond to business challenges. This talk will cover the challenges and innovations for AI at scale for the Industires such as Healthcare and Automotive , the AI ladder and AI life cycle and infrastructure architecture considerations.
This talk gives an introduction about Healthcare Use cases - The AI ladder and Lifestyle AI at Scale Themes The iterative nature of the workflow and some of the important components to be aware in developing AI health care solutions were being discussed. The different types of algorithms and when machine learning might be more appropriate in deep learning or the other way will also be discussed. Use cases in terms of examples are also shared as part of this presentation .
Healthcare has became one of the most important aspects of everyones life. Its importance has surged due to the latests outbreaks and due to this latest pandemic it has become mandatory to collaborate to improve everyones Healthcare as soon as possible.
IBM has reacted quickly sharing not only its knowledge but also its Artificial Intelligence Supercomputers all around the world.
Those Supercomputers are helping to prevail this outbreak and also future ones.
They have completely different features compared to proposals from other players of this Supercomputers market.
We will try to make a quick look at the differences of those AI focused Supercomputers and how they can help in the R&D of Healthcare solutions for everyone, from those ones with access to a big IBM AI Supercomputer to those ones with access to only one small IBM AI focused server.
Healthcare has became one of the most important aspects of everyones life. Its importance has surged due to the latests outbreaks and due to this latest pandemic it has become mandatory to collaborate to improve everyones Healthcare as soon as possible.
IBM has reacted quickly sharing not only its knowledge but also its Artificial Intelligence Supercomputers all around the world.
Those Supercomputers are helping to prevail this outbreak and also future ones.
They have completely different features compared to proposals from other players of this Supercomputers market.
We will try to make a quick look at the differences of those AI focused Supercomputers and how they can help in the R&D of Healthcare solutions for everyone, from those ones with access to a big IBM AI Supercomputer to those ones with access to only one small IBM AI focused server.
Moving object recognition (MOR) corresponds to the localization and classification of moving objects in videos. Discriminating moving objects from static objects and background in videos is an essential task for many computer vision applications. MOR has widespread applications in intelligent visual surveillance, intrusion detection, anomaly detection and monitoring, industrial sites monitoring, detection-based tracking, autonomous vehicles, etc. In this session, Murari provided a poster about the deep learning algorithms to identify both locations and corresponding categories of moving objects with a convolutional network. The challenges in developing such algorithms have been discussed.
Clarisse Hedglin from IBM presented this as part of 3 days International Summit .. She shared the scenarios AI can solve for today using the IBM AI infrastructure.
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
Maruthi Prithivirajan, Head of ASEAN & IN Solution Architecture, Neo4j
Get an inside look at the latest Neo4j innovations that enable relationship-driven intelligence at scale. Learn more about the newest cloud integrations and product enhancements that make Neo4j an essential choice for developers building apps with interconnected data and generative AI.
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
2. Machine Learning Basics
Machine learning is a field of computer science that gives
computers the ability to learn without being explicitly
programmed
Methods that can learn from and make predictions on data
Labeled
Data
Labeled
Data
Machine
Learning
algorithm
Learned
model
Prediction
Training
Predictio
n
3. Types of Learning
01 Supervised: Learning with a labeled training set
Example: email classification with already labeled
emails
02 Unsupervised: Discover patterns in unlabeled data
Example: cluster similar documents based on text
03 Reinforcement learning: learn to act based on
feedback/reward
Example: learn to play Go, reward: win or lose
Regression
class A
class B
Classification Clustering
4. Machine Learning vs Deep Learning
● Most machine learning methods work well because of human-designed
representations and input features
● ML becomes just optimizing weights to best make a final prediction.
● Thus, Machine learning uses algorithms to parse data, learn from that
data, and make informed decisions based on what it has learned.
Deep learning structures algorithms in layers to create an “artificial
neural network” that can learn and make intelligent decisions on its
own.
● In practical terms, deep learning is just a subset of machine learning. It
technically is machine learning and functions in a similar way (hence
why the terms are sometimes loosely interchanged), but its capabilities
are different.
5. Machine Learning vs Deep Learning
https://www.xenonstack.com/blog/static/public/uploads/media/machine-learning-vs-deep-learning.png
6. What is Deep Learning?
● Deep learning is a subset of machine learning in Artificial Intelligence
(AI) that has networks capable of learning unsupervised from data that
is unstructured or unlabeled. Also known as Deep Neural Learning or
Deep Neural Network.
In Simple Terms,
● Deep learning algorithms attempt to learn (multiple levels of)
representation by using a hierarchy of multiple layers
● If you provide the system tons of information, it begins to understand it
and respond in useful ways.
7. Why Deep Learning is useful?
● Manually designed features are often over-specified,
incomplete and take a long time to design and validate
● Learned Features are easy to adapt, fast to learn
● Deep learning provides a very flexible, universal, learnable
framework for representing world, visual and linguistic
information.
● Can learn both unsupervised and supervised
● Effective end-to-end joint system learning
● Utilize large amounts of training data
8. Introduction to Neural Networks
● A deep neural network consists of a hierarchy of layers, whereby
each layer transforms the input data into more abstract
representations (e.g. edge -> nose -> face).
The output layer combines those features to make predictions.
● It consists of one input, one output and multiple fully-connected
hidden layers in between.
Each layer is represented as a series of neurons and
progressively extracts higher and higher-level features of the input
until the final layer essentially makes a decision about what the
input shows.
The more layers the network has, the higher level features it will
learn.
10. What is a neuron?
● Neurons are trained to filter and detect specific features or
patterns (e.g. edge, nose) by receiving weighted input,
transforming it with the activation function and passing it to the
● An artificial neuron contains an activation function and has several
incoming and outgoing weighted connections.
11. Activation Functions
● They define the output of that node given an input or set of inputs
● They decide whether a neuron should be activated or not.
● The Activation Functions can be basically divided into 2 types-
● Linear Activation Function
Identity Activation Function
● Non-linear Activation Functions
Sigmoid or Logistic Activation Function
Tanh or hyperbolic tangent Activation Function
ReLU (Rectified Linear Unit) Activation Function
Leaky ReLU
12. Types of Activation Functions
● Linear Activation Function
1. Identity Activation Function
● Non-linear Activation Functions
1. Sigmoid or Logistic Activation Function
2. Tanh or hyperbolic tangent Activation Function
3. ReLU (Rectified Linear Unit) Activation Function
4. Leaky ReLU
13. Linear Activation Function - Identity Function
● The output of the functions will not
be confined between any range.
● Equation : f(x) = x
● Range : (-infinity to infinity)
● It doesn’t help with the complexity
or various parameters of usual
data that is fed to the neural
networks. Identity Activation Function
14. Non-Linear Activation Function - Sigmoid
● It exists between (0 to 1).
● Used for models where we have
to predict the probability as an
output, sigmoid is the right choice.
● The function is differentiable, we
can find the slope of the sigmoid
curve at any two points.
● The function is monotonic but
function’s derivative is not.
Sigmoid Activation Function
15. Non-Linear Activation Function - Tanh
● The range of the tanh function is
from (-1 to 1). Tanh is also sigmoidal
● Negative inputs will be mapped
strongly negative and the zero
inputs will be mapped near zero in
the tanh graph.
● It is differentiable and is monotonic
while its derivative is not monotonic.
● The tanh function is mainly used
classification between two classes
● The function is monotonic but
function’s derivative is not. Hyperbolic Tangent Activation Function
16. Non-Linear Activation Function - ReLU (Rectified
Linear Unit)
● The ReLU is half rectified (from
bottom). f(z) is zero when z is less
than zero and f(z) is equal to z when
z is above or equal to zero.
● Range: [ 0 to infinity)
● The function and its derivative both
are monotonic.
● Issue: Negative values become
zero immediately which decreases
the ability of the model to fit or train
from the data properly
ReLU Activation Function
17. Non-Linear Activation Function - Leaky ReLU
(Rectified Linear Unit)
● The leak helps to increase the range
of the ReLU function. Usually, the
value of a is 0.01 or so.
● When a is not 0.01 then it is called
Randomized ReLU.
● Range: (-infinity to infinity).
● Both Leaky and Randomized ReLU
functions are monotonic in nature.
Also, their derivatives also
monotonic in nature.
ReLU vs Leaky ReLU Activation Function
19. Overfitting
● Learned hypothesis may fit the
training data very well, including the
outliers (noise) but may thus fail to
generalize to new examples (test
data)
● How to overcome this:
1. Use of validation data
2. Regularization:
Dropout Layers
20. Regularization - Dropout
Dropout
● Randomly drop units (along with
their connections) during
training.
● Each unit retained with fixed
probability p, independent of
other units.
● Hyper-parameter p to be chosen
(tuned)
Dropout
Srivastava, Nitish, et al. "Dropout: a simple way to prevent neural
networks from overfitting." Journal of machine learning research (2014)
21. Regularization - L2 weight decay and
Early-Stopping
L2 = weight decay
● Regularization term that penalizes
big weights, added to the objective
● Weight decay value determines how
dominant regularization is during
gradient computation
● Big weight decay coefficient → big
penalty for big weights
Early-stopping
● Use validation error to decide when
to stop training
● Stop when monitored quantity has
not improved after ‘n’ subsequent
epochs
● Where, ‘n’ is called patience
22. Convolutional Neural Network
● Convolutional neural networks form a subclass of
feedforward neural networks that have special weight
constraints, individual neurons are tiled in such a way that
they respond to overlapping regions.
ConvNet has two parts:
● Feature learning (Conv, Relu,and Pool)
● Classification(FC and softmax)
24. Convolutional Layers
The objective of a Convolutional layer is to extract
features of the input volume.
SOME COMMON TERMS:
● Filter, Kernel, or Feature Detector: is a small matrix
used for features detection.
● Convolved Feature, Activation Map or Feature Map: is
the output volume formed by sliding the filter over the
image and computing the dot product.
● Receptive field is a local region of the input volume that
has the same size as the filter.
25. Convolutional Layers
● Depth is the number of filters.
● Stride has the objective of producing smaller output
volumes spatially.
● Zero-padding adds zeros around the outside of the input
volume so that the convolutions end up with the same
number of outputs as inputs. If we don’t use padding the
information at the borders will be lost after each Conv
layer
27. WORKING OF A CNN
Reference: https://cdn-images-
1.medium.com/max/800/1*_34EtrgYk6cQxlJ2br51HQ.gif
28. Pooling Layers
Reference:
https://shafeentejani.github.io/assets/images/pooling.gif
● Pool Layer performs a function to
reduce the spatial dimensions of the
input, and the computational
complexity of our model.
● It also controls overfitting. It
operates independently on every
depth slice of the input.
● There are different functions such as
Max Pooling, Average
Pooling, or L2-Norm Pooling.
● Max pooling (most used) only takes
the most important part (the value of
the brightest pixel) of the input.
Example-Max Pooling
29. Fully Connected (FC) Layers
● Fully_Connected Layer (FC):
Fully connected layers connect
every neuron in one layer to every
neuron in another layer.
● The last fully-connected layer may
use a softmax activation function for
classifying the generated features of
the input image into various classes
based on the training dataset. Fully Connected Layer
30. PoC : Scalable Implementation of Blood Cells
Classification Using Deep Learning
- Aditya Mitkari, Sayali Deshpande, Saurabh Kshirsagar
The existing method for blood cell classification may
contribute to inaccuracy, inconsistency and poor reliability
diagnosis that may lead to false diagnosis situation.
Thus it is a need to provide a fast, simple and efficient,
rotational and scale invariant blood cell identification system
which can be used in automating laboratory reporting.
Potential scenarios
● Accurate and faster blood cell analysis
● Real time diagnosis of patient condition
31. PoC : Scalable Implementation of Blood Cells
Classification Using Deep Learning
PowerAI
Training Data
from Kaggle
Predicted
Output
CNN
layers
Convolutio
nal Layer
RELU
Fully
Connected
Layers
4
classes
CNN
layers
Dropout
layers
Data
Preprocessing RELU
Processed
Data
32. PoC : Scalable Implementation of Blood Cells
Classification Using Deep Learning
The existing method for blood cell classification may
contribute to inaccuracy, inconsistency and poor reliability
diagnosis that may lead to false diagnosis situation.
Thus it is a need to provide a fast, simple and efficient,
rotational and scale invariant blood cell identification system
which can be used in automating laboratory reporting.
Potential scenarios
● Accurate and faster blood cell analysis
● Real time diagnosis of patient condition
33. PoC : Results
With the currently used
dataset we have achieved:
● A testing accuracy of
58%
● loss of 0.25 after
training for 5000 steps.
FUTURE SCOPE
By using a larger dataset with more
number of data points that reflect the
variance of the data we can improve the
model’s accuracy. We also plan to
implement our model for classification of
cell types other than white blood cells.
34. OpenPOWER Foundation and OpenPOWER
Academia
OpenPOWER Foundation is a consortium formed by companies including
Google, IBM , Nvidia , Mellanox and Tyan revolving around Power
Architecture.
It is an open technical community based on the POWER architecture,
enabling collaborative development and opportunity for member
differentiation and industry growth.
It is the intent of OpenPOWER Foundation to:
● Opening the POWER architecture to give the industry the ability to
innovate across the full Hardware and Software stack
35. OpenPOWER Academia
● The OpenPOWER Academia Discussion Group (ADG) serves as a
platform within the OpenPOWER Foundation for the academic
members.
● It comprises a broad range of academic institutions with strong
competence, e.g., in providing and operating high-performance
computing facilities or in developing scientific applications, scalable
numerical methods, new approaches for processing extreme scale data
or new methods for data analytics.
● The main motive is of OpenPOWER Academia is to encourage and
implement solutions to solve real world problems.
36. System Overview - OpenPOWER and
IBMPowerAI
OpenPOWER™ architecture as a way to innovate and evangelize the POWER
technologies.
IBMPowerAI software platform for deep learning with IBM® Power
Systems™, for rapid deployment of a fully optimized and supported platform
for AI with blazing performance.
IBM® Power Systems™ S822LC (Minsky)
● 128 threads, 4 x NVIDIA P100 GPUs with NVLink
● PushToCompute™ for compiling POWER8 applications and deploying
directly to the Nimbix Cloud
37. System Details
POWER9 – Premier Acceleration Platform
o State of the Art I/O and Acceleration Attachment Signaling
○ PCIe Gen 4 (16G) x 48 lanes
○ 192 GB/s duplex bandwidth
○ 25G Link x 48 lanes
○ 300 GB/s duplex bandwidth
o Robust Accelerated Compute Options with OPEN standards
○ On-Chip Acceleration
○ Gzip x1, 842 Compression x2, AES/SHA x2
○ CAPI 2.0
○ OpenCAPI 3.0 – High bandwidth, low latency and open interface
using 25G Link
○ NVLink 2.0 – Next generation of GPU/CPU bandwidth and
integration