3. COGNITION
Cognition is "the mental action or process of acquiring knowledge
and understanding through thought,experience,and the senses”.
Cognitive processes use existing knowledge and generate new
knowledge.
The concept of cognition is closely related to abstract
concepts such as mind and intelligence.
Technologies – natural language processing, machine learning
4. Cognitive technologies are products of the field of
artificial intelligence.
So what will be a cognitive system..??
They are able to perform tasks that only humans used to
be able to do.
5. In order to achieve this new level of
computing, cognitive systems must be
1. Adaptive
2. Interactive
3. Iterative and stateful
4. Contextual
7. COGNITIVE COMPUTING
Mimicking the function of a human brain and handling
human kind of problems.
Combination of technologies to understand human
interaction and provide answers.
Cognitive computing systems use machine learning
algorithms.
Such systems continually acquire knowledge from the
data.
8. Three eras of computing
1. Tabulating system(1900)-adding objects
2. Programmable system(1950)-processing numbers
3. Cognitive computing- make sense of data, prediction
9. WHEN COMPUTER BECOMES BRAINS
The human brain integrates memory and processing together
It reacts to things in its environment
But a computer has separate memory and processing.
It does its work by a clock.
10. For the first time in computing history, it’s possible for machines
to learn from experience and penetrate the complexity of data to
identify associations.
The field is called cognitive analytics™—inspired by how the
human brain processes information, draws conclusions instead of
depending on predefined rules and structured queries.
Cognitive analytics can push past the limitations of human
cognition, allowing us to process and understand big data in real
time.
Context-based hypotheses can be formed by exploring massive
numbers of permutations of potential relationships of influence
and causality—leading to conclusions .
11. ARCHITECTURE
Three major elements..
a) Executive layer
planner and simulation
b) Coordination layer
diagnoser, model base, monitor, and executor
c) Execution layer
effector and perceptor.
13. WATSON
Super computer developed by IBM Research
Named for IBM’s founder –Thomas J Watson
Research team led by Principal Investigator
Dr. David Ferrucci.
Programmed by 25 IBM scientists
Not connected to Internet
14. • Watson is an artificial intelligent computer system
• Capable of answering questions posed in natural languages
• Able to build knowledge and learn ,understand natural
language and interact more naturally with human beings
• More than hundred techniques are used to analyse natural
language
WATSON
15. TECHNOLOGY
Question answering technology
Deep understanding of human language
Software-DeepQA
Runs on cluster of power 750 computers
10 racks holding 90 servers
2880 3.55GHz power processor cores
16 TB memory
Hold approximately one million books
18. GOAL OF WATSON
According to IBM, "The goal is to have computers start to
interact in natural human terms across a range of applications and
processes, understanding the questions that humans ask and
providing answers that humans can understand and justify.“
The goal of cognitive computing is to create automated IT
systems that are capable of solving problems without requiring
human assistance.
19. APPLICATIONS
•HEALTHCARE
The Data Explosion-Medical data are expected to double every 73
days by 2020.
The Great Unknown-80% of health data are invisible to current
systems because they are unstructured. Watson Health can see them.
A Quick Study-Watson can read 40 million documents in 15
seconds.
20. • Cognitive Computing is making agriculture more productive.
• Cognitive Computing is making cars smarter.
• Cognitive Computing is making our homes and offices more
secure, as well as our borders.
21. FUTURE ASPECTS
Augmented Virtual Reality
Sentient Systems
Specialized Deep Learning on Big Data sets
Generalized Artificial Intelligence systems
22. CONCLUSION
Cognitive computing has been characterized as a set of
autonomous and perceptive knowledge processing theories and
technologies mimicking the mechanisms of the brain beyond
conventional imperative data processing