The "hard" problem of concsiousness can be shown to be a non-problem because it is formulated using a seriously defective concept (the concept of "phenomenal consciousness" defined so as to rule out cognitive functionality and causal powers).
So the hard problem is an example of a well known type of philosophical problem that needs to be dissolved (fairly easily) rather than solved. For other examples, and a brief introduction to conceptual analysis, see http://www.cs.bham.ac.uk/research/projects/cogaff/misc/varieties-of-atheism.html
In contrast, the so-called "easy" problem requires detailed analysis of very complex and subtle features of perceptual processes, introspective processes and other mental processes, sometimes labelled "access consciousness": these have cognitive functions, but their complexity (especially the way details change as the environment changes or the perceiver moves) is considerable and very hard to characterise.
"Access consciousness" is complex also because it takes many different forms, since what individuals are conscious of and what uses being conscious of things can be put to, can vary hugely, from simple life forms, through many other animals and human infants, to sophisticated adult humans,
Finding ways of modelling these aspects of consciousness, and explaining how they arise out of physical mechanisms, requires major advances in the science of information processing systems -- including computer science and neuroscience.
There are empirical facts about introspection that have generated theories of consciousness but some of the empirical facts go unnoticed by philosophers.
The notion of a virtual machine is introduced briefly and illustrated using Conway's "Game of life" and other examples of virtual machinery that explain how contents of consciousness can have causal powers and can have intentionality (be able to refer to other things).
The beginnings of a research program are presented, showing how more examples can be collected and how notions of virtual machinery may need to be developed to cope with all the phenomena.
Historically, one of the toughest nuts to crack in both philosophy and science has been the so-called “hard problem of consciousness”:
How is it possible for the hard-wiring of our brain to produce inner perception, sentient experience, and self-awareness?
It’s a much more difficult conundrum to solve than one might expect upon first glance.
Because science is inherently objective, while consciousness is subjective, it has been very difficult for experts working from both sides of the divide to bridge this gap – to explain the precise mechanisms of conscious experience.
Join us in exploring the mysterious relationship between the brain and conscious experience.
Overview of the course. Introduction to image sciences, image processing and computer vision. Basics of machine learning, terminologies, paradigms. No-free lunch theorem. Supervised versus unsupervised learning. Clustering and K-Means. Classification and regression. Linear least squares and polynomial curve fitting. Model complexity and overfitting. Curse of dimensionality. Dimensionality reduction and principal component analysis. Image representation, semantic gap, image features, and classical computer vision pipelines.
Historically, one of the toughest nuts to crack in both philosophy and science has been the so-called “hard problem of consciousness”:
How is it possible for the hard-wiring of our brain to produce inner perception, sentient experience, and self-awareness?
It’s a much more difficult conundrum to solve than one might expect upon first glance.
Because science is inherently objective, while consciousness is subjective, it has been very difficult for experts working from both sides of the divide to bridge this gap – to explain the precise mechanisms of conscious experience.
Join us in exploring the mysterious relationship between the brain and conscious experience.
Overview of the course. Introduction to image sciences, image processing and computer vision. Basics of machine learning, terminologies, paradigms. No-free lunch theorem. Supervised versus unsupervised learning. Clustering and K-Means. Classification and regression. Linear least squares and polynomial curve fitting. Model complexity and overfitting. Curse of dimensionality. Dimensionality reduction and principal component analysis. Image representation, semantic gap, image features, and classical computer vision pipelines.
[Paper] Multiscale Vision Transformers(MVit)Susang Kim
multiscale feature hierarchies with the transformer model.
a multiscale pyramid of features with early layers operating at high spatial resolution to model simple low-level visual information visual pathway with a hierarchical model
COM2304: Introduction to Computer Vision & Image Processing Hemantha Kulathilake
At the end of this lesson, you should be able to;
Describe image.
Describe digital image processing and computer vision.
Compare and Contrast image processing and computer vision.
Describe image sources.
Identify the array of imaging application under the EM Image source.
Describe the components of Image processing system and computer vision system.
Anatomy of the posterior cerebral circulation from the neuroradiology point of view. Anatomy of the vertebral artery. Anatomy of the basilar artery. Important for Neuroradiologists and Neurointerventionalists.
Imagen: Photorealistic Text-to-Image Diffusion Models with Deep Language Unde...Vitaly Bondar
A presentation about a new Google Research paper in the text-to-image task - Imagen.
This latent diffusion-based model outperforms DALLE-2 and other models and produces incredibly realistic images.
Lung Cancer Detection on CT Images by using Image Processingijtsrd
This project is mainly based on image processing technique. In this work MATLAB have been used through every procedure made. Image processing techniques are widely use in bio-medical sector. The objective of our work is noise removal operation, thresholding, gray scale imaging, histogram equalization, texture segmentation, and morphological operation. Detection of lung cancer from computed tomography (CT) images is done by using MATLAB software. By using these methods the work has been done on CT images and the final tumor area has been shown with pixel values. Bindiya Patel | Dr. Pankaj Kumar Mishra | Prof. Amit Kolhe"Lung Cancer Detection on CT Images by using Image Processing" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-2 | Issue-3 , April 2018, URL: http://www.ijtsrd.com/papers/ijtsrd11674.pdf http://www.ijtsrd.com/engineering/electronics-and-communication-engineering/11674/lung-cancer-detection-on-ct-images-by-using-image-processing/bindiya-patel
[Paper] Multiscale Vision Transformers(MVit)Susang Kim
multiscale feature hierarchies with the transformer model.
a multiscale pyramid of features with early layers operating at high spatial resolution to model simple low-level visual information visual pathway with a hierarchical model
COM2304: Introduction to Computer Vision & Image Processing Hemantha Kulathilake
At the end of this lesson, you should be able to;
Describe image.
Describe digital image processing and computer vision.
Compare and Contrast image processing and computer vision.
Describe image sources.
Identify the array of imaging application under the EM Image source.
Describe the components of Image processing system and computer vision system.
Anatomy of the posterior cerebral circulation from the neuroradiology point of view. Anatomy of the vertebral artery. Anatomy of the basilar artery. Important for Neuroradiologists and Neurointerventionalists.
Imagen: Photorealistic Text-to-Image Diffusion Models with Deep Language Unde...Vitaly Bondar
A presentation about a new Google Research paper in the text-to-image task - Imagen.
This latent diffusion-based model outperforms DALLE-2 and other models and produces incredibly realistic images.
Lung Cancer Detection on CT Images by using Image Processingijtsrd
This project is mainly based on image processing technique. In this work MATLAB have been used through every procedure made. Image processing techniques are widely use in bio-medical sector. The objective of our work is noise removal operation, thresholding, gray scale imaging, histogram equalization, texture segmentation, and morphological operation. Detection of lung cancer from computed tomography (CT) images is done by using MATLAB software. By using these methods the work has been done on CT images and the final tumor area has been shown with pixel values. Bindiya Patel | Dr. Pankaj Kumar Mishra | Prof. Amit Kolhe"Lung Cancer Detection on CT Images by using Image Processing" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-2 | Issue-3 , April 2018, URL: http://www.ijtsrd.com/papers/ijtsrd11674.pdf http://www.ijtsrd.com/engineering/electronics-and-communication-engineering/11674/lung-cancer-detection-on-ct-images-by-using-image-processing/bindiya-patel
Karl Fristonが提唱している「自由エネルギー原理(free-energy principle = FEP)」について、北大文学部の聴衆を対象にして、物理学や機械学習の知識の前提抜きにして、説明を行い、その意義を説明したものです。FEPの意識研究への応用に向けて、FEPとエナクション説の近接性について強調したものとなっております。
Argues that the qualitative character distinctive of conscious perceiving also occurs without being conscious, so that being conscious is not essential to such qualitative character.
Living and Learning with the Brain in MindMioara Iacob
SAPTAMANA ALTFEL 2016 IN PARTENETIAT CU SCOALA GIMNAZIALA 150, BUCURESTI
Proiectul "From Brain Walking to Brain Jogging" (De la rezultate satisfacatoare, la performanta scolara) – eveniment organizat in cadrul programului ”Să știi mai multe, să fii mai bun!”, desfasurat in perioada 18 – 22 aprilie 2016 – cuprinde activitati educationale de informare si invatare a unor tehnici de studiu bazate pe modelul de functionare a creierului (brain-based learning techniques) ca o completare a metodelor traditionale.
Activitatile imbina exercitii de observare, reflectie, luare a deciziilor si rezolvare creativa a problemelor prin jocuri lingvistice si teste/ quiz-uri. Se urmareste dezvoltarea inteligentei emotionale si a constientizarii proceselor de invatare pentru a atinge urmatoarele obiective: gandire critica, comunicare, colaborare si creativitate. In cadrul acestor activitati vom cauta raspunsuri la cateva intrebari:
- Sunt resursele web indispensabile in procesul invatarii? Este posibila optimizarea abilitatilor cognitive cu ajutorul jocurilor video? Tehnologia – prieten sau dusman?
- Sunt copiii generatiei digitale mai destepti decat generatiile anterioare? Desi invata mai repede, cum se explica rezultatele slabe ale unora si esecul total al altora?
- Exista stiluri de invatare bazate pe tipul de inteligenta care ii caracterizeaza pe copii?
- Abilitati cognitive – tehnici de optimizare a memoriei, atentiei si a functiei executive a creierului.
Concluzii:
Creierul are o capacitate uriasa de a se adapta, schimba si a se optimiza, deci orice elev are sansa de a recupera si de a face performanta.
Fiecare dintre noi suntem responsabili pentru modul si tipul de performanta pentru care „ne antrenam” creierul.
Functia executiva a creierului – cheia eliminarii stresului in procesul de invatare si a eficientizarii acestuia.
Mindfulness training exercises from Alchemy of Love Mindfulness Training Books: Mindful Being, Conscious Parenting, Mindful Eating, Chanting mantras with best chords, Ama Alchemy of Love. About meditation, living our highest potential, self-remembering and mindfulness, willpower exercises, new ways of thinking, how to nurture creative thinking, practice focus with love, train emotional intelligence.
Presentation by Hiten Shah at HustleCon 2014. The talk is three things that both non-technical and technical founders can work on constantly when they aren't working on "making" things.
Past, Present and Future of AI: a Fascinating Journey - Ramon Lopez de Mantar...PAPIs.io
Possibly the most important lesson we have learned after 60 years of AI research is that what seemed to be very difficult to achieve, such as accurate medical diagnosis to playing chess at the level of a Grand Master, turned out to be relatively easy whereas what seemed easy, such as visual object recognition or deep language understanding, turned out to be extremely difficult. In my talk I will try to explain the reasons for this apparent contradiction by briefly reviewing the past and present of AI and projecting it into the near future.
Ramon Lopez de Mantaras is Research Professor of the Spanish National Research Council (CSIC) and Director of the Artificial Intelligence Research Institute of the CSIC. Technical Engineer EE (Electrical Engineering) from the Technical Engineering School of Mondragón (Spain) in 1973. Master of Sciences in Automatic Control from the University of Toulouse III (France) in 1974, Ph.D. in Physics from the University of Toulouse III (France), in 1977, with a thesis in Robotics (done at LAAS, CNRS). Master of Science in Engineering (ComputerScience) from the University of California at Berkeley (USA) in 1979. Ph.D. in Computer Science, from the Technical University of Catalonia, Barcelona (Spain) in 1981.
Helping Darwin: How to think about evolution of consciousness (Biosciences ta...Aaron Sloman
ABSTRACT
Many of Darwin's opponents, and some of those who accepted the theory of evolution as regards physical forms, objected to the claim that human mental functions, and
consciousness in particular, could be products of evolution. There were several reasons for this opposition, including unanswered questions as to how physical mechanisms could produce mental states and processes an old, and still surviving, philosophical problem.
A new answer is now available. Evolution could have produced the "mysterious" aspects of consciousness if, like engineers developing computing systems in the last six or seven decades, evolution encountered and "solved" increasingly complex problems of representation and control (including self-monitoring and self-control) by using systems with increasingly abstract mechanisms based on virtual machines, including most
recently self-monitoring virtual machines.
These capabilities are, like many capabilities of computer-based systems, implemented in non-physical virtual machinery which, in turn, are implemented in lower level physical mechanisms.
This would require far more complex virtual machines than human engineers have so far created. Noone knows whether the biological virtual machines could have been
implemented in the discrete-switch technology used in current computers.
These ideas were not available to Darwin and his contemporaries: most of the concepts, and the technology, involved in creation and use of sophisticated virtual machines were developed only in the last half century, as a by-product of a large number of design decisions by hardware and software engineers solving different problems.
Why symbol-grounding is both impossible and unnecessary, and why theory-tethe...Aaron Sloman
Introduction to key ideas of semantic models,
implicit definitions and symbol tethering, using ideas from philosophy of science and model theoretic semantics to explain why symbol ground theory is misguided: there is no need for all symbols used by an intelligent agent to be 'grounded' in terms of experience, or sensory-motor patterns. Rather, most of the meaning of a symbol may come from its role in a powerful explanatory theory, though the theory should have some connection with experiments and observations in order to be applicable to the world. That is not the same as requiring every symbol to be linked to experiences, experiments or measurements.
Symbol grounding theory is a modern version of the philosophical theory of 'concept empiricism', which was refuted by the philosopher Immanuel Kant in the 18th century.
Ontologies for baby animals and robots From "baby stuff" to the world of adul...Aaron Sloman
In contrast with ontology developers concerned with a symbolic or digital environment (e.g. the internet), I draw attention to some features of our 3-D spatio-temporal environment that challenge young humans and other intelligent animals and will also challenge future robots. Evolution provides most animals with an ontology that suffices for life, whereas some animals, including humans, also have mechanisms for substantive ontology extension based on results of interacting with the environment. Future human-like robots will also need this. Since pre-verbal human children and many intelligent non-human animals, including hunting mammals, nest-building birds and primates can interact, often creatively, with complex structures and processes in a 3-D environment, that suggests (a) that they use ontologies that include kinds of material (stuff), kinds of structure, kinds of relationship, kinds of process (some of which are process-fragments composed of bits of stuff changing their properties, structures or relationships), and kinds of causal interaction and (b) since they don't use a human communicative language they must use information encoded in some form that existed prior to human communicative languages both in our evolutionary history and in individual development. Since evolution could not have anticipated the ontologies required for all human cultures, including advanced scientific cultures, individuals must have ways of achieving substantive ontology extension. The research reported here aims mainly to develop requirements for explanatory designs. The attempt to develop forms of representation, mechanisms and architectures that meet those requirements will be a long term research project.
Do Intelligent Machines, Natural or Artificial, Really Need Emotions?Aaron Sloman
(Updated on 14 Jan 2014 -- with substantial revisions.)
Many people believe that emotions are required for intelligence. I argue that this is mostly based on (a) wishful thinking and (b) a failure adequately to analyse the variety of types of affective states and processes that can arise in different sorts of architectures produced by biological evolution or required for artificial systems. This work is a development of ideas presented by Herbert Simon in the 1960s in his 'Motivational and emotional controls of cognition'.
Complexity and Context-Dependency (version for Bath IOP Seminar)Bruce Edmonds
Extended version of my Complexity and Context-Dependency Talk given at the IOP seminar on "Complexity of Complexity" in Bath 19th Dec 2011.
This version talks more about looking for context in data.
Reorganised several times since first uploaded: most recently 25 Jan 2016
-------------------------------------------------------------------------------------------------------
Slides include link to video of lecture (158MB) http://www.cs.bham.ac.uk/research/projects/cogaff/movies/#ailect2-2015
-------------------------------------------------------------------------------------------------------------
Two questions are shown to have deep connections: What are the functions of vision in animals? and How did human languages evolve? The answer given here is that the functions of vision need to be supported by richly structured internal languages (forms of representation used for acquiring, storing, manipulating, deriving and using information), from which it follows that internal languages must have evolved before languages for communication.
---------------------------------------------------------------------------------------------------------------
The account of the functions of vision mentions early AI vision, the impact of Marr and the even greater impact of Gibson, but argues that they did not recognize all the functions of vision, e.g. the uses of vision in making mathematical discoveries leading to Euclid's elements.
---------------------------------------------------------------------------------------------------------------
Many questions are left unanswered by this research, which is part of the Meta-Morphogenesis project, introduced here:
http://www.cs.bham.ac.uk/research/projects/cogaff/misc/meta-morphogenesis.html
---------------------------------------------------------------------------------------------------------------
A slideshare presentation on "origins of language" by Jasmine Wong, adds some useful additional evidence, but presents a simpler theory:
http://www.slideshare.net/JasmineWong6/origins-of-language
---------------------------------------------------------------------------------------------------------------
Minor corrections+ additions 30-Mar-2015, 1-Apr-2015, 15-Apr-2015 12-Nov-2015
Mario TamaGetty Images NewsGetty ImagesLearning Objectiv.docxendawalling
Mario Tama/Getty Images News/Getty Images
Learning Objectives
After reading this chapter, you should be
able to:
• Explain how Tolman’s concept of latent learn-
ing sparked a need for theoretical accounts
that went beyond basic classical and operant
conditioning models.
• Describe the beginning of the cognitive revo-
lution and the importance of George Miller’s
memory studies.
• Explain how the computer has been used as
a metaphor for the way the mind processes
information.
• Describe how the cognitive counterrevolution
was different from the behavioral revolution
before it.
• List and define some of the foundational
concepts of cognitive approaches.
• Characterize George Kelly’s personal construct
theory and his view of humans as scientists.
Cognitive and Cognitive-Behavioral
Approaches to Personality 6
Chapter Outline
Introduction
6.1 Metacognition: Thinking About Your
Thinking
6.2 Major Historical Figures and Theories
• Edward Tolman and Latent Learning
• Chomsky and the Role of Language in
Cognition
• George Miller: The Cognitive Revolution
• Von Neumann and McCulloch: A Conceptual
Leap From Computers to the Mind
• Describe the contributions of the social learning theorists Julian Rotter, Albert Bandura, and Walter Mischel and
their contributions to such concepts as reward value, behavioral expectancies, self-efficacy, self-regulation, self-verification,
and motivated reasoning.
• Describe the process of modeling in both human and animal models.
• Compare and contrast the cognitive theories of Albert Ellis, Aaron Beck, and Marsha Linehan.
• Name and briefly describe some of the more commonly employed assessment tools used in the cognitive perspective.
Lec81110_06_c06_163-194.indd 163 5/21/15 12:39 PM
CHAPTER 6
Introduction
“There are no facts, only interpretations.” F. Nietzche
It’s December 14, 2012, at approximately 9:30 a.m. It’s an otherwise normal day
at Sandy Hook Elementary School in the small town of Newtown, Connecticut.
Then, 20-year-old Adam Lanza arrives on the scene carrying an arsenal of high-
capacity weapons, and he proceeds to gun down 20 helpless children and 6 adults
and terrorize many more. Authorities would later discover that Lanza had also
shot and killed his mother prior to coming to the school. Unfortunately, the mas-
sacre was not an isolated incident in American history, but it was the second dead-
liest school shooting in American history. Some claim that this incident points to
the disturbed nature of some individuals. Others point to the need for stricter gun
control to limit access to high-capacity weapons. Still others suggest that violence
in the media or video games were to blame. What was your take on these events?
If the goal were to predict your behavioral response to this event, could we do
so simply by understanding the environment in which you were brought up in
and currently live, or would we also need to understand how you cognitively pro-.
Mario TamaGetty Images NewsGetty ImagesLearning Objectiv.docxalfredacavx97
Mario Tama/Getty Images News/Getty Images
Learning Objectives
After reading this chapter, you should be
able to:
• Explain how Tolman’s concept of latent learn-
ing sparked a need for theoretical accounts
that went beyond basic classical and operant
conditioning models.
• Describe the beginning of the cognitive revo-
lution and the importance of George Miller’s
memory studies.
• Explain how the computer has been used as
a metaphor for the way the mind processes
information.
• Describe how the cognitive counterrevolution
was different from the behavioral revolution
before it.
• List and define some of the foundational
concepts of cognitive approaches.
• Characterize George Kelly’s personal construct
theory and his view of humans as scientists.
Cognitive and Cognitive-Behavioral
Approaches to Personality 6
Chapter Outline
Introduction
6.1 Metacognition: Thinking About Your
Thinking
6.2 Major Historical Figures and Theories
• Edward Tolman and Latent Learning
• Chomsky and the Role of Language in
Cognition
• George Miller: The Cognitive Revolution
• Von Neumann and McCulloch: A Conceptual
Leap From Computers to the Mind
• Describe the contributions of the social learning theorists Julian Rotter, Albert Bandura, and Walter Mischel and
their contributions to such concepts as reward value, behavioral expectancies, self-efficacy, self-regulation, self-verification,
and motivated reasoning.
• Describe the process of modeling in both human and animal models.
• Compare and contrast the cognitive theories of Albert Ellis, Aaron Beck, and Marsha Linehan.
• Name and briefly describe some of the more commonly employed assessment tools used in the cognitive perspective.
Lec81110_06_c06_163-194.indd 163 5/21/15 12:39 PM
CHAPTER 6
Introduction
“There are no facts, only interpretations.” F. Nietzche
It’s December 14, 2012, at approximately 9:30 a.m. It’s an otherwise normal day
at Sandy Hook Elementary School in the small town of Newtown, Connecticut.
Then, 20-year-old Adam Lanza arrives on the scene carrying an arsenal of high-
capacity weapons, and he proceeds to gun down 20 helpless children and 6 adults
and terrorize many more. Authorities would later discover that Lanza had also
shot and killed his mother prior to coming to the school. Unfortunately, the mas-
sacre was not an isolated incident in American history, but it was the second dead-
liest school shooting in American history. Some claim that this incident points to
the disturbed nature of some individuals. Others point to the need for stricter gun
control to limit access to high-capacity weapons. Still others suggest that violence
in the media or video games were to blame. What was your take on these events?
If the goal were to predict your behavioral response to this event, could we do
so simply by understanding the environment in which you were brought up in
and currently live, or would we also need to understand how you cognitively pro-.
Chapter 8 Thought andLanguage8.1 The Organization ofJinElias52
Chapter 8 Thought and
Language
8.1 The Organization of Knowledge
Concepts and Categories 315
Working the Scientific Literacy Model: Priming and Semantic Networks
318
Module 8.1a Quiz 319
Memory, Culture, and Categories 319
Module 8.1b Quiz 323
Module 8.1 Summary 323
8.2 Problem Solving, Judgment, and Decision Making
Defining and Solving Problems 325
Module 8.2a Quiz 328
Judgment and Decision Making 328
Working the Scientific Literacy Model: Maximizing and Satisficing in
Complex Decisions 332
Module 8.2b Quiz 334
Module 8.2 Summary 335
8.3 Language and Communication
What Is Language? 337
Module 8.3a Quiz 341
The Development of Language 341
Module 8.3b Quiz 344
Genes, Evolution, and Language 344
Working the Scientific Literacy Model: Genes and Language 344
Module 8.3c Quiz 348
Module 8.3 Summary 348
Module 8.1 The Organization of
Knowledge
Dmitry Vereshchagin/Fotolia
Learning Objectives
Know . . . the key terminology associated with concepts and categories.
Understand . . . theories of how people organize their knowledge about
the world.
Understand . . . how experience and culture can shape the way we
organize our knowledge.
Apply . . . your knowledge to identify prototypical examples.
8.1a
8.1b
8.1c
8.1d
When Edward regained consciousness in the hospital, his family
immediately noticed that something was wrong. The most obvious
problem was that he had difficulty recognizing faces, a relatively common
disorder known as prosopagosia. As the doctors performed more testing,
it became apparent that Edward had other cognitive problems as well.
Edward had difficulty recognizing objects—but not all objects. Instead, he
couldn’t distinguish between different vegetables even though he could
use language to describe their appearance. His ability to recognize most
other types of objects seemed normal.
Neurological patients like Edward may seem unrelated to your own life.
However, for specific categories of visual information to be lost, they must
have been stored in similar areas of the brain before brain damage
occurred. Therefore, these cases give us some insight into how the brain
stores and organizes the information that we have encoded into memory.
Focus Questions
1. How do people form easily recognizable categories from complex
information?
2. How does culture influence the ways in which we categorize
information?
Each of us has amassed a tremendous amount of knowledge in the course of
our lifetime. Indeed, it is impossible to put a number on just how many facts each
of us knows. Imagine trying to record everything you ever learned about the
world—how many books could you fill? Instead of asking how much we know,
psychologists are interested in how we keep track of it all. In this module, we will
explore what those processes are like and how they work. We will start by
learning about the key terminology before presenting theories about how
Analyze . . . the claim that the language we speak determines how we
t ...
Construction kits for evolving life -- Including evolving minds and mathemati...Aaron Sloman
Darwin's theory of evolution by natural selection does not adequately explain the generative power of biological evolution. For that we need to understand the mechanisms involved in producing new options for natural selection, without which there would always be the same set of possibilities available. This applies also to the construction kits: evolution can produce new construction kits, "Derived" construction kits, based on the Fundamental construction kit provided by the physical universe and its originally lifeless physical and chemical mechanisms. It turns out that life needs both concrete and abstract construction kits, of ever increasing complexity. This paper introduces some basic ideas, though far more empirical and theoretical research is required, combining multiple disciplines. Slideshare no longer allows presentations to be updated, so I no longer use it. For a later version search for: Sloman "Construction kits for evolving life" cogaff. Most of my slideshare presentations have newer versions in the CogAff web site at the University of Birmingham, UK. (Not Alabama)
The Turing Inspired Meta-Morphogenesis Project -- The self-informing universe...Aaron Sloman
This replaces an earlier version. The latest version with clickable links is available at Versions with clickable links available at http://www.cs.bham.ac.uk/research/projects/cogaff/misc/meta-morphogenesis.html
Virtuality, causation and the mind-body relationshipAaron Sloman
Extends my previous introductions to virtual machines and their role both in artefacts and products of biological evolution. This attempts to correct various erroneous assumptions about computation, functionalism, supervenience, life, information, and causation. See also http://www.cs.bham.ac.uk/research/projects/cogaff/misc/vm-functionalism.html
How to Build a Research Roadmap (avoiding tempting dead-ends)Aaron Sloman
What's a Research Roadmap For?
Why do we need one?
How can we avoid the usual trap of making bold promises to do X, Y and Z,
then hope that our previous promises will not be remembered the next time we apply for funds to do X, Y and Z?
How can we produce a sensible, well informed roadmap?
Originally presented at the euCognition Research Roadmap discussion in Munich on 12 Jan 2007
This suggests a way to avoid tempting dead ends (repeating old promises that proved unrealistic) by examining many long term goals, including describing existing human and animal competences not yet achieved by robots, then working backwards systematically by investigating requirements for those competences, and requirements for meeting those requirements, etc. Insread of generating a single linear roadmap this should produce a partially ordered network of intermediate targets, leading back, to short term goals that may be achievable starting from where we are.
Such a roadmap will inevitably have mistakes: over-optimistic goals, missing preconditions, unrecognised opportunities. But if the work is done in many teams in a fully open manner with as much collaboration as possible, it should be possible to make faster, deeper, progress than can be achieved by brain-storming discussions of where we can get in a few years.
If learning maths requires a teacher, where did the first teachers come from?
or
Why (and how) did biological evolution produce mathematicians?
Presentation at Symposium on Mathematical Cognition AISB2010
Part of the Meta-Morphogenesis Project. See also this discussion of toddler theorems:
http://www.cs.bham.ac.uk/research/projects/cogaff/misc/toddler-theorems.html
Evolution of human mathematics from earlier abilities to perceived, use and reason about affordances, spatial possibilities and constraints.
The necessity of mathematical truth does not imply infallibility of mathematical reasoning. (Lakatos).
Toddlers discover theorems without knowing it. Later they may learn to reflect on and talk about what they have learnt. Compare Annette Karmiloff-Smith on "Representational re-description".
Why is it still so hard to give robots and AI systems the ability to reason spatially as mathematicians do (except for simple special cases, e.g. where space is discretised.)
A multi-picture challenge for theories of visionAaron Sloman
(Modified 7th June 2013 to include some droodles.)
Some informal experiments are presented whose results help to challenge most theories of vision and proposed mechanisms of vision.
A possible explanatory information-processing architecture is proposed, based on multiple dynamical systems, grown during an individual's life time, most of which are dormant most of the time, but which can be very rapidly activated and instantiated so as to build a multi-ontology interpretation of the currently, and recently, available visual information -- e.g. turning a corner into a busy street in an unfamiliar city. As far as I know, there is no working implementation of such a system, though a very early prototype called Popeye (implemented in Pop2) around 1976 is summarised. Many hard unsolved problems remain, though most of them are ignored by research on vision that makes narrow assumptions about the functions of biological vision.
Meta-Morphogenesis, Evolution, Cognitive Robotics and Developmental Cognitive...Aaron Sloman
How could a planet, condensed from a cloud of dust, produce minds -- and products of minds, along with microbes, mice, monkeys, mathematics, music, marmite, murder, megalomania, and all other forms and products of life on earth (and possibly elsewhere)?
This presentation introduces the ambitious, multi-disciplinary Meta-Morphogenesis project, partly inspired by Turing's 1952 paper on morphogenesis. It may lead to an answer, by identifying the many transitions between different types and mechanisms of biological information processing, including transitions that changed the mechanisms of change, altering forms of evolution, development, learning, culture and ecosystem dynamics. One of the questions raised is whether chemical information-processing is capable of supporting processes that would be infeasible or impossible on a Turing machine or conventional computer.
A 2hour 30 min recording of this tutorial was made by Adam Ford, available here: http://www.youtube.com/watch?v=BNul52kFI74 (new version installed on 14 Jun 2013 with titles and audio problem fixed). Also available here
http://www.cs.bham.ac.uk/research/projects/cogaff/movies/#m-m-tut
"Information" here is used in Jane Austen's sense, not Claude Shannon's sense. See http://www.cs.bham.ac.uk/research/projects/cogaff/misc/austen-info.html
More information about the project is available here: http://www.cs.bham.ac.uk/research/projects/cogaff/misc/meta-morphogenesis.html
Adam Ford interviewed the author about some of these topics at the AGI conference in December 2012 in this video: http://www.youtube.com/watch?v=iuH8dC7Snno
Related PDF presentations can be found here http://www.cs.bham.ac.uk/research/projects/cogaff/talks
What is computational thinking? Who needs it? Why? How can it be learnt? ...Aaron Sloman
What is computational thinking?
Who needs it? Why? How can it be learnt?
Can it be taught? How?
Slides for invited presentation at Conference of ALT (Association for Learning Technology) 11th Sept 2012, University of Manchester.
PDF available (easier for printing, selecting text, etc.):
http://www.cs.bham.ac.uk/research/projects/cogaff/talks/#talk105
A video of the actual presentation (using no slides because of a projector problem) is now available here
http://www.youtube.com/watch?v=QXAFz3L2Qpo
It also has been made available as "slide 47" after the PDF presentation on this page.
I attempt to generalise Jeannette Wing's notion of "Computational thinking" (ACM 2006) to include attempting to understand much biological information processing, and try to show the necessity for educators to do deep computational thinking if they wish to facilitate processes of learning.
What's vision for, and how does it work? From Marr (and earlier)to Gibson and...Aaron Sloman
ABSTRACT
Very many researchers assume that it is obvious what vision (e.g. in humans) is for, i.e. what functions it has, leaving only the problem of explaining how those functions are fulfilled. So they postulate mechanisms and try to show how those mechanisms can produce the required effects, and also, in some cases, try to show that those postulated mechanisms exist in humans and other animals and perform the postulated functions. The main point of this presentation is that it is far from obvious what vision is for - and J.J. Gibson's main achievement is drawing attention to some of the functions that other researchers had ignored. I'll present some of the other work, show how Gibson extends and improves it, and then point out much more there is to the functions of vision and other forms of perception than even Gibson had noticed.
In particular, much vision research, unlike Gibson, ignores vision's function in on-line control and perception of continuous processes; and nearly all, including Gibson's work, ignores meta-cognitive perception, and perception of possibilities and constraints on possibilities and the associated role of vision in reasoning. If we don't understand that we cannot understand how biological mechanisms arising from requirements for being embodied in a rich, complex and changing 3-D environment underpin human mathematical capabilities, including the ability to reason about topology and Euclidean geometry.
Last updated: 1st March 2014, 10 June 2015 (additional links)
Slides prepared for a broadcast presentation to members of Computing at School http://www.computingatschool.org.uk/, about why computing education should be about more than the science and technology required for useful or entertaining applications. Instead, learning about forms of information processing systems can give us new, deeper ways of thinking about many old phenomena, e.g. the nature of mind and the evolution of minds of various kinds. This supports the claim that the study of computation is as much a science as physics or psychology, rather than just a branch of engineering -- as famously suggested by Fred Brooks.
Possibilities between form and function (Or between shape and affordances)Aaron Sloman
I discuss the need for an intelligent system, whether it is a robot, or some sort of digital companion equipped with a vision system, to include in its ontology a range of concepts that appear not to have been noticed by most researchers in robotics, vision, and human psychology. These are concepts that lie between (a) concepts of "form", concerned with spatially located objects, object parts, features, and relationships and (b) concepts of affordances and functions, concerned with how things in the environment make possible or constrain actions that are possible for a perceiver and which can support or hinder the goals of the perceiver.
Those intermediate concepts are concerned with processes that *are* occurring and processes that *can* occur, and the causal relationships between physical structures/forms/configurations and the possibilities for and constraints on such processes, independently of whether they are processes involving anyone's actions or goals.
These intermediate concepts relate motions and constraints on motion to both geometric and topological structures in the environment and the kinds of 'stuff' of which things are composed, since, for example, rigid, flexible, and fluid stuffs support and constrain different sorts of motions.
They underlie affordance concepts. Attempts to study affordances without taking account of the intermediate concepts are bound to prove shallow and inadequate.
Notes for invited talk at Dagstuhl Seminar: ``From Form to Function'' Oct 18-23, 2009 http://www.dagstuhl.de/en/program/calendar/semhp/?semnr=09431
Virtual Machines and the Metaphysics of Science Aaron Sloman
Philosophers regularly use complex (running) virtual machines (not virtual realities) composed of enduring interacting non-physical subsystems (e.g. operating systems, word-processors, email systems, web browsers, and many more). These VMs can be subdivided into different kinds with different types of functions, e.g. "specific-function VMs" and "platform VMs" (including language VMs, and operating system VMs) that provide support for a variety of different (possibly concurrent) "higher level" VMs, with different functions.
Yet, almost all ignore (or misdescribe) these VMs when discussing functionalism, supervenience, multiple realisation, reductionism, emergence, and causation.
Such VMs depend on many hardware and software designs that interact in very complex ways to maintain a network of causal relationships between physical and virtual entities and processes.
I'll try to explain this, and show how VMs are important for philosophy, in part because evolution long ago developed far more sophisticated systems of virtual machinery (e.g. running on brains and their surroundings) than human engineers so far. Most are still not understood.
This partly accounts for the apparent intractability of several philosophical problems.
E.g. running VM subsystems can be disconnected from input-output interactions for extended periods, and some can have more complexity than the available input/output bandwidth can reveal.
Moreover, despite the advantages of VMs for self-monitoring and self control, they can also lead to self-deception.
For a lot of related material see Steve Burbeck's web site http://evolutionofcomputing.org/Multicellular/Emergence.html
(A related presentation debunking the "hard problem" of consciousness is also in this collection.)
What is science? (Can There Be a Science of Mind?) (Updated August 2010)Aaron Sloman
This presentation gives an introduction to philosophy of science, though a rather idiosyncratic one, stressing science as the search for powerful new ontologies rather than merely laws. You can't express a law unless you have
an ontology including the items referred to in the law (e.g. pressure, volume, temperature). The talk raises a
number of questions about the aims and methods of science, about the differences between the physical sciences and
the science of information-processing systems (e.g. organisms, minds, computers), whether there is a unique truth
or final answers to be found by science, whether scientists ever prove anything (no -- at most they show that some
theory is better than any currently available rival theory), and why science does not require faith (though
obstinacy can be useful). The slides end with a section on whether a science of mind is possible, answering yes, and explaining how.
Distinguishes Humean (statistics-based) notions of causation and Kantian (deterministic, structure-based) notions of causation, arguing that intelligent robots and animals need both, but each requires a combination of competences, and various kinds of partial competence of both kinds are possible.
What designers of artificial companions need to understand about biological onesAaron Sloman
Discusses some of the requirements for artificial companions (ACs), contrasting relatively easy (Type 1) goals and much harder, but more important, (Type 2) goals for designers of ACs. Current techniques will not lead to achieving Type 2 goals, but progress can be made towards those goals, by looking more closely at how the relevant competences develop in humans, and their architectural and representational requirements. A human child develops a lot of knowledge about space, time, causation, and many varieties of 3-D structures and processes. This knowledge provides the background and foundation for many kinds of learning, and will be required for development of ACs that act sensibly and give sensible advice. No AI systems are anywhere near the competences of young children. Trying to go directly to systems with adult competences, especially statistics based systems, will produce systems that are either very restricted, or very brittle and unreliable (or both).
Kantian Philosophy of Mathematics and Young Robots: Could a baby robot grow u...Aaron Sloman
There is a sequel to this, with more emphasis on 'toddler theorems' and kinds of child science here:
http://www.cs.bham.ac.uk/research/projects/cogaff/talks/#toddler
It is not yet stable enough to be uploaded to slideshare.
Evolution of minds and languages: What evolved first and develops first in ch...Aaron Sloman
SLIDESHARE NOW STUPIDLY DOES NOT ALLOW SLIDES TO BE UPDATED. To find the latest version of these slides go to http://www.cs.bham.ac.uk/research/projects/cogaff//talks/#talk111
The version posted here was last updated on 16 March 2015. There have been several changes since then on the alternative site. Why did Slideshare take such a stupid decision (after being bought by Linkedin?)
A theory is presented according to which "languages" with structural variability and compositional semantics evolved in several species for *internal* use (e.g. in perception, planning, learning, forming goals, deciding, etc.) before *external* languages evolved for communication. The theory implies that such internal languages develop in young humans before a language for communication.
It is is also noted that the standard notion of 'compositional semantics' has to allow for the propagation of semantic content from parts to wholes to be potentially context sensitive at every stage: i.e. current context, speaker intentions, user knowledge, shared goals, can all affect how semantics of larger parts are derived from semantics of smaller parts+syntactic structure. This applies as much to non-verbal languages as to verbal ones.
This theory of how human languages evolved from earlier 'internal languages' (GLs) is inconsistent with the best known published theories of evolution or development of language.
But that does not make it wrong. Moreover, this theory is supported by empirical evidence including the example of deaf children in Nicaragua: http://en.wikipedia.org/wiki/Nicaraguan_Sign_Language
Instructions for Submissions thorugh G- Classroom.pptxJheel Barad
This presentation provides a briefing on how to upload submissions and documents in Google Classroom. It was prepared as part of an orientation for new Sainik School in-service teacher trainees. As a training officer, my goal is to ensure that you are comfortable and proficient with this essential tool for managing assignments and fostering student engagement.
Biological screening of herbal drugs: Introduction and Need for
Phyto-Pharmacological Screening, New Strategies for evaluating
Natural Products, In vitro evaluation techniques for Antioxidants, Antimicrobial and Anticancer drugs. In vivo evaluation techniques
for Anti-inflammatory, Antiulcer, Anticancer, Wound healing, Antidiabetic, Hepatoprotective, Cardio protective, Diuretics and
Antifertility, Toxicity studies as per OECD guidelines
Introduction to AI for Nonprofits with Tapp NetworkTechSoup
Dive into the world of AI! Experts Jon Hill and Tareq Monaur will guide you through AI's role in enhancing nonprofit websites and basic marketing strategies, making it easy to understand and apply.
Palestine last event orientationfvgnh .pptxRaedMohamed3
An EFL lesson about the current events in Palestine. It is intended to be for intermediate students who wish to increase their listening skills through a short lesson in power point.
Francesca Gottschalk - How can education support child empowerment.pptxEduSkills OECD
Francesca Gottschalk from the OECD’s Centre for Educational Research and Innovation presents at the Ask an Expert Webinar: How can education support child empowerment?
Synthetic Fiber Construction in lab .pptxPavel ( NSTU)
Synthetic fiber production is a fascinating and complex field that blends chemistry, engineering, and environmental science. By understanding these aspects, students can gain a comprehensive view of synthetic fiber production, its impact on society and the environment, and the potential for future innovations. Synthetic fibers play a crucial role in modern society, impacting various aspects of daily life, industry, and the environment. ynthetic fibers are integral to modern life, offering a range of benefits from cost-effectiveness and versatility to innovative applications and performance characteristics. While they pose environmental challenges, ongoing research and development aim to create more sustainable and eco-friendly alternatives. Understanding the importance of synthetic fibers helps in appreciating their role in the economy, industry, and daily life, while also emphasizing the need for sustainable practices and innovation.
Model Attribute Check Company Auto PropertyCeline George
In Odoo, the multi-company feature allows you to manage multiple companies within a single Odoo database instance. Each company can have its own configurations while still sharing common resources such as products, customers, and suppliers.
Welcome to TechSoup New Member Orientation and Q&A (May 2024).pdfTechSoup
In this webinar you will learn how your organization can access TechSoup's wide variety of product discount and donation programs. From hardware to software, we'll give you a tour of the tools available to help your nonprofit with productivity, collaboration, financial management, donor tracking, security, and more.
June 3, 2024 Anti-Semitism Letter Sent to MIT President Kornbluth and MIT Cor...Levi Shapiro
Letter from the Congress of the United States regarding Anti-Semitism sent June 3rd to MIT President Sally Kornbluth, MIT Corp Chair, Mark Gorenberg
Dear Dr. Kornbluth and Mr. Gorenberg,
The US House of Representatives is deeply concerned by ongoing and pervasive acts of antisemitic
harassment and intimidation at the Massachusetts Institute of Technology (MIT). Failing to act decisively to ensure a safe learning environment for all students would be a grave dereliction of your responsibilities as President of MIT and Chair of the MIT Corporation.
This Congress will not stand idly by and allow an environment hostile to Jewish students to persist. The House believes that your institution is in violation of Title VI of the Civil Rights Act, and the inability or
unwillingness to rectify this violation through action requires accountability.
Postsecondary education is a unique opportunity for students to learn and have their ideas and beliefs challenged. However, universities receiving hundreds of millions of federal funds annually have denied
students that opportunity and have been hijacked to become venues for the promotion of terrorism, antisemitic harassment and intimidation, unlawful encampments, and in some cases, assaults and riots.
The House of Representatives will not countenance the use of federal funds to indoctrinate students into hateful, antisemitic, anti-American supporters of terrorism. Investigations into campus antisemitism by the Committee on Education and the Workforce and the Committee on Ways and Means have been expanded into a Congress-wide probe across all relevant jurisdictions to address this national crisis. The undersigned Committees will conduct oversight into the use of federal funds at MIT and its learning environment under authorities granted to each Committee.
• The Committee on Education and the Workforce has been investigating your institution since December 7, 2023. The Committee has broad jurisdiction over postsecondary education, including its compliance with Title VI of the Civil Rights Act, campus safety concerns over disruptions to the learning environment, and the awarding of federal student aid under the Higher Education Act.
• The Committee on Oversight and Accountability is investigating the sources of funding and other support flowing to groups espousing pro-Hamas propaganda and engaged in antisemitic harassment and intimidation of students. The Committee on Oversight and Accountability is the principal oversight committee of the US House of Representatives and has broad authority to investigate “any matter” at “any time” under House Rule X.
• The Committee on Ways and Means has been investigating several universities since November 15, 2023, when the Committee held a hearing entitled From Ivory Towers to Dark Corners: Investigating the Nexus Between Antisemitism, Tax-Exempt Universities, and Terror Financing. The Committee followed the hearing with letters to those institutions on January 10, 202
Honest Reviews of Tim Han LMA Course Program.pptxtimhan337
Personal development courses are widely available today, with each one promising life-changing outcomes. Tim Han’s Life Mastery Achievers (LMA) Course has drawn a lot of interest. In addition to offering my frank assessment of Success Insider’s LMA Course, this piece examines the course’s effects via a variety of Tim Han LMA course reviews and Success Insider comments.
Why the "hard" problem of consciousness is easy and the "easy" problem hard. (And how to make progress)
1. Talk to Language and Cognition Seminar, School of Psychology, UoB. 6 Nov 2009
Why the “hard” problem
of consciousness is easy
and the “easy” problem hard.
(And how to make progress)
Aaron Sloman
School of Computer Science, University of Birmingham
http://www.cs.bham.ac.uk/˜axs/
These PDF slides will be available in my ‘talks’ directory:
http://www.cs.bham.ac.uk/research/projects/cogaff/talks/#cons09
Lang Cog Sem Birmingham 2009 Slide 1 Last revised: January 27, 2010
2. Why? Because:
1. The “hard” problem can be shown to be a non-problem because it is formulated using
a seriously defective concept (explained later as the concept of “phenomenal
consciousness” defined so as to rule out cognitive functionality).
2. So the hard problem is an example of a well known type of philosophical problem that
needs to be dissolved (fairly easily) rather than solved.
For other examples, and a brief introduction to conceptual analysis, see
http://www.cs.bham.ac.uk/research/projects/cogaff/misc/varieties-of-atheism.html
3. In contrast, the so-called “easy” problem requires detailed analysis of very complex
and subtle features of perceptual processes, introspective processes and other mental
processes, sometimes labelled “access consciousness”: these have cognitive
functions, but their complexity (especially the way details change as the environment
changes or the perceiver moves) is considerable and very hard to characterise.
4. “Access consciousness” is complex also because it takes many different forms:
what individuals can be conscious of, and what functions being conscious has, varies hugely, from
simple life forms to sophisticated adult humans, and can vary between humans at different stages of
development from conception to senile dementia. The concept is highly polymorphic.
5. Finding ways of modelling these aspects of consciousness, and explaining how they
arise out of physical mechanisms, requires major advances in the science of
information processing systems – including computer science and neuroscience.
The remaining slides attempt to justify these claims
Lang Cog Sem Birmingham 2009 Slide 2 Last revised: January 27, 2010
3. Some of the topics included, not necessarily in this order
1. Philosophical (and psychological) background to the idea of ‘phenomenal consciousness’ and qualia as the introspectable
contents of consciousness.
2. Brief introduction to philosophers’ notions of supervenience/realisation/implementation.
3. Ned Block’s distinction between access consciousness (which has a functional role) and phenomenal consciousness (which
does not).
4. Why there seems to be an ‘explanatory gap’ (T.H.Huxley) or a ‘hard problem’ (Chalmers). Chalmers uses something close to
Block’s notion of phenomenal consciousness to pose the problem, but it is actually a very old problem.
[Incidentally the explanatory gap was used even by scientists sympathetic to Darwin, as an objection his theory of evolution as
applied to humans.]
5. Chalmers uses the gap to propose a research programme investigating ‘Neural correlates of consciousness’. I’ll try to show
why the notion of phenomenal consciousness (not access consciousness) is incoherent, as is the search for neural correlates.
6. Brief introduction to the technology of virtual machinery (not virtual reality) developed over the last six decades and its
philosophical relevance: processes and events in running virtual machines can be causes and effects, despite being fully
implemented in physical mechanisms. (But specific VM entities, events, and processes, need not have physical correlates.)
7. Explanation of the importance of virtual machines in sophisticated control systems with self-monitoring and self-modulating
capabilities. Why such machines need something like “access consciousness”/qualia – and why they too generate an
explanatory gap – a gap bridged by a lot of sophisticated hardware and software engineering developed over a long time.
Here the explanations are much deeper than mere correlations: we know how the physical and virtual machinery are related.
8. Conjecture that biological evolution discovered those design problems long before we did and produced solutions using virtual
machinery long before we did – in order to enable organisms to deal with rapidly changing and complex information structures
(e.g. in visual perception, decision making, control of actions, etc.).
You can’t rapidly rewire millions of neurons when you look in a new direction. So there’s no alternative to using virtual machinery?
9. Some deep philosophical problems about causation, and about meaning/intentionality, discussed briefly, using simple
(perhaps over-simple) examples.
9. Work still to be done, e.g. finding out precisely what the problems were that evolution solved and how they are solved in
organisms, and how things can go wrong in various ways, and why future intelligent robots will need similar solutions.
Lang Cog Sem Birmingham 2009 Slide 3 Last revised: January 27, 2010
4. Preamble (Historical caricature)
• The problem of explaining how mind and matter are related goes back at least two millenia.
http://plato.stanford.edu/entries/aristotle-psychology/
• The usefulness of talking about consciousness has been much debated by scientists,
During the 20th C. many (English speaking) psychologists tried to avoid talking about anything mental.
• Freud’s theories with his division between id, ego, and superego were disparaged
(though welcomed by many outside psychology departments).
• The rise of cognitive psychology and cognitive science since late 1960s, along with development of
computational theories began to change this.
• More hard-headed scientific psychologists started talking about emotions in the 1980s.
The journal Cognition and Emotion started in 1987. Now there are journals, workshops, projects, ...
• In the 1990s people started talking about “neural correlates of consciousness”(NCCs)
“Progress in addressing the mind-body problem has come from focusing on empirically accessible
questions rather than on eristic philosophical arguments. Key is the search for the neuronal
correlates - and ultimately the causes - of consciousness.” (Scholarpedia) (Compare Crick & Koch 1990)
see http://www.scholarpedia.org/article/Neural_correlates_of_consciousness
• Neural correlates of what??? Some respond: “Phenomenal consciousness”.
In 1995 Ned Block distinguished access consciousness (A-C) (introspective states with cognitive
functions) and phenomenal consciousness (P-C) (states without cognitive functions).
• David Chalmers The Conscious Mind: In Search of a Fundamental Theory (1996) claimed that there was a “hard”
problem of consciousness, namely explaining the existence of P-C in a physical world.
Explaining A-C was thought to be (relatively) easy by comparison.
• This encouraged many to start searching for NCCs as a way of attacking the “hard” problem. What is
the hard problem?
• Now many AI/Robotics papers are sprinkled with references to “emotion” and “consciousness” – mostly
undisciplined and muddled as far as I can see.
Lang Cog Sem Birmingham 2009 Slide 4 Last revised: January 27, 2010
5. Next few slides
• There are many empirical facts about human experience (most of them easily
checked) that support claims about the existence of introspectively accessible entities,
often described as privately accessible contents of consciousness.
Various labels are used for these entities: “phenomenal consciousness”, “qualia” (singular “quale”),
“sense-data”, “sensibilia”, “what it is like to be/feel X” (and others).
For a useful, but partial, overview see http://en.wikipedia.org/wiki/Qualia
• What is not clear is what exactly follows from the empirical facts, and how best they
can be described and explained.
• I start by indicating some of the standard philosophical problems and theories about
the status of the entities allegedly shown to exist.
• They pose a scientific problem of explaining how such entities arise and how they are
related to non-mental mechanisms, e.g. brains.
• I introduce and then criticise the so-called “hard problem”, which is a new formulation
of what was long ago called “the explanatory gap”
• Later I present a new way of dealing with the explanatory gap using the concept of
“running (or active) virtual machine”.
This concept has been largely ignored (or when not ignored, misunderstood) by philosophers and
psychologists even though they use examples every day when writing papers, reading and writing
research papers, using spelling checkers, spreadsheets, internet browsers, anti-virus packages, etc.
Lang Cog Sem Birmingham 2009 Slide 5 Last revised: January 27, 2010
6. Do some “consciousness science”
Stare at each of these two pictures for a while.
Each is ambiguous and should flip between (at least) two very different views.
Try to describe exactly what changes when the flip occurs.
What concepts are needed for the different experiences?
In one case geometrical relations and distances change. In the other case geometry is unchanged, but
biological functions change. Can a cube be experienced as “looking to left or to right”? If not, why not?
Nothing changes on the paper or in the optical information entering your eyes and brain.
Compare the kind of vocabulary used to describe parts and relationships in the two views
of the Necker cube, and in the two views of the “duck-rabbit”.
If the figures do not flip for you, ask a friend to try.
Lang Cog Sem Birmingham 2009 Slide 6 Last revised: January 27, 2010
7. Empirical basis for referring to contents of consciousness
Introspection provides empirical evidence for mental contents “inside” us,
distinct from external causes or internal physical processes:
• Ambiguous figures: as you stare at them, nothing changes out there, only the
experiences/qualia (in your mind) “flip” (E.g. Necker cube, face-vase, old/young lady, etc.)
This one rotates in 3-D in either direction: http://www.procreo.jp/labo/labo13.html
• Optical illusions (of many kinds): Muller-Lyer, Ebbinghaus, motion/colour after-effects.
• Dreams, hallucinations, hypnotic suggestions, effects of alcohol and other drugs.
• Different people see the same things differently. E.g. short-sighted and long-sighted people.
Identical twins look different to people who know them well, but look the same to others.
Cf. Botanical expertise makes similar plants look different. Colour blindness.
• Pressing eyeball makes things appear to move when they don’t, and can undo binocular fusion: you get
two percepts of the same object; crossing your eyes can also do that.
• Put one hand into a pail of hot water, the other into a pail of cold water, then put both into lukewarm
water: it will feel cool to one hand and warm to the other. (A very old philosophical experiment.)
• People and other things look tiny from a great height – without any real change in size.
• Aspect ratios, what’s visible, specularities, optical flow – all change with viewpoint.
• We experience only portions of things. A cube has six faces but we can’t see them all: what you
experience changes as you move.
• Thinking, planning, reminiscing, daydreaming, imagining, can be done with eyes closed ....
• Composing poems, or music, or proving theorems with your eyes shut.
Rich mental contents (not perceptual contents) can be involved in all of these.
Lang Cog Sem Birmingham 2009 Slide 7 Last revised: January 27, 2010
8. 2-D and 3-D Qualia (added 27 Jan 2010)
Here (on the right) is part of a picture by Swedish artist,
¨
Oscar Reutersvard (1934) which you probably see as a
configuration of coloured cubes.
As with the Necker cube you have experiences of both 2-D
lines, regions, colours, relationships and also 3-D surfaces,
edges, corners, and spatial relationships.
You probably also experience various affordances: places you could touch the surfaces, ways you could
grasp and move the various cubes (perhaps two are held floating in place by magnetic fields).
E.g. you can probably imagine swapping two of them, thinking about how you would have to grasp them in
the process – e.g. swapping the white one with the cube to the left of it, or the cube on the opposite side.
The second picture on the right (from which the first one was
extracted) has a richer set of 2-D and 3-D contents.
Again there is a collection of 2-D contents (e.g. a star in the middle),
plus experience of 3-D structures, relationships and affordances: with
new possibilities for touching surfaces, grasping cubes, moving cubes.
The picture is outside you, as would the cubes be if it were not a picture.
But the contents of your experience are in you: a multi-layered set of
qualia: 2-D, 3-D and process possibilities.
But the scene depicted in the lower picture is geometrically impossible,
even though the 2-D configuration is possible and exists, on the screen
or on paper, if printed: the cubes, however, could not exist like that.
So your qualia can be inconsistent!
Lang Cog Sem Birmingham 2009 Slide 8 Last revised: January 27, 2010
9. Implications of the empirical facts
When we perceive our environment or when we think, plan, daydream, reminisce,
compose in our heads, in addition to what’s “out there”, or what’s happening in our brains,
there are mental things going on in us, which we can attend to, think about, ask questions
about, and in some cases use, e.g. to gain information about the environment.
• The process of discovery of these internal entities mainly uses introspection.
Introspection is a kind of perception, directed inwardly, but is inherently “private”.
• Contents of introspection (what we experience as existing in us) are what discussions
of phenomenal consciousness and the hard problem of consciousness are supposed
to be about. Terminology varies: Qualia, sense-data, ideas, impressions, sensibilia, ...
• A science of mind needs to study of these phenomena, even though they are not
accessible by the normal methods of science: “public” observation and measurement.
• The internal states, events and processes (desires, decisions, impulses, “acts of
will”...) are also able to produce physical events and processes
e.g. movements, speech, saccades, smiling, weeping, ...
• The internal entities are distinct from physical matter in many ways:
e.g. they are not amenable to observation and measurement by physicists, chemists, etc. using the
tools of the physical sciences – they are only accessible by their owner.
Even if you could somehow be connected to my brain so as to have my experiences projected into
you, you would then be having your experiences through my senses – you would not be sharing my
experiences. You still could not know for certain, as I can, what my experiences actually are.
• Can we bridge the “explanatory gap”: between physical phenomena and contents of
consciousness, in either direction?
Lang Cog Sem Birmingham 2009 Slide 9 Last revised: January 27, 2010
10. Questions generated
Difference between mental contents and physical phenomena generate many questions.
Some examples:
• Epistemological:
– Can we infer the existence of an “external” world from our experiences?
– Can we infer the existence of “other minds” from our experiences
– Can we tell whether infants, other species, or robots, have them? ... etc....
• Metaphysical:
– What kinds of things exist and how are they related?
Can mind and matter influence each other? Can either exist without the other?
• Conceptual/semantic:
– How can we conceive of and think about things we don’t directly experience?
(e.g. the physical environment, other minds, causal connections, distant past or future events....)
– Is there some other way of acquiring concepts other than by abstraction from
experiences of instances?
Concept empiricism (recently reinvented as “symbol grounding theory”) says no. But Kant and
20th century philosophers of science showed that concept empiricism must be false: there must
be other ways of acquiring concepts, e.g. of gene, subatomic particle, remote galaxy, etc.
• Scientific:
– Can we explain how physical systems can produce mental contents?
E.g. what brain mechanisms could do it?
– Can we account for the evolution of minds with experiences in a physical world?
– Can we understand development of individual minds (e.g. from foetus to philosopher?)
Lang Cog Sem Birmingham 2009 Slide 10 Last revised: January 27, 2010
11. Answers suggested by various thinkers
A (partly) new collection of answers (some going back a very long time):
• Epistemological: (We need to remember our biological origins.)
– We don’t infer an “external” world: we evolved so as to assume that we exist in a
3-D space and able to find out what sorts of things are in it and how they work.
– Likewise evolution, not inference, makes us assume there are “other minds”.
What work is done by those two assumptions, and how, is another matter: discussed later.
• Metaphysical:
– Many kinds of things exist, with complex interactions: science keeps finding more.
• Conceptual/semantic:
– As Kant argued, you can’t have experiences without concepts, so concepts don’t all
come from experience.
Scientific concepts have to be introduced through the theories that use them. We are still learning
(very slowly) how to give machines such capabilities.
http://www.cs.bham.ac.uk/research/projects/cogaff/talks/#grounding
– Recently developed sets of concepts used in designing and implementing virtual
machines whose behaviours are very different from physical processes, give us
new conceptual tools for thinking about minds. (More on this later.)
• Scientific:
– By learning how to replicate the creation of biological minds as virtual machines
running on, and controlling, physical machines, we may become more ready to
answer the really hard scientific questions and close the explanatory gap.
Lang Cog Sem Birmingham 2009 Slide 11 Last revised: January 27, 2010
12. Plato’s cave
Plato’s cave analogy expresses the view that we are like prisoners in a
cave who do not see things in the world as they are, since all we have are
the shadows cast on a wall of the cave (the contents of consciousness).
http://home.earthlink.net/˜johnrpenner/Articles/PlatosCave.html
If people walk past the flame on the
raised pathway, their shadows will be
cast on the wall, and be seen by the
prisoners facing the wall.
The prisoners cannot turn their
heads to see the things that cast the
shadows.
This could be a metaphorical
description of the relationship
between our percepts and their
causes in the environment. Picture found in many locations on the internet
(But I am no Plato Scholar.)
But if the prisoners are scientists, they may be able to develop, and perhaps also test,
theories about the causes of the shadows.
E.g. explaining what happens to some shadows when they move their own heads?
Lang Cog Sem Birmingham 2009 Slide 12 Last revised: January 27, 2010
13. Older theories about mind-body relations
Various attempts have been made to describe the relations between:
(a) physical objects, including bodies, sense organs and brains, and
(b) mental entities, e.g. minds and their contents.
An incomplete sample:
• Monism: There is only one kind of stuff that exists
Monists differ as to whether the stuff is physical, mental, or something else:
– Idealism (everything is mental): to be is to be experienced (Berkeley: “Esse est percipi”)
– Phenomenalism: existence is enduring possibility of being experienced (e.g. Hume?)
– Materialism: only matter exists, and some portions of matter behave as if perceiving, desiring,
thinking, etc.
– Behaviourism: mental phenomena just are behaviours or behavioural dispositions, tendencies,
capabilities.
– Double-aspect theories
∗ There is only one kind of stuff, but it has different aspects (neutral monism)
∗ Mind-brain identity – mental processes just are the physical processes that implement them.
• Dualism:
– Interactionism: Two kinds of stuff exist, and interact causally
– Psycho-physical parallelism: Two streams existing independently – with no causal connections
– One way causation – e.g. matter to mind (Epiphenomenalism), or mind to matter (???-ism).
• Polyism:(??) There are many kinds of stuff interacting in different ways.
• Expressivism: labelling something as conscious expresses a stance or attitude to it.
E.g. Daniel Dennett’s “intentional stance”. A stance is useful, or useless, not true or false.
Lang Cog Sem Birmingham 2009 Slide 13 Last revised: January 27, 2010
14. A-C and P-C
All of the above leave many people feeling that there is an “explanatory gap” between the
physical and the mental.
Ned Block’s 1995 paper made an influential distinction between access consciousness (A-C, which has a
functional role) and phenomenal consciousness (P-C, which does not).
http://cogprints.org/231/0/199712004.html
“The information processing function of phenomenal consciousness in Schacter’s model is the ground of
what I will call “access-consciousness”. A perceptual state is access-conscious roughly speaking if its
content – what is represented by the perceptual state – is processed via that information processing
function, that is, if its content gets to the Executive system, whereby it can be used to control reasoning
and behavior.”
“Let me acknowledge at the outset that I cannot define P-consciousness in any remotely non-circular
way. I don’t consider this an embarrassment.”
“The controversial part is that I take P-conscious properties to be distinct from any cognitive, intentional,
or functional property. (Cognitive = essentially involving thought; intentional properties = properties in
virtue of which a representation or state is about something; functional properties = e.g. properties
definable in terms of a computer program”....) (Note: he should not have referred to computer programs:
there is a deeper, broader notion of function involving causal role – See quotes from Chalmers, below.)
“It is of course P-consciousness rather than access-consciousness or self-consciousness that has
seemed such a scientific mystery.”
Notice that P-C is defined negatively: e.g. it has no functional role, unlike A-C.
This implies that P-C has no cognitive causal powers.
This is a bit like postulating invisible, intangible, fairies at the bottom of the garden produced by the plants
growing, but incapable of influencing anything else.
How can things with no cognitive function generate so much philosophical questioning???
Lang Cog Sem Birmingham 2009 Slide 14 Last revised: January 27, 2010
15. More on gaps
Chalmers (see next slide) used a distinction very close to Block’s distinction between A-C
and P-C, to pose “the hard problem of consciousness” which actually is the very old
problem, expressed thus by T.H. Huxley (1866):
“How it is that anything so remarkable as a state of consciousness comes about as a result of irritating
nervous tissue, is just as unaccountable as the appearance of Djin when Aladdin rubbed his lamp.”
Many scientists around Huxley’s time thought Darwin’s arguments for evolution of animal bodies were
convincing in the light of evidence of small changes in physical structure, but also felt that evolution could
not produce anything like states of consciousness as a result of a succession of gradual changes in
physical matter. (Huxley’s “explanatory gap”.)
Chalmers suggests there are two gaps, which we can describe in Block’s terminology:
• a gap separating physical mechanisms from access consciousness
• a gap separating physical mechanisms from phenomenal consciousness
He thinks bridging the first gap (“the easy problem”) is non-trivial, but much easier than
bridging the second: “the hard problem of consciousness”.
He proposes a research programme investigating ‘Neural correlates of consciousness’ in
order to understand the second gap better.
But if phenomenal consciousness is defined as incapable of having any cognitive
functions, then I shall argue that the concept is incoherent, in which case we can solve
the hard problem easily by showing that it is a non-problem: just a philosophical muddle.
Explaining how A-C can exist in a physical world remains hard, but tractable, if we use the right tools.
Lang Cog Sem Birmingham 2009 Slide 15 Last revised: January 27, 2010
16. Explanatory gaps and the hard problem
David Chalmers re-labelled the ‘explanatory gap’ (T.H.Huxley, 1866) as ‘the hard problem
of consciousness’, which can be seen as the problem of explaining how physical
machinery can produce P-C.
He does not mention Block or Huxley, but essentially makes the same distinction as Block
between A-C and P-C, in different words:
http://www.imprint.co.uk/chalmers.html
Facing Up to the Problem of Consciousness (JCS, 1995)
“The easy problems of consciousness include those of explaining the following phenomena:
the ability to discriminate, categorize, and react to environmental stimuli;
the integration of information by a cognitive system;
the reportability of mental states;
the ability of a system to access its own internal states;
the focus of attention;
the deliberate control of behavior;
the difference between wakefulness and sleep.
.... “There is no real issue about whether these phenomena can be explained scientifically. All of them
are straightforwardly vulnerable to explanation in terms of computational or neural mechanisms. ....
Getting the details right will probably take a century or two of difficult empirical work.” ...
“If any problem qualifies as the problem of consciousness, it is this one. In this central sense of
‘consciousness’, an organism is conscious if there is something it is like to be that organism, and a
mental state is conscious if there is something it is like to be in that state. Sometimes terms such as
‘phenomenal consciousness’ and ‘qualia’ are also used here, but I find it more natural to speak of
‘conscious experience’ or simply ‘experience’....”
He is making essentially the same distinction as Block (using different labels) and
proposing that explaining how physical systems can produce P-C is the “hard problem”.
Lang Cog Sem Birmingham 2009 Slide 16 Last revised: January 27, 2010
17. Figure for Neural Correlates of Consciousness
by Mormann & Koch
From Scholarpedia
http://www.scholarpedia.org/article/Neural_correlates_of_consciousness
Problem: does causation go only one way?
How can motives, decisions, intentions, plans, preferences....
produce behaviour?
Or is it a myth that they do?
By definition, P-C cannot do that. What about A-C? How?
Lang Cog Sem Birmingham 2009 Slide 17 Last revised: January 27, 2010
18. What it is like to be X
I find it astounding that so many thinkers treat the phrase “what it is like to be” as having
sufficient precision to play a role in defining a serious problem for philosophy or science.
It was originally brought into philosophical discussion by Tom Nagel in “What is it like to be a bat?” in 1970
Compare my “What is it like to be a rock?”
http://www.cs.bham.ac.uk/research/projects/cogaff/misc/rock
As a piece of colloquial language the phrase is used in different contexts to express
different things. E.g. there are great differences between what it is like to be
• half asleep
• horizontal
• unpopular
• 25 years old
• in France
• able to cook a meal without help
• drowning
• deaf
• more knowledgeable than most of the population
• unwittingly on the verge of a great discovery
• dead
• unknown
The phrase expresses a polymorphic concept: what is being said depends on the whole
context in which it is being used, and there need not be anything in common to all the
possible interpretations of “what it is like to be” – as with the word “conscious”.
Lang Cog Sem Birmingham 2009 Slide 18 Last revised: January 27, 2010
19. There is another way: notice that minds DO things
What many people forget when they discuss these issues is that minds
don’t just contain bits of consciousness: they contain processes and
those processes do things, including producing, modifying and aborting,
other processes, mental and physical.
• In other words, minds are machines and the philosophical and scientific task is to find
out what they do, and what sorts of physical and non-physical machinery they use.
This is deeper and harder than asking what things minds contain, and how those things are
produced, and what non-mental things correlate with them.
• Intelligent minds produce and use theories, questions, arguments, motives, plans,
percepts: all of which are information structures (including descriptive and control structures).
• Such minds require information-manipulating machinery.
Some people may think their minds contain only sequences of sentences and pictures, or
sentence-like and picture-like entities.
But sentences and pictures are merely two sorts of formats for encoding various kinds of information.
A mind that understands sentences and pictures must have more than sentences and pictures, since
if understanding sentences and pictures amounted to having and understanding more sentences and
pictures, that would require an infinite regress of sentences and pictures.
• The implications of all that tend to be ignored in many discussions of consciousness.
• Most researchers investigating these topics lack powerful conceptual tools
developed in computer science/software engineering for creating, analysing,
understanding information processing machinery of various kinds.
Lang Cog Sem Birmingham 2009 Slide 19 Last revised: January 27, 2010
20. Information processing models of mind
Some researchers have attempted to explain what consciousness is in
terms of information-processing models of mentality, and they often
assume that such models could be implemented on computers.
Several decades ago, Ulric Neisser in “The Imitation of Man by Machine” Science (1963)
http://www.sciencemag.org/cgi/reprint/139/3551/193
distinguished what Robert Abelson later called “cold cognition” (perception, reasoning,
problem solving, etc.) and “hot cognition” (desires, emotions, moods, etc.).
Neisser claimed that the former could be replicated in computers but not the latter,
offering detailed examples to support his case.
Herbert Simon responded to this challenge in “Motivational and emotional controls of cognition”
Psychological Review 1967 http://circas.asu.edu/cogsys/papers/simon.emotion.pdf
arguing that computation can include control as well as “cold” cognition.
The idea was that emotions and other forms of “hot cognition” could be analysed as states and processes
produced by control mechanisms, including motive generators and interrupt mechanisms.
My own earliest attempts were in The Computer Revolution in Philosophy 1978, e.g. chapters 6–10.
http://www.cs.bham.ac.uk/research/projects/cogaff/crp/
Similar ideas were put forward before that and after that, e.g. in P.N. Johnson-Laird’s book The Computer
and the Mind: An Introduction to Cognitive Science (1988)
People objecting to such ideas claim that machines with all the proposed forms of
information processing could still lack consciousness: no matter how convincing their
behaviour, they could be mere “zombies” (defined in the next slide).
Lang Cog Sem Birmingham 2009 Slide 20 Last revised: January 27, 2010
21. What’s a zombie?
Many of the discussions of consciousness by philosophers refer to
zombies: e.g. claiming that no matter how closely you build a machine that
looks and behaves like a human and which contains a physical
information-processing machine mimicking all of the functionality of a
human brain it could still be a zombie.
The idea of a zombie occurs in a certain kind of grisly science fiction.
Here are some definitions:
• “A zombie is a creature that appears in folklore and popular culture typically as a reanimated corpse or
a mindless human being. Stories of zombies originated in the Afro-Caribbean spiritual belief system of
Vodou, which told of the people being controlled as laborers by a powerful sorcerer. Zombies became a
popular device in modern horror fiction, largely because of the success of George A. Romero’s 1968
film Night of the Living Dead.”
• “zombi: a dead body that has been brought back to life by a supernatural force”
• “zombi: a god of voodoo cults of African origin worshipped especially in West Indies”
• “automaton: someone who acts or responds in a mechanical or apathetic way”
Many philosophers have come to regard the explanatory gap as a challenge to
demonstrate that certain sorts of physical machines could not be zombies: can we find a
design that guarantees that its instances have minds and are conscious, and do not
merely behave as if they were conscious.
However if being conscious means having P-C that’s an incoherent challenge!
(As explained below.)
Lang Cog Sem Birmingham 2009 Slide 21 Last revised: January 27, 2010
22. Spurious philosophical arguments: imaginability
Dualists and idealists have been convinced that they can think about the mental as
distinct from the physical/material in a manner that refutes various explanations of what
mental phenomena are, as follows:
• Whatever a theorist proposes as constituting or explaining the existence of mental
phenomena (e.g. certain collections of brain processes, or certain collections of
internal cognitive processes and dispositions), such a dualist responds:
“I can imagine a machine that has all of that and behaves exactly like a conscious human, but is
really a zombie: it has no consciousness – it does not have this” said pointing inwardly.
• Philosophers who can imagine a zombie made according to specifications XYZ often
claim that that proves that XYZ cannot provide an analysis of what consciousness is,
or explain how it arises.
• The A-C/P-C distinction is sometimes appealed to in such arguments: saying that such a design can
produce A-C, but not P-C, not the “what it is like to be” states. So the design does not explain P-C.
• Such imaginability/conceivability of machinery without consciousness proves nothing when the machine
specification is rich and detailed enough to provide the basis for a design for a working system that goes
through all the internal cognitive processes required for functioning animals, including: interpretation of
sensory inputs, reasoning, learning, remembering, imagining, generating motives, deliberating,
deciding, controlling actions, and monitoring internal processes. (We are nowhere near such a design.)
When enough such detail has been achieved, claiming that there is something that has all
that richness but lacks contents of consciousness is a form of self delusion, for what is
claimed to be imagined is actually incoherent, as I’ll try to explain. In other words,
the concept of functionless, causally disconnected contents of consciousness (P-C) is incoherent.
Lang Cog Sem Birmingham 2009 Slide 22 Last revised: January 27, 2010
23. Beware of arguments from imaginability
People can think they are imagining something while fooling themselves –
because what they are referring to is either totally impossible or else
semantically detached from reality, or mathematically impossible.
• Someone may claim to be able to imagine it being 3am on the surface of the sun, or
noon at the centre of the earth.
But “the time at X” depends on angular elevation of the sun above the horizon at X.
Where there is no angular elevation, talk about “The time at X” makes no sense.
• Another old example is the claim to imagine the whole universe moving slowly in an
north-west direction at three miles per hour, or being able to imagine everything in the
universe expanding at a uniform rate, so that the expansion is not detectable.
In both those cases the claim is incoherent because what is imagined is defined so
as to undermine preconditions for what is imagined to exist!
• Someone may claim to be able to imagine invisible, intangible fairies at the bottom of
the garden smiling when the sun comes out.
But such imaginings are worthless because since the fairies have no effects on anything, the
imagined hypothesis can take infinitely many forms that are indistinguishable: two fairies, three
fairies, four fairies; talking English, or German, or Swahili, or ...
See http://www.cs.bham.ac.uk/research/projects/cogaff/09.html#pach
• Someone may claim to be able to imagine finding a set of objects that produce
different numbers when counted in different orders, though the objects do not change.
People can imagine many things that are later proved mathematically impossible.
Lang Cog Sem Birmingham 2009 Slide 23 Last revised: January 27, 2010
24. Representability does not imply possibility
Arguments of the form
“I can imagine something that has XYZ without phenomenal consciousness, therefore it
is possible for XYZ to exist without phenomenal consciousness”,
are reminiscent of this claim in Wittgenstein’s Tractatus (1922):
3.0321 “We could present spatially an atomic fact which contradicted the laws of physics, but not one
which contradicted the laws of geometry.”
Well, here is a spatial representation of a round blue square, seen edge-on:
[ ] demonstrating that imaginability, does not imply possibility.
¨
Moreover, in the picture by Oscar Reutersvard, shown
earlier, there is a fairly rich and detailed, but nevertheless
impossible configuration of cubes. So:
Even quite detailed description, depiction, or representation, of some
alleged possibility may conceal a deep contradiction or incoherence.
Untutored philosophical imaginings or intuitions about possible kinds of
information-processing systems, or minds, should be treated with deep
scepticism, like untutored mathematical imaginings, e.g. imagining
discovering the largest prime number.
Someone without experience of designing working information
systems may be unable to imagine sophisticated working systems
in detail, and may be deceived by assumed implications of
superficial (amateur) imaginings – e.g. imagined zombies.
Lang Cog Sem Birmingham 2009 Slide 24 Last revised: January 27, 2010
25. The problem of identity of spatial regions
Someone who is new to philosophy of space and time, and has not studied
much physics might ask:
Where is the volume of space that was occupied by my nose exactly ten minutes ago?
• If you are sitting in your office working at your desk and have not moved much in the
last ten minutes, the answer could be: “It is now still where it was ten minutes ago”.
• But suppose you were sitting in a railway carriage reading a book for the last ten
minutes: you could give different answers, e.g. the location in the carriage, or the
portion of the railway track you were close to ten minutes ago.
• Even while at your desk you were on a planet both rotating and hurtling through
space, so perhaps your nose is thousands of miles away from its previous location?
• Such examples may lead you to conclude that volumes of space do not have an
identity except relative to a reference frame provided by a set of space occupants.
(As Leibniz did – though Newton disagreed with him.)
• But a defender of absolute space (e.g. Newton) might ask: Where is that volume of
space in itself, not just in relation to you or me or a carriage or even the earth?
• Such a philosopher might even claim to be able to imagine the volume of space
continuing to exist, so it must be somewhere.
• But if spatial location is necessarily relative to space-occupants, then the question
where the volume is in itself is incoherent: there is no answer.
• We could label that incoherent problem: the hard problem of spatial re-identification!
Lang Cog Sem Birmingham 2009 Slide 25 Last revised: January 27, 2010
26. Incoherent hard problems
Wondering about the colours of invisible fairies in the garden, the present location of a
previously occupied volume of space, the direction of motion of the whole universe, about
the time at the centre of the earth, or whether some numbers get hungry at night are all
fairly obviously concerns about incoherent questions, and I am suggesting that some
problems about consciousness (but not all) are similarly disguised nonsense.
Examples of disguised nonsense questions are:
• whether you have qualia like mine not only in their functional relations but in their intrinsic non-functional
qualities (compare the problem of absolute identity of places)
• whether the colour and shape qualia you had yesterday are the same as the ones you have today in the
same situations
• how brains can produce functionless phenomenal consciousness
One benefit of developing sharp philosophical intuitions (not part of general education,
alas) is learning to detect disguised nonsense: like the “hard” problem of consciousness.
• Of course that still leaves the deep and difficult problems of explaining the widely agreed and easily
verified empirical phenomena, of the sorts described earlier, including many phenomena that are only
privately accessible.
• Part of the problem is to explain why they are only privately accessible: and the answer proposed here
is: because the contents of internal data-structures of complex virtual machines are not accessible from
behaviours and not measurable via physical measuring devices (e.g. brain scanners).
• So they need their own modes of investigation, which includes forming rich and detailed conjectures
about how the systems work, and testing those conjectures in terms of their detailed consequences,
including consequences observable in working models.
Lang Cog Sem Birmingham 2009 Slide 26 Last revised: January 27, 2010
27. Argument from infallibility
Some philosophers have mistakenly been impressed by the infallibility of
certain kinds of introspection.
“We are infallible about what we experience.”
“We have ‘direct access’ to our own states of consciousness so we cannot be mistaken
about them.”
Descartes: “I can doubt everything but not that I am doubting.”
“I can be wrong about whether there is a dagger before me,
but not about whether there seems to me to be a dagger before me.”
But this is exactly like a certain sort of infallibility of every measuring device –
it cannot be mistaken about what it reports!
A voltmeter can be wrong about what the voltage is, if it is faulty, but it cannot be wrong about what it
reports the voltage to be.
In the same way, we can be mistaken about what we have seen, but not about what we
seem to have seen.
However, nothing of any interest follows from this tautology.
In particular, it does not rule out being mistaken about what we have perceived.
Elaboration/Digression:
The next three slides present an experiment that demonstrates that people sometimes can be mistaken
about what they have experienced, though the experiment does not work for everyone.
Lang Cog Sem Birmingham 2009 Slide 27 Last revised: January 27, 2010
28. Experiment demonstrating fallibility about what you see
Experiment (slide 1)
The experiment is presented on the next slide, where a sentence is presented inside a
frame.
Your task is to look carefully at what is in the frame and decide whether there is anything
wrong with the sentence.
Some people think something is wrong, while others don’t see anything wrong.
When you have decided please go to the following slide, where you will be asked some
questions.
The experiment does not work for everyone.
In particular, it will not work for you if you have seen this sort of display previously.
Now look at the next slide. If you are sharing this display with someone else, neither of
you should say anything about what is in the frame until you are instructed to
communicate in the third slide.
Note: this experiment is also available on its own on the internet, here:
http://www.cs.bham.ac.uk/research/projects/cogaff/misc/unconscious-seeing.html
Lang Cog Sem Birmingham 2009 Slide 28 Last revised: January 27, 2010
29. This may not work for everyone.
Experiment (slide 2)
If you are doing this test with someone else, do not say anything out loud
about what you see here:
Trespassers will
will be prosecuted
Is there anything wrong with the sentence in the frame?
Look carefully at the sentence in the frame, and decide whether there is
anything wrong with it, and when you have decided (either way) continue
to the next slide, without looking back at this one.
Do not say anything to anyone else doing this experiment at the same time, as you may
spoil it for them.
Lang Cog Sem Birmingham 2009 Slide 29 Last revised: January 27, 2010
30. Post-test debriefing
Experiment (slide 3)
If you saw something wrong with the sentence in the frame while you were looking at it
then this experiment has not worked for you.
If you saw nothing wrong please try to answer these two questions without looking back
at the previous slide:
• How many words were in the frame?
• Where was the word “will” ?
If thinking about one or other of those questions (without looking back) makes you realise
that what you saw had an error that you did not previously notice, then the experiment
has worked for you.
If the experiment worked for you, then you were able, simply by asking yourself a new
question, to notice an aspect of your experience that had earlier escaped your attention.
That demonstrates that you can be mistaken about the contents of your own
consciousness when looking at the original display because the information you got from
the display included something you thought was not present: you thought nothing was
wrong. (This happens often during proof-reading of text.)
The information about the duplicated word must have been in your mind if you
could answer either of the above questions without looking back at the display.
(Now you may talk, and look back at the previous slide.)
Lang Cog Sem Birmingham 2009 Slide 30 Last revised: January 27, 2010
31. Mythical Infallibility
The alleged infallibility of knowledge about the contents of one’s
consciousness is supposed to be evidence that people have direct
understanding of the nature of phenomenal consciousness.
Arguments based on that kind of infallibility are flawed in two ways:
• As we have seen, there are aspects of their own contents of consciousness about
which people can be mistaken (e.g. when failing to spot errors in texts they are
reading).
• Insofar as there are real examples of infallibility, they are the tautological
consequences of trivial truths similar to
– I cannot be wrong about how things seem to me.
– A voltmeter cannot be wrong about the voltage it has measured.
Claims that such infallibility demonstrates the undeniable existence of something
(seemings, qualia, ....) are therefore not to be taken seriously.
Lang Cog Sem Birmingham 2009 Slide 31 Last revised: January 27, 2010
32. Summary of the last few slides
A: Arguments of the form “I can imagine something with mechanisms of
type X, but lacking this sort of experience” are not to be taken seriously
as arguments against attempts to explain having experiences in terms
of mechanisms of type X, since humans, including sophisticated
philosophers, can imagine things that are impossible without realising
that that is what they are doing.
Sometimes that can be because they imagine the wrong (over-simple) things of type
X, e.g. if they don’t know enough about varieties of information processing systems.
B: Claims to know about the nature of consciousness on the basis of
infallible types of introspection are not to be taken seriously, since
(a) introspection can be fallible
(b) the infallible aspect is tautological and similar to infallibility found in all
representational devices: they can’t be wrong about what they are representing,
though what they represent can be wrong.
The upshot is that common philosophical arguments intended to refute the claim that
minds are information-processing systems can be discounted.
But we still lack a deep theory about relevant types of information processing systems:
including how they evolved and how they work. But we have made some progress.
Lang Cog Sem Birmingham 2009 Slide 32 Last revised: January 27, 2010
33. We need to understand minds with architectures
The preceding experiment gives us clues about the architecture of a
human mind, and some aspects of the structure of visual experience:
• A mind is not a unitary entity whose contents of consciousness at a time are sharply
defined.
• Rather a mind is a complex information-processing system with an architecture that
involves different subsystems working in parallel, and the subsystems may acquire,
use and store different information from the same sensory input. (Different kinds of
minds have different interacting subsystems.)
• In particular, the subsystem that is engaged in communication with someone else (e.g.
trying to answer the question I posed for you earlier about the contents of the box)
may fail to detect part of what is perceived, even though another subsystem has all
that information and is capable of being interrogated by the communication subsystem
even after the original physical source of the information has been removed.
• What is perceived (the contents of visual experience) can have different layers of
information, including a layer concerned with the distribution of colour and texture in
the visual field, a layer concerned with “chunks” in the field, layers concerned with
more abstract entities, such as phrases and sentences, and many others ...
See Chapter 9 of The Computer Revolution in Philosophy (1978)
http://www.cs.bham.ac.uk/research/projects/cogaff/crp/chap9.html
Lang Cog Sem Birmingham 2009 Slide 33 Last revised: January 27, 2010
34. Varieties of having and using information
It is often assumed that if something has and uses information it also has
the information that it has the information: but this is a fallacy.
• For organisms, much (often all) of the information used is transient control information.
If an organism senses a noxious substance and shrinks away from it, the information that led to the
action could be transient: there need be no subsequent record of it.
• Being able to use information for control purposes is one thing: retaining the
information for future use is another.
It requires a more complex information-processing architecture, allowing previously used information
to be stored and accessed: we don’t notice that when we experience or remember change.
• One important reason for storing information after use is that it can make change
detection possible: otherwise an organism merely has at any instant information about that instant.
• The ability to detect change (and processes, flows, speeds at which things happen)
requires more than just storage after use
the architecture needs to include mechanisms that compare stored and new information: to produce
information about the change (e.g. this is crucial for homeostatic control mechanisms).
• Information about change also could be either transient, or recorded for future use.
• People are surprised at change blindness: instead they should be puzzled by
change-detection, and by detection of no-change!
• So detecting change or persistence, using change, detecting change detection,
storing information about what does and does not change, all require additional
information processing mechanisms beyond merely having and using information.
Similar comments apply to detection of spatial differences, curvature, gradients, boundaries, etc.
Lang Cog Sem Birmingham 2009 Slide 34 Last revised: January 27, 2010
35. The mythical “it”
We have a noun “consciousness” and we assume there must be
something it refers to: forgetting that it is derived from an adjective
“conscious”, whose primary use is before a prepositional phrase “of X”.
Many spurious problems arise from such reifications, e.g. asking what the noun “emotion” refers to
instead of analysing the adjectives “emotional”, and “moved”, and related adjectives, “afraid”,
“despondent”, “grateful”, “infatuated”, “devastated”, etc.
The use of the noun “consciousness” suggests there is one thing referred to and invites
questions about “it”, like
What are its functions? How did it evolve?
Which animals have it? Can machines have it?
At what stage does a foetus have it? What causes it?
Which bits of brain produce it? How does the brain produce it?
Attending to the use of the adjective with “of ...”, reveals that there is no one thing.
There is no “it” about which we can ask such questions, because there are huge differences between
being conscious (or not) of a toothache, of how late it is, of the changing weather, of the day of the week,
of being puzzled, of being unable to remember a name, of being unpopular, of being in favour of capital
punishment, of a likely collision, and many more.
“Conscious” is a highly polymorphic concept. (Familiar in computing.)
Adjectives like “efficient”, and “useful” are also polymorphic, in similar ways.
There is no one thing that being efficient, or useful, or expensive or dangerous is: in each case, what the
adjective indicates has to be specified in the context, e.g. being an efficient (or dangerous) lawn-mower
or microwave oven, proof strategy, government department, or murderer.
Lang Cog Sem Birmingham 2009 Slide 35 Last revised: January 27, 2010
36. Semantic polymorphism: multiple “ITs”
It is well known that the meaning of the adjective “tall” does not specify a minimum height
required for being tall: some additional noun or noun phrase is required to specify a
comparison class, e.g.
A tall squirrel, a tall giraffe, a tall building for a small rural town, a tall flea?
These are cases of “parametric polymorphism”: an extra parameter is required to specify
which of several interpretations the word has.
The same applies to “A is conscious of X”: What A and X are make a big difference.
There are many consequences
• Different animals can be conscious of different things, and with different consequences
and in different ways (e.g. echolocation vs using feelers vs using vision, etc.)
• So there are different varieties of consciousness that evolved at different times in
different species – these are not minor differences of degree.
• Different information-processing mechanisms and architectures are required
– in some cases the mechanisms require introspective mechanisms (e.g. being conscious of feeling
annoyed or surprised), but it’s likely that not all animals can do that;
– other kinds of consciousness (e.g. being conscious of an obstacle to be avoided or of something
large approaching quickly) seem to be in many more species;
– some forms of human consciousness are not possible for infants or toddlers: the information
processing architecture is still too primitive (e.g. being conscious of finding algebra difficult);
– some are only possible in particular cultures, e.g. being conscious of having begun a new millenium,
was impossible for our early homo sapiens ancestors. (Why?)
Lang Cog Sem Birmingham 2009 Slide 36 Last revised: January 27, 2010
37. How did consciousness evolve? How did what? evolve?
Can we draw a line between things with and without consciousness?
• Some people (e.g. Susan Greenfield) think it’s all just a matter of degree:
evolution produced more and more complex kinds of consciousness.
BUT evolution cannot produce continuous change –
so we need to study many small discontinuities rather than one big one or only smooth transitions.
• One way to think about this:
– instead of trying to draw a single boundary, regard anything that acquires and uses information (i.e.
all living things) as in some way conscious
– but they are conscious of (acquire information about) different things; and their consciousness has
different functions and effects. There are differences in:
∗ what information is acquired
∗ how it is acquired
∗ how it is processed
∗ how it is used, and whether different sub-systems use it in different ways
∗ how it is stored
∗ how it is encoded/represented, and whether it is stored in different ways in one individual,
∗ how much of it is represented in features of the environment rather than entirely in the user.
– we can then try to understand the many different problems posed to different species, and the many
different solutions produced by various combinations of biological evolution, cultural evolution,
individual development and learning.... a survey of types of information-processing system
– Instead of differences of degree we have differences of kind – not necessarily linearly ordered.
• This requires good conceptual tools for thinking about such systems, how they work,
and how they differ: What tools? – tools for designing testable working systems!
Lang Cog Sem Birmingham 2009 Slide 37 Last revised: January 27, 2010
38. The ordinary notion vs the scientific notion
Scientists and philosophers who don’t understand the diversity associated
with the polymorphism of “consciousness” as ordinarily understood, are
tempted to formulate pointless, unanswerable questions about “it”.
The ordinary notion of consciousness (being conscious of XX) is highly polymorphic.
There are major differences in varieties of consciousness (being conscious of something) that depend on
what XX is: different XXs require different information-processing competences and architectures.
Some of those are products of recent evolution, or cultural development, or individual learning, while
others are based on evolutionarily old mechanisms. (See previous slides.)
People who don’t notice this diversity/polymorphism talk about consciousness as one “it”:
• How did it evolve? or When did it evolve?
• What are its functions?
• Which things have it? (At what stage is it in a foetus?)
• What are its neural correlates?
Instead of assuming that there’s one thing all these questions refer to, we need to allow
that there are different things and each question has a variety of answers.
So searching for neural correlates of “it” is pointless: though there may be neural
correlates of some of “them”. (As explained later, some may have no neural correlates.)
However it takes hard analytical research (requiring philosophical sophistication) to
identify the diversity of “them”, so as to clarify what the various research problems are.
Lang Cog Sem Birmingham 2009 Slide 38 Last revised: January 27, 2010
39. None of this matters for ordinary usage
The problems arise only when philosophers and scientists make false
assumptions about what “consciousness” and related words mean, since
in our ordinary usage we take the polymorphism in our stride: we have
learnt to make use of context in interpreting uses of many polymorphic
concepts, including “tall”, “efficient”, “useful”, “dangerous”, “prevention”,
“aware”, “attend”, and “conscious”.
• These are examples of the extreme complexity of unconscious processing that goes
on when we learn and use language: we acquire and use very many complex
syntactic rules that we are totally unaware of, some of which are quite hard to
discover, as linguists have found.
• In addition to ordinary informal usage, as in “I have been conscious of your hostility to
me”, there are many medical uses of the adjective “conscious” and noun
“consciousness” which do not make any explicit claims about what consciousness is,
e.g. when a doctor asks “when did the patient regain consciousness?”
• It is only when consciousness itself becomes the subject of research, as if it were a
well-defined unitary topic, that the confusions I am criticising tend to arise.
• If consciousness is not a unitary phenomenon with specific neural correlates, we can
ask how the variety of non-physical phenomena found in minds can occur as a result
of the operation of different kinds of physical machinery, some produced at different
stages of biological evolution, as found in brains, sensors and effectors.
Lang Cog Sem Birmingham 2009 Slide 39 Last revised: January 27, 2010
40. Causation is at the heart of the problem
Is this an acceptable picture of causal relations between brain and mind?
If M1, M2, M3 fit the definition of
“phenomenal consciousness” then all the
arrows with “?” must be removed.
P-C is defined to be epiphenomenal – a side
effect of physical processes, but lacking any
functional role, which means P-C has no
causal powers.
However, if M1, M2, M3 fit the definition of A-C, they can have cognitive functions and causal powers.
But how can they influence physical events if all of P1, P2, P3 are determined by previous physical events?
Worries about this leads some people to propose gaps in physical causation (e.g. quantum indeterminacy)
to allow mental causation to play a role.
We don’t need any gaps in physical causation.
In the last half century we have learnt how to build active virtual machines (VMs) that run
in parallel with physical machinery on which they depend for their existence, and with the
ability to control behaviours of those physical machines.
Objections to this claim are based on an inadequate understanding of causation, as if it
were some sort of fluid flowing through time.
Instead we need to understand causation in terms of networks of structural relations that
constrain what can happen, including relations and constraints involving entities and
processes in running virtual machines, e.g. operating systems, spelling checkers, ....
Lang Cog Sem Birmingham 2009 Slide 40 Last revised: January 27, 2010
41. Towards a meta-theory of information-processing systems
• The most advanced such theories come from computer science/software engineering.
• They have led to profound technological advances in information processing
machinery.
• But most computer scientists ignore the biological information processing systems that
preceded the development of electronic computers.
• So their theories are not general enough
• Some very important. complex and technical, ideas have been developed since the
1950s and we can use them, and extend them in trying to understand products of
biological evolution.
• Among the most important is the idea of a running virtual machine (RVM), not to be
confused with the abstract mathematical structure defining a type of VM.
Lang Cog Sem Birmingham 2009 Slide 41 Last revised: January 27, 2010
42. The 20th C Philosophical breakthrough: Virtual machinery
Brief introduction to the technology of virtual machinery (not virtual reality)
developed over the last six decades and its philosophical relevance:
Processes and events in running virtual machines can be causes and
effects, despite being implemented in deterministic physical mechanisms.
Instead of the picture on the left, which implies that there is only one-way causation from
physical to mental, we need to understand how running virtual machinery can co-exist
with, and influence, underlying physical machinery, which it helps to control, even though
the virtual machinery is all fully implemented in the physical machinery.
The virtual machinery (e.g. a chess program, or email program) chunks processes at a
level of abstraction that cannot be defined in the language of physics, as explained in this
presentation: http://www.cs.bham.ac.uk/research/projects/cogaff/talks/#mos09
Lang Cog Sem Birmingham 2009 Slide 42 Last revised: January 27, 2010
43. Virtual machines are everywhere
How many levels of (relatively) virtual machinery does physics itself require?
Lang Cog Sem Birmingham 2009 Slide 43 Last revised: January 27, 2010
44. A brief history of evolution!
Growing environmental demands, and increasing complexity of the organisms
themselves, led to many important changes in the requirements for information-
processing mechanisms and architectures, forms of representation, and ontologies.
Eventually the complexity of the control problems demanded use of virtual machinery
running on the biological physical machinery.
(Physical reconfiguration (e.g. neural re-growth) is much too slow for animals in dynamic environments!)
Lang Cog Sem Birmingham 2009 Slide 44 Last revised: January 27, 2010
45. Mechanisms required to support running VMs
We don’t know exactly what problems evolution faced, and what solutions it came up with,
and what mechanisms it used, in creating virtual machinery running in animal bodies, to
control increasingly complex biological organisms (and societies), but perhaps we can
learn from some of the developments supporting increasing use of increasingly
sophisticated VMs in artificial systems over the last half century (a small sample follows):
• The move from bit-level control to control of and by more complex and abstract patterns.
• The move from machine-level instructions to higher level languages (using compilers that ballistically
translate to machine code and especially interpreters that “translate” dynamically, informed by context).
• Memory management systems make physical memory reference context-dependent. (
• Virtual memory (paging and cacheing) and garbage collection switch virtual memory contents between
faster and slower core memories and backing store, and between different parts of core memory:
constantly changing PM/VM mappings. (These support multiple uses of limited resources.)
• Networked file systems change apparent physical locations of files.
• Device interfaces translate physical signals into “standard” VM signals and vice versa.
• Devices can themselves run virtual machines with buffers, memories, learning capabilities...
• Device drivers (software) handle mappings between higher level and lower level VMs – and allow
devices to be shared between VMs (e.g. interfaces to printers, cameras, network devices).
• Context-dependent exception and interrupt handlers distribute causal powers over more functions.
• Non-active processes persist in memory and can have effects on running processes through shared
structures. (It’s a myth that single-cpu machines cannot support true parallelism.)
• Multi-cpu systems with relocatable VMs allow VM/PM mappings to be optimised dynamically.
• Multiplicity of concurrent functions continually grows – especially on networked machines.
• Over time, control functions increasingly use monitoring and control of VM states and processes.
Lang Cog Sem Birmingham 2009 Slide 45 Last revised: January 27, 2010
46. Different requirements for virtual machinery
The different engineering developments supporting new kinds of virtual
machinery helped to solve different sorts of problems. E.g.
• Sharing a limited physical device between different users efficiently.
• Optimising allocation of devices of different speeds between sub-tasks.
• Setting interface standards so that suppliers could produce competing solutions.
• Allowing re-use of design solutions in new contexts.
• Simplifying large scale design tasks by allowing components to “understand” more
complex instructions (telling them what to do, leaving them to work out how to do it).
• Specifying abstract kinds of functionality that could be instantiated in different ways in
different contexts (polymorphism).
• Improving reliability of systems using unreliable components.
• Allowing information transfer/information sharing to be done without constant
translation between formats.
• Simplifying tasks not only for human designers but also for self-monitoring,
self-modulating, self-extending systems and sub-systems.
These are solutions to problems that are inherent in the construction and improvement of complex
functioning systems: they hare not restricted to artificial systems, or systems built from transistors, or ...
Conjecture: Similar problems were encountered in biological evolution (probably many
more problems) and some of the solutions developed were similar, while some were a lot
more sophisticated than solutions human engineers have found so far.
Lang Cog Sem Birmingham 2009 Slide 46 Last revised: January 27, 2010
47. Virtual machinery and causation
Virtual machinery works because “high level” events in a VM can control
physical machinery.
Accordingly, bugs in the design of virtual machines can lead to disasters, even though
there’s nothing wrong with the hardware.
As stated previously
Processes and events in running virtual machines can
be causes and effects, despite being implemented in
deterministic physical mechanisms.
Engineers (but not yet philosophers, psychologists, neuroscientists?)
now understand how running virtual machinery can co-exist with,
and influence, underlying physical machinery, which it helps to
control, even though the virtual machinery is all fully implemented in
the physical machinery.
This works because
• We have learnt how to set up physical mechanisms that enforce constraints between abstract patterns
(in contrast with mechanisms that enforce constraints between physical or geometric relations).
• Chains of such constraints can have complex indirect effects linking different patterns.
• Some interactions involve not only causation but also meaning: patterns are interpreted as including
descriptive information (e.g. testable conditions) and control information (e.g. specifying what to do).
See “What enables a machine to understand?” (IJCAI 1985)
http://www.cs.bham.ac.uk/research/projects/cogaff/81-95.html#4
Could biological evolution have solved similar problems?
Lang Cog Sem Birmingham 2009 Slide 47 Last revised: January 27, 2010
48. Physical ⇐⇒ virtual interfaces at different levels
Starting with simple physical devices implementing interacting discrete
patterns, we have built layers of interacting patterns of ever increasing
spatial and temporal complexity, with more and more varied functionality.
• Physical devices can constrain continuously varying states so as to allow only a small
number of discrete stable states (e.g. only two)
(e.g. using mechanical ratchets, electronic valves (tubes), aligned magnetic molecules, transistors etc.)
• Networks of such devices can constrain relationships between discrete patterns.
E.g. the ABCD/XY example: a constraint can ensure that if devices A and B are in states X and Y
respectively then devices C and D will be in states Y and X (with or without other constraints).
So, a device network can rule out some physically possible combinations of states of components,
and a new pattern in part of the network will cause pattern-changes elsewhere via the constraints.
Compare: one end of a rigid lever moving down or up causes the other end to be moving up or down.
• Such networks can form dynamical systems with limited possible trajectories,
constraining both the possible patterns and the possible sequences of patterns.
• A network of internal devices can link external interfaces (input and output devices)
thereby limiting the relationships that can exist between patterns of inputs and patterns of outputs,
and also limiting possible sequences of input-output patterns.
• Patterns in one part of the system can have meaning for another part, e.g.
– constraining behaviour (e.g. where the pattern expresses a program or ruleset) or
– describing something (e.g. where the pattern represents a testable condition)
• Such patterns and uses of such patterns in interacting computing systems may result
from design (e.g. programming) or from self-organising (learning, evolving) systems.
• Some useful patterns need not be describable in the language of physics.
Lang Cog Sem Birmingham 2009 Slide 48 Last revised: January 27, 2010
49. Try playing with interacting abstract patterns
There are movies showing some videos of some interacting virtual machines
implemented in the (freely available) Poplog SimAgent Toolkit here:
http://www.cs.bham.ac.uk/research/projects/poplog/packages/simagent.html
John Conway’s “Game of Life” provides many examples of interacting abstract patterns.
An excellent online version can be run in a (java-enabled) web browser here:
http://www.bitstorm.org/gameoflife/ (It allows mouse interaction with a running process.)
Anyone can play with interacting patterns by clicking on squares to make an initial pattern or set of patterns
(or selecting a pre-built pattern) from the menu labelled ‘Clear’, and then stepping through the interactions
by repeatedly pressing ‘Next’, or launching the process by pressing ‘Start’.
For example, a starting pattern of five squares like the pattern (a)
very quickly fizzles out, whereas moving one of the blue squares
to form the starting pattern (b) on the right produces a totally
different result: once started the display continues changing
hundreds of times without repeating itself until it eventually (after
more than 1100 changes – if the board is big enough) gets stuck
with a collection of static or repeating small patterns.
Such demonstrations illustrate ways in which abstract patterns can be made to interact
causally (parts of patterns cause other parts to change), by harnessing physical
machinery to operate according to a set of constraints (Conway’s four rules).
The constraints can be implemented via mechanical linkages, or via electronic circuits, or via a computer
program that forces the computer hardware to behave as if wired to obey the rules: as a result a group of
squares changing state can cause a new pattern to come into existence.
Lang Cog Sem Birmingham 2009 Slide 49 Last revised: January 27, 2010
50. Which patterns interact?
When Conway’s machine runs, do patterns on the screen interact
causally? NO!
The screen merely displays what is going on in a virtual machine running in the computer.
In the running VM the abstract data-structures in the 2-D grid in the VM interact.
Those changing VM structures are represented by changing squares on the screen
merely to inform us what is happening inside.
• In a different physical design the screen could be part of the implementation of the VM.
Most Conway implementations would go on running even if the screen were disconnected: they would
merely cease to display anything. That is how some people understand contents of consciousness –
i.e. as causally/functionally ineffective (e.g. P-C as defined by Block).
(Perhaps they think zombies are possible because they imagine turning off an internal screen.)
• If the grid information used in applying the rules were stored in devices implementing the pixels of the
screen display, then we could say that the cells in the screen interact – but not their visual effects.
• However, if you are inspired, by what you see on the screen, to interact by clicking in the grid to
change a pattern, altering future behaviour, then the visible screen patterns are causally effective in
changing how the process develops, and your eyes and brain become part of the physical
implementation of a very sophisticated distributed virtual machine!
The causation observed in a Conway machine is normally only in the (invisible) virtual
machine that causes what is on the screen to change: the screen display is (relatively)
epiphenomenal, like changing expressions on people’s faces, when they are alone!
For more examples and references see http://en.wikipedia.org/wiki/Conway%27s_Game_of_Life
Lang Cog Sem Birmingham 2009 Slide 50 Last revised: January 27, 2010
51. More complex interacting patterns
If a run of the machine starts with two separated patterns, then for a time
each may change internally without influencing the other, even if each
pattern has many internal interactions.
But that can change after a while.
• The distances between cells influenced by the two patterns may gradually be reduced
either because the patterns grow, or because they move (e.g. so-called “gliders” in
Conway’s machine).
• The results of the interaction may feed back into both patterns making them behave
thereafter very differently from how they would have behaved on their own.
• The form of interaction can depend not only on the starting shape of each pattern, but
exactly how far apart they are initially (and whether they “bump into” the grid frame).
• In other cases they may interact via a communicating bridge pattern.
• In all cases the interactions are caused by pattern structures and pattern relationships
insofar as the same patterns placed in different starting locations of the Conway
machine will always produce the same results: so the precise implementation
mechanism is less important than the patterns and their rules of behaviour.
It is also possible to produce variants of the Conway machine whose rules are
probabilistic (stochastic) rather than deterministic.
The form of causation in such machines is then probabilistic, except when pattern structures constrain
possible outcomes despite the probabilistic rules.
Lang Cog Sem Birmingham 2009 Slide 51 Last revised: January 27, 2010
52. Reactive vs deliberative interacting patterns
A Conway machine uses real or simulated concurrency: behaviour of each
square depends only on the previous states of its eight neighbours and
nothing else.
On a computer the concurrency is achieved by time-sharing, but it is still real concurrency.
Consider what happens when two virtual machines running on a computer compete in a
chess game, sharing a virtual board, and interacting through moves on the board, each
can sense or alter the state of any part of the (simulated) chess board.
• In general, programs on a computer are not restricted to local interactions.
• In some cases, the interacting processes are purely reactive: on every cycle every
square immediately reacts to the previous pattern formed by its neighbours.
• If two instances of a chess program (or instances of different chess programs) interact
by playing chess in the same computer, their behaviour is typically no longer purely
reactive. Good ones will often have to search among possible sequences of future
moves to find a good next move – and only then actually move.
In addition, one or both of the chess virtual machines may do some searching in advance while
waiting for the opponent’s next move.
• Then each instance is a VM with its own internal states and processes interacting
richly, and a less rich interaction with the other VM is mediated by changes in the
shared board state (represented by an abstract data-structure).
For more on varieties of deliberation see:
http://www.cs.bham.ac.uk/research/projects/cosy/papers/#dp0604
Lang Cog Sem Birmingham 2009 Slide 52 Last revised: January 27, 2010
53. Intentionality in a virtual machine
A running chess program (a VM) takes in information about the state of the
board after the opponent moves, and builds or modifies internal structures
that it uses to represent the board and the chess pieces on it, and their
relationships, including threats, opportunities, possible traps, etc.
• In particular it uses those representations in attempting to achieve its goals.
So, unlike the interacting Conway patterns mentioned earlier, some of the patterns in the chess
virtual machine are treated by the machine as representations, that refer to something.
• During deliberation, some created patterns will be treated as referring to non-existent
but possible future board states, and as options for moves in those states.
They are treated that way insofar as they are used in considering and evaluating possible future move
sequences in order to choose a move which will either avoid defeat (if there is a threat) or which has
a chance of leading to victory (check-mate against the opponent).
• In this case the chess VM, unlike the simplest interacting Conway patterns, exhibits
intentionality: the ability to refer. (NB. The programmer need not know about the details.)
Since the Conway mechanism is capable of implementing arbitrary Turing machines, it could in
principle implement two interacting chess virtual machines, so there could be intentionality in virtual
machines running on a Conway machine – probably requiring a very big fairly slow machine.
• The intentionality of chess VMs is relatively simple because they have relatively few
types of goal, relatively few preferences, and their options for perceiving and acting
are limited by being constrained to play chess:
For a human-like, or chimp-like, robot the possibilities would be much richer, and a far more complex
architecture would be required. See
http://www.cs.bham.ac.uk/research/projects/cogaff/03.html#200307
Lang Cog Sem Birmingham 2009 Slide 53 Last revised: January 27, 2010
54. Adding meta-semantic competence
If a virtual machine playing chess not only thinks about possible board
states and possible moves, and winning, moving, threats, traps, etc. but
also thinks about what the opponent might be thinking, then that requires
meta-semantic competences: the ability to represent things that
themselves represent and use information.
• It is very likely that biological evolution produced meta-semantic competences in some
organisms other than humans because treating other organisms (prey, predators,
conspecifics to collaborate with, and offspring as they learn) as mere physical
systems, ignoring their information-processing capabilities, will not work well
(e.g. hunting intelligent prey, or avoiding intelligent predators).
• Another application for meta-semantic competences is self-monitoring, self evaluation,
self-criticism, self-debugging: you can’t detect and remedy flaws in your thinking,
reasoning, planning, hypotheses etc. if you are not able to represent yourself as an
information user.
• It is often assumed that social meta-semantic competences must have evolved first,
but that’s just an assumption: it is arguable that self-monitoring meta-semantic
competences must have evolved first
e.g. because an individual has relatively direct access to (some of) its own information-processing
whereas the specifics of processing in others has to be inferred in a very indirect way (even if
evolution produced the tendency to use information about others using information).
See A. Sloman, 1979, The primacy of non-communicative language,
http://www.cs.bham.ac.uk/research/projects/cogaff/81-95.html#43
Lang Cog Sem Birmingham 2009 Slide 54 Last revised: January 27, 2010
55. Emerging varieties of functionality
Computer scientists and engineers and AI/Robotics researchers have
been learning to add more and more kinds of control, kinds of pattern, and
ways of interpreting patterns of varying levels of abstraction.
• A simple machine may repeatedly take in some pattern and output a derived pattern,
e.g. computing the solution to an arithmetical problem.
• More complex machines can take in a pattern and a derivation-specification (program)
and output a derived pattern that depends on both.
• Other machines can continually receive inputs (e.g. from digitised sensors) and
continually generate outputs (e.g. to digitally controlled motors).
• More sophisticated machines can
– solve new problems by searching for new ways of relating inputs to outputs, i.e. learning;
– interpret some patterns as referring to the contents of the machine (using a somatic ontology) and
others to independently existing external entities, events, processes (using an exosomatic ontology)
– extend their ontologies and theories about the nature and interactions of external entities
– perform tasks in parallel, coordinating them,
– monitor and control some of their own operations – even interrupting, modulating, aborting, etc.
(Including introspecting some of their sensory and other information contents: qualia.)
– develop meta-semantic ontologies for representing and reasoning about thinking, planning, learning,
communicating, motives, preferences, ...
– acquire their own goals and preferences, extending self-modulation, autonomy, unpredictability, ...
– develop new architectures which combine multiple concurrently active subsystems.
– form societies, coalitions, partnerships ... etc.
• Biological evolution did all this and more, long before we started learning how to do it.
Lang Cog Sem Birmingham 2009 Slide 55 Last revised: January 27, 2010
56. Causal networks linking layered patterns
How can events in virtual machines be causes as well as effects, even
causing physical changes?
The answer is
through use of mechanisms that allow distinct patterns of states and sequences
of patterns to be linked via strong constraints to other patterns of states and
sequences of patterns (as in the ABCD/XY example, and the Conway machines,
mentioned above). (Some VMs may use probabilistic/stochastic constraints.)
What many people find hard to believe is that this can work for a virtual machine whose
internal architecture allows for divisions of functionality corresponding to a host of
functional divisions familiar in human minds, including
• interpreting physical structures or abstract patterns as referring to something (intentionality)
• generation of motives,
• selection of motives,
• adoption of plans or actions,
• perceiving things in the environment,
• introspecting perceptual structures and their changes,
• extending ontologies,
• forming generalisations,
• developing explanatory theories,
• making inferences,
• formulating questions,
• and many more.
Lang Cog Sem Birmingham 2009 Slide 56 Last revised: January 27, 2010
57. Biological unknowns: Research needed
Many people now take it for granted that organisms are information-processing systems,
but much is still not known, e.g. about the varieties of low level machinery available (at
molecular and neuronal mechanisms) and the patterns of organisation for purposes of
acquiring and using information and controlling internal functions and external behaviours.
Steve Burbeck’s web site raises many of the issues:
“All living organisms, from single cells in pond water to humans, survive by constantly processing
information about threats and opportunities in the world around them. For example, single-cell E-coli
bacteria have a sophisticated chemical sensor patch on one end that processes several different
aspects of its environment and biases its movement toward attractant and away from repellent
chemicals. At a cellular level, the information processing machinery of life is a complex network of
thousands of genes and gene-expression control pathways that dynamically adapt the cell’s function to
its environment.”
http://evolutionofcomputing.org/Multicellular/BiologicalInformationProcessing.html
“Nature offers many familiar examples of emergence, and the Internet is creating more.
The following examples of emergent systems in nature illustrate the kinds of feedback between
individual elements of natural systems that give rise to surprising ordered behavior. They also illustrate
the trade off between the number of elements involved in the emergent system and the complexity of
their individual interactions. The more complex the interactions between elements, the fewer elements
are needed for a higher-level phenomenon to emerge. ... networks of computers support many sorts of
emergent meta-level behavior because computers interact in far more complex ways than air and water
molecules or particles of sand ... Some of this emergent behavior is desirable and/or intentional, and
some (bugs, computer viruses, dangerous botnets, and cyber-warfare) are not.”
http://evolutionofcomputing.org/Multicellular/Emergence.html
Lang Cog Sem Birmingham 2009 Slide 57 Last revised: January 27, 2010
58. An old idea.
The idea of the analogy expressed in this diagram is
very old, but we are only slowly understanding the
variety of phenomena on the left hand side, extending
our appreciation of what might be going on on the right.
The simple-minded notion that the left hand side involves a
program in a computer is seriously deficient, (a) because it ignores
the requirement for the program to be running and (b) because we
know now that there are far more types of information-processing
system than a computer running a single program, as explained in
other slides.
Simple-minded views about virtual machines lead to
easily refutable computational theories of mind, e.g. the
theory that virtual machines are simple finite state
machines, as illustrated on the right (“Atomic state
functionalism”). See
Ned Block: Functionalism http://cogprints.org/235/
A. Sloman, The mind as a control system, 1993,
http://www.cs.bham.ac.uk/research/projects/cogaff/81-95.html#18
Forget what you have learnt about Turing machines: that’s a simple
abstraction which is surprisingly useful for theorising about classes
of computations – but not so useful for modelling complex
multi-component systems interacting asynchronously with a rich
and complex environment.
Lang Cog Sem Birmingham 2009 Slide 58 Last revised: January 27, 2010
59. More realistic models
As crudely indicated here we need to allow multiple concurrent inputs (blue) and outputs
(red), multiple interacting subsystems, some discrete some continuous, with the ability to
spawn new subsystems/processes as needed.
Artificial VM on artificial PM Biological VM on biological PM
In both cases there are multiple feedback loops involving the environment.
Lang Cog Sem Birmingham 2009 Slide 59 Last revised: January 27, 2010
60. Architectures
We need good conceptual tools for talking about architectures.
The CogAff architecture schema crudely depicted on the left specifies types of
components of an architecture.
Particular architectures can be specified by filling in the boxes and indicating connections
(information flow and causal influences) between subsystems.
A particular way of filling in the boxes has been explored in the cognition and affect
project as a move towards more human like systems.
A crude schematic representation of that option, the H-CogAff architecture (strictly
another, more constrained, architecture schema), is shown on the right.
For more detail see the CogAff papers and presentations. Also Marvin Minsky The Emotion Machine
Lang Cog Sem Birmingham 2009 Slide 60 Last revised: January 27, 2010