Professional Documents
Culture Documents
The precondition for vision is the existence of light. Light is electromagnetic radiation that
can be described in terms of wavelength. Humans can perceive only a small range of the
wavelengths that exist; the visible wavelengths are from 380 to 750 nanometers (Figure 3.6;
Starr, Evers, & Starr, 2007). Vision begins when light passes through the protective covering
of the eye (Figure 3.7). This covering, the cornea, is a clear dome that protects the eye. The
light then passes through the pupil, the opening in the center of the iris. It continues through
the crystalline lens and the vitreous humor. The vitreous humor is a gel-like substance that
comprises the majority of the eye. Eventually, the light focuses on the retina where
electromagnetic light energy is transduced—that is, converted—into neural electrochemical
impulses (Blake, 2000). Vision is most acute in the fovea, which is a small, thin region of the
retina, the size of the head of a pin. When you look straight at an object, your eyes rotate so
that the image falls directly onto the fovea. Although the retina is only about as thick as a
single page in this book, it consists of
three main layers of neuronal tissue
(Figure 3.8). The first layer of neuronal
tissue—closest to the front,
outward-facing surface of the eye—is
the layer of ganglion cells, whose axons
constitute the optic nerve. The second
layer consists of three kinds of
interneuron cells. Amacrine cells and
horizontal cell make single lateral (i.e.,
horizontal) connections among adjacent
areas of the retina in
the middle layer of cells. Bipolar cells
make dual connections forward and
outward to
the ganglion cells, as well as backward and inward to the third layer of retinal cells.
The third layer of the retina contains the photoreceptors, which convert light
energy into electrochemical energy that is transmitted by neurons to the brain.
There are two kinds of photoreceptors—rods and cones. Each eye contains roughly
120 million rods and 8 million cones. Rods and cones differ not only in shape but
also in their compositions, locations, and responses to light. Within the rods and
cones are photopigments, chemical substances that react to light and transform
physical electromagnetic energy into an electrochemical neural impulse that can be
understood by the brain. The rods are long and thin photoreceptors. They are more
highly concentrated in the periphery of the retina than in the foveal region. The
rods are responsible for night vision and are sensitive to light and dark stimuli. The cones
are short and thick photoreceptors and allow for the perception of colour. They are more
highly concentrated in the foveal region than in the periphery of the retina (Durgin, 2000).
The neurochemical messages processed by the rods and cones of the retina travel via
the bipolar cells to the ganglion cells (see Goodale, 2000a, 2000b). The axons of the ganglion
cells in the eye collectively form the optic nerve for that eye. The optic nerves of the two eyes
join at the base of the brain to form the optic chiasma (see Figure 2.8 in Chapter 2). At this
point, the ganglion cells from the inward, or nasal, part of the retina—the part closer to your
nose—cross through the optic chiasma and extend to the opposite hemisphere of the brain.
The ganglion cells from the outward, or temporal area of the retina closer to your temple go
to the hemisphere on the same side of the body. The lens of each eye naturally inverts the
image of the world as it projects the image onto the retina. In this way, the message sent to
your brain is literally upside-down and backward. After being routed via the optic chiasma,
about 90% of the ganglion cells then go to the lateral geniculate nucleus of the thalamus.
From the thalamus, neurons carry information to the primary visual cortex (V1 or striate
cortex) in the occipital lobe of the brain. The visual cortex contains several processing areas.
Each area handles different kinds of visual information relating to intensity and quality,
including colour, location, depth, pattern, and form.
Bottom-Up Perception
Definition and Characteristics
Definition: Bottom-up perception, often referred to as data-driven perception, begins with
raw sensory input. It's a process where perception is constructed from the ground up, guided
primarily by sensory data.
Characteristics:
It follows a "building-block" approach, starting with the analysis of basic sensory features
and progressively building toward more complex interpretations.
There's minimal influence from prior knowledge or expectations at the initial stages.
Real-Life Examples
When you look at a jigsaw puzzle, your brain processes the colors, shapes, and edges of
individual puzzle pieces without any preconceived notions.
Detecting sudden movements, such as a car suddenly swerving on the road.
Applications:
Crucial for processing new or unfamiliar information.
Fundamental in early stages of perceptual processing, such as detecting edges and basic
visual features.
Top-Down Perception
Definition and Characteristics
Definition: Top-down perception, also known as concept-driven perception, is an approach
where prior knowledge, context, and cognitive processes influence the interpretation of
sensory input.
Characteristics:
Prior knowledge and context significantly shape perception, often guiding the interpretation
of sensory data.
It involves using existing concepts and mental models to understand sensory information.
Real-Life Examples
Reading a book where you automatically correct a typo because of your knowledge of
grammar and language rules.
Recognizing a friend's face in different lighting or clothing due to your familiarity with their
facial features.
Applications
Vital for understanding language, context, and social cues.
Allows us to quickly recognize familiar objects and make predictions based on prior
knowledge.
Interaction Between Bottom-Up and Top-Down Perception
Complementary Process: In reality, perception often involves a dynamic interplay between
bottom-up and top-down processing. The two processes complement each other, helping us
make sense of the world.
Role of Attention: Attention plays a crucial role in the interaction. It directs our focus to
specific sensory input, and top-down processes can influence where we allocate our attention.
Example: Consider being in a crowded room, and someone calls your name (top-down
processing), guiding your attention to that specific voice. Simultaneously, your sensory input
(bottom-up) helps you identify the source of the voice.
Real-Life Significance
Everyday Perception: We engage in both bottom-up and top-down perception continuously
throughout our daily lives. Understanding these processes is essential for grasping how we
navigate and interact with the world.
Cognitive Psychology: Top-down and bottom-up perception are at the core of cognitive
psychology, which delves into the intricacies of how our mental processes shape our
perceptions and behaviors.
Limitations and Ongoing Research
Ongoing research continues to explore the complexities of how bottom-up and top-down
processes interact. Researchers aim to uncover how these mechanisms can be more
effectively harnessed in various applications, including improving artificial intelligence,
enhancing human-computer interactions, and gaining a deeper understanding of human
cognition.
In conclusion, the interplay between bottom-up and top-down perception is pivotal to how we
make sense of the world. Understanding these processes illuminates the core principles of
how humans process information, and ongoing research further unravels the intricacies of
these mechanisms, guiding developments in various scientific and applied fields.
1. Introduction
The human perceptual system is remarkably adept at making sense of the world around us.
Even when presented with limited or ambiguous information, our brains can often quickly
and accurately recognize objects, events, and contexts. The Pattern and Context Recognition
Theory posits that perception is an active process where the brain matches incoming sensory
information to stored patterns, and then interprets this information within a broader context.
2. Basic Concepts
a) Pattern Recognition: This refers to the ability of the brain to identify regularities in sensory
input. For instance, upon seeing the letters "c", "a", and "t" in sequence, our brain recognizes
the pattern and interprets it as the word "cat".
b) Context Recognition: This is the process by which the brain uses surrounding information
to aid perception. For example, the word "lead" can be pronounced as "leed" or "led"
depending on the context in which it's used.
For example, reading handwritten notes relies on both processes. While the shapes of the
letters guide our recognition (bottom-up), our knowledge of the language and the context
helps fill in gaps or interpret ambiguous letters (top-down).
7. Studies and Experiments
a) The Word Superiority Effect (Reicher, 1969): Participants were quicker at recognizing
letters presented within a word compared to those presented in isolation or within a
non-word, highlighting the influence of context in recognition.
b) Contextual Cueing (Chun & Jiang, 1998): Participants were faster at locating a target in a
scene when the scene's configuration remained consistent across trials, even if they weren't
consciously aware of the repeating patterns.
Facial Recognition (Sternberg Pg. 116-118)
Introduction
Facial recognition is a specialized process in which the human visual system identifies and
processes faces differently from other objects. Faces convey a wealth of information, from
identity to emotions, intentions, and health. Given the social nature of humans, accurately
recognizing and interpreting faces is crucial for successful interpersonal interactions.
2. The Configural System
Facial recognition relies heavily on the configural system, which refers to the spatial
relationships between facial features. Unlike other objects, slight changes in the configuration
of facial features can lead to significant changes in perceived identity.
There are three primary components of configural processing:
a) First-order relational information: This is the basic configuration of facial features that are
common to all faces (e.g., two eyes above a nose, which is above a mouth).
b) Second-order relational information: This pertains to the specific spatial relationships
between features on an individual's face (e.g., the distance between the eyes or the size of the
nose relative to the mouth).
c) Holistic processing:This is the ability to perceive the face as a whole, rather than as a
collection of individual features. When we view a face, we don't just see eyes, a nose, and a
mouth; we see an integrated whole, and this gestalt perception is central to recognizing
familiar faces.
Example: In a busy cafe, as we speak to someone, their words are cues, so are their lip
movements and gestures. With the café's ambient noise, understanding them isn't just about
hearing. You're unconsciously triangulating meaning from audio, visual, and even contextual
cues.
4. Evaluation of Cues
Each cue provides a level of evidence for or against each potential perceptual category. Using
fuzzy logic, each cue is given a degree of membership to each category, ranging from 0 (not a
member) to 1 (full member).
*Real-life Example:* When your friend's lips form the word "coffee", but the noise drowns
out some sound, the lip-reading might offer a 0.90 confidence they said "coffee", while the
sound alone might only provide 0.60.
5. Independent Integration
The magic of the FLMP is in how it combines these cues. The model asserts that cues are
combined independently, which contrasts with other models that suggest cues interact.
Real-life Scenario: combining ingredients for a unique café recipe. The lip movement (0.90)
and voice sound (0.60) independently contribute, blending to shape the perception.
Mathematically, their integration multiplies them, giving a combined confidence of 0.90 x
0.60 = 0.54 for the word "coffee."
Deficits in Perception
Sternberg, Pg. 127-131
https://www.youtube.com/watch?v=QV59aAtznSg ;
https://www.visioncenter.org/conditions/motion-blindness/
https://medlineplus.gov/genetics/condition/achromatopsia/#causes
https://my.clevelandclinic.org/health/diseases/23421-visual-agnosia
https://www.britannica.com/science/optic-ataxia
https://www.britannica.com/science/prosopagnosia
https://www.britannica.com/science/color-blindness
https://byjus.com/biology/colour-blindness/