How the internal senses may connect sight and sound

Certain information is associated across the senses. Some of these crossmodal associations are shared by most people. For example, in the Bouba/Kiki-effect, more than 95% of people around the world match the spoken word “Bouba” with a rounded shape and the spoken word “Kiki” with an angular shape. Other crossmodal associations are subjective; while some people see colours when hearing music, others read braille in colour. And it seems these subjective associations between the external senses may be closely related to the internal senses. (See our blog for How anxious individuals perceive odours, Emotional perceptions associated with sound environments, and Growing into one’s own body.)

I have invited Dr Marina Iosifian, School of Divinity, University of St. Andrews to write this post about crossmodal associations between visual paintings and sounds. Dr Iosifian has contributed to several scientific papers and public outreach events on how the internal senses might create crossmodal associations between vision and hearing.

Have you ever noticed that certain colours seem to “fit” certain sounds? For example, dark red might feel like it matches a low, deep voice, while pink feels more like a high, light voice. These kinds of connections between different senses—such as sight and hearing—are called cross-modal associations. Researchers study them to understand how our brain brings together information from different senses to form a unified picture of the world, even though each sense works separately (our eyes only see, our ears only hear).

Why do these associations happen? One possible explanation involves emotion. For instance, dark red and a low voice might both feel connected to sadness, while pink and a high voice might both be linked to happiness or playfulness.

But emotions aren’t the only reason. Another explanation has to do with the body’s movements and sensations. For example, when people are asked to name two round tables—one large and one small—they often call the large one “mal” and the small one “mil.” This may be because of how our mouths move when saying these sounds: “mal” requires a wider, more open mouth shape, similar to something large, while “mil” involves smaller, tighter movements, like something small.

Girl in garden scene with a cat and a dog
The Garden Walk, by Emile Friant. Retrieved from WIKIART

In our study, we explored these bodily mechanisms—the ways our physical sensations and actions might shape how we connect sights and sounds—to better understand how cross-modal associations arise.

To explore these associations, we collected a set of sounds produced by the human body, such as the sound of someone drinking. We called these embodied sounds. To provide a contrast, we also included sounds that cannot be produced by the human body, such as electronic or synthesized sounds, which we called synthetic sounds.

Because we were interested in how sounds are connected with visual experiences, we also gathered a collection of images. These included two types of paintings: figurative paintings, which show recognizable subjects like people or objects (eg, The Garden Walk by Emile Friant), and abstract paintings, which do not represent specific things (eg, Sky above clouds by Georgia O’Keefe). We then paired the paintings with the sounds and asked our participants a simple question: “Does this sound and this painting fit together?”

Glowing horizon with fluffy white clouds below
Sky above clouds, by Georgia O’Keefe. Retrieved from Custom Prints, Georgia O’Keefe Museum

We found that embodied sounds were more often associated with figurative paintings, while synthetic sounds were more often linked with abstract paintings. This suggests that the body—and the way we experience sensations physically—plays an important role in how people connect what they see with what they hear.

Why might these associations occur? One possible explanation lies in the difference between concrete and abstract ways of thinking. Figurative paintings depict familiar, tangible things—people, objects, and scenes—so they may evoke more concrete thinking. Abstract paintings, on the other hand, invite a more imaginative or distant mindset.

Interestingly, previous research has shown that people tend to associate abstract art with more distant situations—whether in time or space—compared to figurative art. This idea is related to the psychological concept of psychological distance, where concrete things feel close to us and abstract things feel farther away. Our results suggest that this distinction between the concrete and the abstract may also shape how we connect sights and sounds.

Some researchers believe that psychological distance is one of the main concepts which can help us understand how the mind works. They developed the Construal Level Theory or CLT – which explains how our mental distance from things – called psychological distance – affect the way we think about them. Psychological distance can take many forms: something can feel distant in time (happening in the future or past), in space (far away), in social distance (involving people unlike us), or in hypotheticality (something uncertain or imaginary). It is suggested that people think about things that feel close to them—such as events happening soon or nearby—in a more concrete and detailed way. In contrast, things that feel distant in time or space, are understood in a more abstract and general way.

If abstract thinking is linked to distant, less embodied experiences, and concrete thinking to close, bodily ones, then the way we perceive and connect sounds and images may depend on how “distant” or “close” they feel to us psychologically. In other words, our sense of distance—both mental and sensory—may shape how we integrate what we see and hear.

Thus, the concept of abstraction offers valuable insight into how people interpret and understand the world around them. Art, in particular, provides a powerful way to explore these processes. Recent research suggests that engaging with beauty in art can encourage people to think in more abstract ways, making art an especially meaningful tool for studying perception and the connections between our senses.

See our blog for Activities; especially 85-87.

Some suggestions for further reading, listening, and watching:

Applying Bodily Sensation Maps to Art-Elicited Emotions

From Perception to Pleasure

From Sensory to Movement

How Does Your Body React to Art?

How Music Changes Your Mind

Processing Internal Sensory Messages

See What Your Brain Does When You Look at Art

How infants who are blind integrate tactile and auditory information

The brain appears to integrate simultaneous information from all the senses from birth. (See our blog for Multisensory processing.) However, when the infant is fully sighted, vision most often takes the lead. So what happens when vision is impaired?

This time, I have invited Stefania Petri, Unit for Visually Impaired People (the UVIP Lab), the Italian Institute of Technology, to write about the integration of tactile and auditory cues in infants. Stefania is part of the MySpace project, which investigates how infants and children who are blind process audio-tactile information. The project is led by Dr Monica Gori, Head of the UVIP Lab, and Stefania contributes to the development of the early intervention system iReach.

For newborns, vision is not only about recognising faces and objects. Sight guides movement, play, and exploration. It allows infants to coordinate their actions, interact with caregivers, and gradually make sense of the world. When vision is missing or severely impaired, these basic experiences are disrupted from the very beginning of life. Indeed, infants with visual impairments often face delays in motor development, difficulties in social interaction, and challenges in learning how to explore space.

 

Why Touch and Sound Matter

Vision usually guides the other senses, helping infants build a coherent sense of space. For a sighted child, seeing a toy, hearing its sound, and touching it all come together to form a single, integrated experience. To construct this spatial map, infants who are blind must rely on other senses, such as touch and hearing.

Both senses are present from birth, and both provide spatial cues: touch gives direct, body-centered information, while hearing allows orientation toward events and objects at a distance. Understanding how these two senses work together in the absence of vision is crucial for developing strategies that support the growth of children who are blind.

 

Studying multisensory spatial perception

To explore this, we used a well-established paradigm – presenting auditory and tactile stimulations on the hands of the infants. We used a non-invasive device and collected behavioural data. The stimulation could be presented in a congruent way, with touch and sound on the same side of the body. Or, in an incongruent way, for example, touch on the right and sound on the left-hand side. By comparing the responses from infants who were blind and infants who were sighted, it became possible to explore how the two groups oriented and how quickly they reacted under different conditions.

This method may seem simple, but it addresses a fundamental question: when vision is absent, how do infants resolve conflicts between touch and sound? And do they still benefit when both cues point in the same direction?

 

What We Found

The results revealed clear differences between the two groups:

  • When touch and sound are in conflict — for example, when a vibration is felt on one hand, but the sound comes from the opposite side — infants who are blind are less likely than their sighted peers to orient toward the sound. This suggests that they rely more strongly on tactile cues when making spatial decisions.
  • When touch and sound are congruent, infants who are blind show evidence of multisensory integration. Specifically, their reaction times are faster when both cues are presented together compared to when they are presented separately. While sighted infants tend to integrate such cues more efficiently, infants who are blind nevertheless reveal that they can combine information across senses in a beneficial way.

(Top) Four experimental conditions: auditory stimulation alone, congruent audio-tactile stimulation, incongruent audio-tactile stimulation, and tactile stimulation alone. (Bottom) Results: (a) percentage of orienting responses directed toward the auditory stimulus and (b) reaction times to the stimulus. Blue bars represent sighted infants (S), and yellow bars infants who were severely visually impaired. Bold black lines indicate statistically significant differences between conditions.

Published with permission. Gori, M., Campus, C., Signorini, S., Rivara, E., & Bremner, A. J. (2021). Multisensory spatial perception in visually impaired infants. Current Biology, 31(22), 5093-5101.e5. https://doi.org/10.1016/j.cub.2021.09.011

These findings highlight an important message: even without vision, multisensory stimulation, particularly the integration of sound and touch—can enhance performance and support the gradual development of spatial and motor skills.

 

Practical Implications

These insights are not just theoretical. They guide the development of both habilitation and rehabilitation strategies and supportive technologies. For instance, play-based training sessions that combine vibration with sound in congruent way could strengthen early sensorimotor skills and might help infants who are blind practice reaching and moving toward objects.
One practical example inspired by this research is iReach. iReach is a small, wearable system made of two units, or tags that communicate via wireless. By attaching an anchor tag to a bracelet on the child’s wrist and another tag to a toy, the device allows infants to sense changes in vibration and sound as they approach the object: As the child moves closer to the toy, the bracelet changes its vibration and sound, giving intuitive feedback about distance.

An early prototype has been tested in a safe, playful setting with sighted children who were blindfolded. In one of the activities, the children had to place objects into a box positioned farther away, which contained the spatial reference tag.

a) iReach units: Tag (left); Anchor (right). b) Example of Tag positions: Anchor on infant’s body midline (left); external object (right). c) Example of use of iReach: The sound emitter and waveform icons represent the auditory and tactile stimuli, respectively. An increase in icons size indicates a corresponding increase in feedback intensity and frequency.

Published with permission: Gori, M., Petri, S., Riberto, M., & Setti, W. (2025). iReach: New multisensory technology for early intervention in infants with visual impairments. Frontiers in Psychology, 16(May) https://do.org/10.3389/fpsyg.2025.1607528

When wearing the iReach bracelet, the children completed the task both faster and with more accurate movements. These early observations suggest that iReach can make exploration more intuitive and engaging for children who are blind.

Importantly, iReach is not a sensory substitution device, which often overload users with complex signals, it uses a child-friendly “language” of touch and sound to encourage active movement and exploration.

 

Conclusion

Infants who are blind grow up in a world where touch and hearing are the main senses that support their exploration of the world. Our studies show that they rely more on touch than on sound when the senses are in conflict, but they also benefit from integrating the two when the information is aligned. Recognizing how touch and sound work together, we can take important steps toward creating early interventions that respect children’s natural abilities and provide them with the best possible start in life.

See our blog for Activities; especially 79-81.

 

Some suggestions for further listening and watching:

Baby’s Fine and Gross Motor Skills

Baby Hearing Development

Beyond the Basic Senses

Get “Inside the Mind of a Baby”

Multisensory spatial perception in visually impaired infants

The Tactile System & Body Awareness In The First 3 Months

Vision Development: Newborn to 12 Months

What Your Baby Sees

Your baby’s sense of touch