Showing posts with label timbre. Show all posts
Showing posts with label timbre. Show all posts

Saturday, May 02, 2026

Isn't musicality more than rhythm?

Last month, we organized a follow-up to the 2014 Lorentz Workshop on Musicality in Leiden, The Netherlands. Twelve years later, it felt both exciting and meaningful to return to Leiden with a renewed focus: spectral percepts

While rhythm cognition has received substantial attention over the past decade, key perceptual dimensions of melodic cognition—especially timbre and pitch—remain comparatively underexplored. Many comparative studies still rely on simplified stimuli, such as pure tones, which may limit our understanding of how non-human animals perceive melodic structure. Recent findings suggest that pitch and timbre do not map uniformly across species, inviting us to rethink how these percepts are studied. 

We therefore deliberately shifted attention away from rhythm perception and production toward the perceptual and affective dimensions of melody, harmony, and timbre. In doing so, we revisited Darwin’s idea that animals may not only perceive melodies, but may also take pleasure in them (see workshop proposal). 

What made this workshop especially rewarding was the remarkable diversity of backgrounds and expertise in the room. Researchers from neurobiology, psychology, ethnomusicology, musicology, and evolutionary theory came together to examine the evolutionary and perceptual roles of pitch, timbre, and consonance. This breadth of perspectives allowed us to explore how these percepts vary across species, cultures, and contexts in ways no single discipline could address alone. 

By bringing together such a broad and inspiring group of researchers, the workshop generated new insights, strengthened interdisciplinary collaborations, and laid the groundwork for a more coherent framework on the evolution and cognition of musicality. 

A special issue is planned for Spring 2027, in which we will summarize the workshop’s findings, develop new research ideas, and outline a future agenda for musicality research. 

Photo credits: (cc) 2026 Bas Cornelissen and Lorentz Center.

Thursday, November 06, 2025

Can birds imitate Artoo-Detoo?

The research summarized in an infographic (Dam et al., 2025).

When you think of birds imitating sounds, parrots and starlings might come to mind. They’re famous for copying human speech, car alarms, and even ringtone melodies. But what happens when you challenge them with something really complex, like the electronic beeps and boops of R2-D2, the beloved Star Wars droid? Researchers from the University of Amsterdam and Leiden University put nine species of parrots and European starlings to the test.

Starlings versus parrots

It turns out that starlings had the upper hand when it came to mimicking the more complex 'multiphonic sounds. Thanks to the unique morphology of their vocal organ, the syrinx, which has two sound sources. This allows starlings to reproduce multiple tones at once—perfect for R2-D2-style chatter.

Parrots, on the other hand, are limited to producing one tone at a time (just like humans). Still, they held their own when it came to the simpler “monophonic” beeps of R2-D2. Interestingly, it weren’t the famously chatty African grey parrots or amazon parrots that did best, but the smaller species, like budgerigars and cockatiels. These little birds, often thought of as less impressive vocalists, actually outperformed the larger species in this specific task, likely by using different strategies to imitate sounds.

Even sounds from science fiction can teach us something real

The researchers call their study a fun but powerful window into how anatomy, like the structure of a bird’s vocal organ, can shape the limits and possibilities of their vocal skills. It is the first time that so many different species all produced the same complex sounds, which finally allows for a direct comparison. This shows that even sounds from science fiction can teach us something real about the evolution of communication and learning in animals.

And here’s the cool part: much of the sound data came from pet owners and bird lovers participating in citizen science through the Bird Singalong Project. With their help, the researchers were able to gather a richer, more diverse collection of bird sounds than ever before, proving that science doesn't always have to happen in a lab.

Reference

Dam, N.C.P., Honing, H. & M.J. Spierings (2025). What imitating an iconic robot reveals on allospecific vocal imitation in parrots and starlings. Scientific Reports, 15, 36816. https://doi.org/10.1038/s41598-025-23444-7

Friday, October 03, 2025

Piano touch unraveled. Touché?

[From Kuromiya et al., 2025: Figure 1B]

 

[Adapted from interview by Elleke Bal, Trouw, 3 October 2025]

Two professional pianists may perform on the exact same piano, in the same concert hall, and even play the same notes at the same tempo. Yet, through the way they touch the keys, they are able to produce strikingly different sounds from the instrument.

This so-called timbre—the tonal color or quality of sound—has long been a subject of fascination and debate among musicians and listeners alike. Consider, for example, the crystalline clarity of Glenn Gould versus the warmth of Sviatoslav Richter. But what exactly constitutes clarity or warmth? For pianists, these are intuitive concepts. For scientists, however, the challenge has been to find objective evidence that such distinctions arise from unique motor actions at the keyboard.

The researchers examined which motor skills underpinned these differences. They found that timbral variation was strongly associated with a limited set of keystroke parameters: the velocity of key descent, the temporal spacing between successive key presses, and the synchrony between hands. Crucially, one factor emerged as particularly decisive: the acceleration of the key movement at the precise moment the hammer disengages. According to the authors, this acceleration largely determines the resulting timbre (Kuromiya et al., 2025).

(c) 2025 Trouw

This study demonstrates the extraordinary precision achieved by highly trained pianists. Nuances in timing and velocity of a few milliseconds can shape timbre in ways that are musically significant. These micro-timing differences are the product of extensive practice and experimentation at the instrument. However, the study overlooked one key factor: the velocity of the hammer striking the string. Without measuring hammer dynamics, the account of timbre remains incomplete. Companies such as Yamaha have long recognized this; their Disklavier Pro system, for example, replicates hammer velocity to convincingly reproduce the playing of pianists like Glenn Gould.

Ultimately, it is the hammer which, at the moment it is released from the piano’s action mechanism, independently carries the cumulative timing and dynamic input of the performer. Its subsequent trajectory toward the string—and the resulting timbre—is determined by its momentum, defined by the combined effects of its velocity and mass.

Does reducing the artistry of great pianists to numerical parameters diminish the magic of a performance? I don’t think so: This research only reinforces the extraordinary dexterity, control, and timing that distinguish master pianists.

References 
Kuromiya, K., et al. (2025). Motor origins of timbre in piano performance, Proc. Natl. Acad. Sci.,122 (39) e2425073122, https://doi.org/10.1073/pnas.2425073122 
Goebl, W., & Palmer, C. (2008). Tactile feedback and timing accuracy in piano performance. Experimental Brain Research, 186 (3), 471-479 DOI: 10.1007/s00221-007-1252-1 

 

Tuesday, September 02, 2025

Are there controversies in pitch and timbre perception research? [in 333 words]

European Starling (Sturnus vulgaris)

At the heart of human musicality lie fundamental questions about how we perceive sound. In the coming academic year our group will dedicate several meetings on exploring and clarifying the spectral percepts that might underlie musicality with an agenda set around some enduring controversies. These span the roles of learning, culture, and cross-species comparisons, as well as evolutionary explanations for why music holds such sway over human minds. 

Among the most debated topics is the relationship between pitch and timbre perception. Both pitch and timbre are percepts: mental constructs arising from acoustic input. In humans, pitch perception is central to melodic recognition. When we hear a melody, we tend to identify it by its sequence of relative pitches—hearing it as the “same” tune regardless of changes in timbre, loudness, or duration. This reliance on relative pitch is a cornerstone of human music cognition. 

But is pitch such a universal perceptual anchor? For years, researchers assumed so, pointing to songbirds as an obvious parallel. Birds, it was thought, must also use pitch cues, though often in the form of absolute rather than relative pitch. Yet recent evidence complicates this narrative. In a striking study, Bregman et al. (2016) reported that European starlings do not, in fact, rely on pitch when recognizing sequences of complex harmonic tones. Instead, they appear to attend more closely to spectral shape, or the broader distribution of energy across frequencies. 

This finding raises a further question: is it really the spectral envelope (i.e. spectral shape) that matters, or something more subtle? Because the methods used—particularly contour-preserving noise vocoding—leave open another possibility: birds may actually be attuned to fine spectral-temporal modulations, the intricate contours woven into sound. Such results remind us that perceptual categories humans take for granted may not map cleanly onto other species, and that the universality of pitch as a cognitive anchor remains an open, and fascinating, controversy (cf. Patel, 2017; ten Cate & Honing, 2025). 

N.B. These entries are part of a new series of explorations on the notion of Spectral Percepts (in 333 words each). 

Bregman, M. R., Patel, A. D. & Gentner, T. Q. (2016). Songbirds use spectral shape, not pitch, for sound pattern recognition. Proceedings of the National Academy of Sciences, 113(6), 1666–1671. doi: 10.1073/pnas.1515380113 

Patel, A. D. (2017). Why Doesn’t a Songbird (the European Starling) Use Pitch to Recognize Tone Sequences? The Informational Independence Hypothesis. Comparative Cognition & Behavior Reviews, 12, 19–32. doi: 10.3819/CCBR.2017.120003 

ten Cate, C. & Honing, H. (2025). Precursors of music and language in animals. In D. Sammler (Ed.), The Oxford Handbook of Language and Music. Oxford University Press. doi: 10.1093/oxfordhb/9780192894700.001.0001

Sunday, August 31, 2025

Is consonance a biological or a cultural phenomenon? [in 333 words]

Chick in consonance experiment (Chiandetti & Vallortigara, 2011).

The distinction between consonance and dissonance has long occupied a central place in the scientific study of auditory perception and music cognition. Consonant intervals are typically described as stable, harmonious, or pleasing, whereas dissonant intervals are often characterized as tense, unstable, or even harsh. Yet even these seemingly straightforward descriptions quickly lead to methodological debate. 

A central difficulty arises from the frequent conflation of “dissonance” with “roughness.” Roughness refers to a physiological effect caused by closely spaced frequencies interacting on the basilar membrane of the inner ear. This phenomenon is measurable, consistent, and largely universal across listeners. Consonance, however, is not reducible to physiology alone. Recent research emphasizes that consonance is a multidimensional construct, shaped by both acoustic properties such as harmonicity and by layers of cognitive and cultural familiarity (Lahdelma & Eerola, 2020). 

This controversy can be framed around two major questions (Harrison, 2021). First, do humans possess an innate preference for consonance over dissonance? Second, if such a preference exists, how might it be explained in evolutionary terms? A landmark study by McDermott et al. (2016) with the Tsimane’, an Amazonian group minimally exposed to Western music, found no consistent preference for consonant over dissonant intervals. Their conclusion was that what many listeners call “pleasant” is primarily shaped by cultural experience. 

This interpretation has been vigorously challenged. Bowling et al. (2017) cite empirical evidence from human infants (Trainor et al., 2002) and even non-human animals (Chiandetti & Vallortigara, 2011) that points toward at least some innate, hardwired auditory sensitivity. If so, consonance may reflect evolutionary selective pressures, possibly related to the spectral composition of human vocalizations and the neurophysiological mechanisms underlying pitch perception and auditory scene analysis. 

In the end, consonance appears to be neither purely biological nor purely cultural. Our ears detect roughness and harmonicity, but our minds interpret these sensations through cultural frameworks. What sounds stable in one tradition may sound unfamiliar in another. The consonance controversy thus highlights music cognition as an intricate interplay between biology and culture. 

N.B. These entries are part of a new series of explorations on the notion of Spectral Percepts (in 333 words each).

References

Bowling, D. L., Hoeschele, M., Gill, K. Z. & Fitch, W. T. (2017). The nature and nurture of musical consonance. Music Perception, 118–121. 

Chiandetti, C. & Vallortigara, G. (2011). Chicks like consonant music. Psychological Science, 22(10), 1270– 1273. https://doi.org/10.1177/0956797611418244 
 
Harrison, P. M. C. (2021). Three Questions concerning Consonance Perception. Music Perception, 337–339. https://doi.org/10.1525/MP.2021.38.3.337 
 
Lahdelma, I. & Eerola, T. (2020). Cultural familiarity and musical expertise impact the pleasantness of consonance/dissonance but not its perceived tension. Scientific Reports, 10(1), 8693. https://doi.org/10.1038/s41598-020-65615-8 
 
McDermott, J. H., Schultz, A. F., Undurraga, E. A. & Godoy, R. A. (2016). Indifference to dissonance in native Amazonians reveals cultural variation in music perception. Nature, 25, 21–25. https://doi.org/10.1038/nature18635 
 
Trainor, L. J. & Unrau, A. (2012). Development of Pitch and Music Perception. In Human Auditory Development (pp. 223–254). Springer. https://doi.org/10.1007/978-1-4614-1421-6_8

Wednesday, August 27, 2025

What makes two melodies feel like the same song? [in 333 words]

(cf. Krumhansl, 1989).

One of the most intriguing questions in music cognition research is also one of the simplest: when are two melodies experienced as the same?

At first glance, the answer might seem obvious — they share the same notes, in the same order, with the same rhythm. But a closer look, across cultures and even across species, reveals a more complex picture. What our brains latch onto when recognizing a tune involves a web of spectral percepts — the fundamental features of sound that guide humans and other animals in interpreting auditory patterns. This may sound like a niche research topic, but it lies at the heart of debates about authorship, originality, and musical ownership.

Consider hearing a melody played in a different key or on an unfamiliar instrument. Most people can still recognize it. How is this possible? Explanations often point to intervallic structure — the sequence of pitch intervals between notes — the contour, which is the overall shape of a melody as it rises and falls, or timbre, often described as the “color” of sound, including brightness, texture, and loudness.

For decades, music research treated timbre as secondary — something layered over supposedly “meaningful” musical features like pitch and rhythm (cf. McAdams & Cunible, 1992). Increasing evidence now suggests timbre is not merely decorative but a core perceptual building block. Timbre may also support “relative listening,” the ability to track patterns of change across different features. Exploring it carefully could reveal flexible and universal aspects of music cognition previously underestimated.

Recognizing that humans and non-human animals may rely on different spectral cues is equally crucial for understanding music’s evolutionary roots. A melody meaningful to humans may not register as such for a zebra finch — and vice versa.

By broadening music cognition research to include timbre, spectral contour, and species-specific strategies, scientists hope to uncover the shared perceptual foundations of musicality. Such work moves us closer to answering a deceptively simple but deeply complex question: what truly makes two melodies feel like the same song?

N.B. These entries are part of a new series of explorations on the notion of Spectral Percepts (in 333 words each).

References

McAdams, S, & Cunible, J-C (1992). Perception of timbral analogies. Philosophical Transactions of the Royal Society B: Biological Sciences, 336, 383-389. 

Krumhansl, C. L. (1989). Why is musical timbre so hard to understand? In S. Nielzén & O. Olsson (Eds.), Structure and perception of electroacoustic sound and music (pp. 43– 53). Elsevier.