Your last paragraph, as a non-musician, is what my rigid world view was not understanding. I was convinced that an instantaneous snapshot of the character of the sound was timbre. My apologies to all. The mix of various Fourier components changes through the entire duration of a note on an instrument. The piano strings when struck and as the sound decays give a changing mix of harmonics and whatever other frequencies make up the character of a piano sound. The entire event of a note is what identifies it as having the timbre of a piano created note. Gotcha.Musical definitions (from memory, hopefully correct):
There are a few other more esoteric harmonic terms (musical and otherwise) but those are the main ones IMO.
- harmonic = multiples of the fundamental pitch (frequency, note)
- subharmonic = submultiples of the fundamental pitch
- enharmonic = different notation of the same pitch (e.g. D#, Eb are the same frequency but notated differently per the key signature, and may be slightly different in pitch depending upon the chord)
- inharmonic = discordant, not harmonically related (typical example is a drum or other percussion instrument, but other instruments can create inharmonic tones depending upon the instrument and the player)
Formant refers to a region of resonance, or amplitude peak, with the center of the region generally referred to as the formant frequency. Usually refers to vocal sounds, though occasionally more disparate things like spread spectrum systems or even the grouping of strings for a piano note.
A reminder that timbre involves not only frequency but also time-varying amplitude information. To distinguish timbre, spectral analysis must reflect the time-varying nature of the waveforms and not just basic frequency information.
Yes! Each tone (harmonic or note) can have a different envelope, i.e. attack, decay, sustain, and release characteristics. That is part of timbre and one reason it is hard to analyze. Think of a sound from two different instruments, one that has the higher frequencies dying off quickly, and another the lower frequencies, for a handwaving thought problem. The percussive attack of a piano hammer vs. the softer attack of a flute playing the same pitch, for instance, and how the piano string causes the sound to decay less rapidly than when the flautist stops the note by stopping the air flow. Not only is the instantaneous spectral content different over time, but the overall envelope of individual frequencies within the sound varies.Your last paragraph, as a non-musician, is what my rigid world view was not understanding. I was convinced that an instantaneous snapshot of the character of the sound was timbre. My apologies to all. The mix of various Fourier components changes through the entire duration of a note on an instrument. The piano strings when struck and as the sound decays give a changing mix of harmonics and whatever other frequencies make up the character of a piano sound. The entire event of a note is what identifies it as having the timbre of a piano created note. Gotcha.
It's a bit geeky, but I love listening to piano notes as they decay. The interplay, decay and sometimes the re-emergence of a previously faded harmonic is very satisfying...The piano strings when struck and as the sound decays give a changing mix of harmonics and whatever other frequencies make up the character of a piano sound.
Cymbals do still have a fundamental, though in normal use it may be hard to discern.Yes, best instrument examples are cymbals, lots of frequencies no pitch but they have timbre.
Maybe, maybe not. There have been experiments done that show that we identify some instruments more from the initial transient than the decay of the note, by for example attaching the initial transient of a piano note to the decay of another instrument.Your last paragraph, as a non-musician, is what my rigid world view was not understanding. I was convinced that an instantaneous snapshot of the character of the sound was timbre. My apologies to all. The mix of various Fourier components changes through the entire duration of a note on an instrument. The piano strings when struck and as the sound decays give a changing mix of harmonics and whatever other frequencies make up the character of a piano sound. The entire event of a note is what identifies it as having the timbre of a piano created note. Gotcha.
Excite the cymbal's edge with a bow (bass bows are probably best) and you can readily hear the fundamental and the harmonics.Cymbals do still have a fundamental, though in normal use it may be hard to discern.
Question: How were the listeners (test subjects) in these experiments "qualified*" or were they even qualified at all? Skilled musicians, those familiar with the instruments in the test, would be probably far more difficult, if not impossible, to duped by splicing the attack of one instrument onto the sustained tone of another. Hiding the splices, even if done in the digital domain, would be very challenging.Maybe, maybe not. There have been experiments done that show that we identify some instruments more from the initial transient than the decay of the note, by for example attaching the initial transient of a piano note to the decay of another instrument.
I’ll need to find the details again, but I believe the abilities of the listeners weren’t specified in the summary I read for these tests.Question: How were the listeners (test subjects) in these experiments "qualified*" or were they even qualified at all? Skilled musicians, those familiar with the instruments in the test, would be probably far more difficult, if not impossible, to duped by splicing the attack of one instrument onto the sustained tone of another. Hiding the splices, even if done in the digital domain, would be very challenging.
* Assessed for age, previous experience, relevant skills, aural health, etc.
Very good point. I’d better find a more accurate term for those non harmonic frequencies. Sloppy of me. What do you call them? I’ve seen ‘inharmonic overtones’ used to describe them just now while searching.
100% agree that often the attack is the most important part of timbre, but I'll also argue that it's pretty hard to identify most instruments based on a static snapshot.Maybe, maybe not. There have been experiments done that show that we identify some instruments more from the initial transient than the decay of the note, by for example attaching the initial transient of a piano note to the decay of another instrument.
Once again, our brains take shortcuts. Sometimes we need to identify a sound instantaneously (our ancestors couldnt wait for the lion's entire roar), and it is also helpful to identify quickly the notes of one instrument in an ensemble.
Question: Does your additive synthesizer produce (or account for) the non-harmonically correlated sounds typically associated with natural (I.e., physical) instruments? If not or if not systematically, it would non-sensical to assume poor timbral identification in any test based on simple additive synthesis to be anything other than the result of an incomplete (or inadequate) instrumental approximation.With additive synthesizers, you can "freeze" the harmonics and listen to them without any variation over time. When you do that, most of the time the timbre is entirely lost. It's a lot like looking at a single row of pixels from a photo and trying to identify what it is.
Pretty critical information, n'est pas?I’ll need to find the details again, but I believe the abilities of the listeners weren’t specified in the summary I read for these tests.
Yes, the kind of thing I'm talking about is really just FFT based resynthesis, so it takes a recording of the instrument and lets you manipulate the partials, time-stretch, and so on.Question: Does your additive synthesizer produce (or account for) the non-harmonically correlated sounds typically associated with natural (I.e., physical) instruments? If not or if not systematically, it would non-sensical to assume poor timbral identification in any test based on simple additive synthesis to be anything other than the result of an incomplete (or inadequate) instrumental approximation.
I think there is a substantial difference.
When we 'hit' an instrument it produces a tone with a specific timbre (spectrum that has attack, sustain, decay and changes during that time).
A speaker, when 'hit' with a pulse (be it DC or Dirac pulse) generates a dry sounding 'tick' with very short attack, no sustain and quick decay.
When comparing speakers the 'tick' may well sound a bit different but does not produce a 'tone' with a clearly recognizable 'timbre'.
What a speaker will do, because of FR variations and resonances, is change (affect) the timbre of reproduced instruments but doesn't have one by itself.
A speaker does have a distinct tonal balance but not a timbre (it isn't an instrument).
A speaker can also change the timbre.
A (bass)guitar speaker in a cabinet can be said to be part of the instrument and determines the timbre (through FR irregularities and resonances) but those speakers are part of an instrument. One does not play a guitar amp with Dirac pulses to create a certain 'sound' either but uses them to change(modify) the sound of an instrument.
So have timbre does not equal affect timbre because timbre is from an instrument which a speaker isn't.
It’s clearly a 1987 Nissan Maxima horn.Yes, the kind of thing I'm talking about is really just FFT based resynthesis, so it takes a recording of the instrument and lets you manipulate the partials, time-stretch, and so on.
I did a quick example, I won't tell you the instrument so as not to give it away, but even if you get it right away, I think you will agree it's not as recognizable as it would be if the harmonics varied over time, even though all the partials are present and accounted for.
We were discussing the wonders of additive synthesis, not FFT based resynthesis. That kind of gets us off target. But for the sake of the game I'll play along. Your sample sounds like a poorly synthesized oboe - might I say it sounds very 1980s Casioesque? - but stripped of the tonal evolution that occurs with real instruments, it easily could be any number of things, even iPunchCholla's 1987 Nissan Maxima horn or the death cry of a gerbil. In short, it doesn't sound remotely like a conventional, real-world instrument, regardless of its ultimate origin. (BTW, I'm rooting for formant based analysis, not FFT.)Yes, the kind of thing I'm talking about is really just FFT based resynthesis, so it takes a recording of the instrument and lets you manipulate the partials, time-stretch, and so on.
I did a quick example, I won't tell you the instrument so as not to give it away, but even if you get it right away, I think you will agree it's not as recognizable as it would be if the harmonics varied over time, even though all the partials are present and accounted for.