Jim Taylor
Major Contributor
- Joined
- Oct 22, 2020
- Messages
- 1,561
- Likes
- 3,367
TAKE MEASUREMENTS.
TAKE MEASUREMENTS.
TAKE MEASUREMENTS.
I like that some of us will depend on this AI machine to interpret audio data into flowery words like we're putting all our "hopes" on the APx555.Many of us throw up our hands when equipment is described as "fast", "slow", "crisp", "warm", etc. It seems impossible to relate these terms to measurable characteristics.
I have a slightly more optimistic view, in that subjective descriptions must be correlated with what people hear, and what people hear tends to be correlated (imperfectly) to measurable output.
To translate subjective descriptions into objective measurements (or the other way around, which might be more interesting), I propose that a machine learning model could be used.
The model would correlate subjective terms used (how often people say a speaker is "crisp", for example) with the measurements of the equipment relative to the median measurements of all equipment in its category.
The output would be a map of how semantically close certain audiophile words are to each other, and how use of those words correlates with measurable characteristics. Imagine a word cloud that groups words like "sharp, bright, tinny" together, and then displays measured characteristics that correlate with those terms - elevated response above 4khz, above-average H3 distortion in that range, etc.
This would be interesting in its own right, but if the ML was sophisticated enough, and fed enough data, it might also reveal trends in preference / subjective experience that go beyond the current preference score models. For example, you might unexpectedly find that some aspect of vertical directivity correlates with "warmth" or "speed", for example. I don't know.
I don't have anywhere near the skills to execute such a project, but it seems like a way to solve the "problem" of people using flowery language that, to many of us, is currently worse than useless. It might also reveal that things audiophiles consider to be "beyond science" are actually very well correlated with simple measurements. Which would be progress in and of itself.
Devil's advocate here: I take 'plankton' to mean 'very fine detail in the recording'. Perception of "plankton" would therefore depend on a good transient response and very low noise and distortion, particularly IMD. So I think you can relate even something like "plankton" to measurable quantities. The difficulty comes in knowing whether anyone interprets a colorful term the same way.
This would be interesting in its own right, but if the ML was sophisticated enough, and fed enough data, it might also reveal trends in preference / subjective experience that go beyond the current preference score models. For example, you might unexpectedly find that some aspect of vertical directivity correlates with "warmth" or "speed", for example. I don't know.
It's even worse than the chicken egg thing.chicken - egg problem. To make measurements have any meaning to individuals with their own 'vocabulary' they need to know how to convert squiglies to their vocabulary (understand a full measurement suite) and that takes quite a bit of knowledge, experience. Even if 'agreed upon' descriptors are used (as an explanation along with the measurements) this might not be fully understood by people with a very different vocabulary. They would have to use the elaborate 'descriptors descriptions' list to understand what the 'agreed upon' descriptors means.
For ASR, for instance, such a vocabulary could be agreed upon.... maybe.
It's even worse than the chicken egg thing.
The point is, the measurements give us the vocabulary we need. Numbers. DONE.
AKG K271 Listening Tests
While not perfect, out of the box tonality was close to what I expected to hear. My reference tracks immediately sounded (almost) right.
I don't think these discussions even matter. Subjectivity is subjectivity. It cannot be agreed upon. Which is why we have measurements.So how would someone that hasn't got the faintest idea what the plot below tells the initiated (the numbers) have to decide if they would like the sound of this headphone ?
How would someone 'describe' the sound of this headphone based on the numbers ?
AKG K271 Listening Tests
While not perfect, out of the box tonality was close to what I expected to hear. My reference tracks immediately sounded (almost) right.
Someone wouldn't. The numbers "speak" for themselves. In other words, the numbers are a language, and that language is separate and different from the language we use to speak to each other.
And like any language, you have to learn it to be fluent in it.
Jim
While that is true the vast majority of people that listen to music is not audiophile minded (music fidelity addicts) and only a small portion of those actually understand the numbers and a much smaller number of them is 'fluent' in interpreting the numbers. The vast majority of 'music consumers' are not going to learn 'numbers language'. They want to read 'flowery and positive' words of 'reviewers' instead. So those words, in practice, are more important than 'numbers' to the vast majority of people. They don't care about 'numbers'.
So a good 'translation' or 'vocabulary/description' is very important.
This 'numbers language' is no option for every day life. Flowery wording and 'descriptions' are a more universal language that is easier to learn.
A 'vocabulary' with descriptions is easier to understand than learning to interpret measurements and taking that skill far beyond 'mount stupid'.
Yes, numbers are a language and when understood say a lot more than the many 'flowery wording vocabularies' as these all differ anyway. But not everyone that goes on vacation has learned to speak fluent Swahili, not even when often visiting the countries it is spoken in.
. Why do ASR's reviews often include a purely subject "Rate this device" poll? At best it is a measurement of how much agreement there is in the value of the measured objective performance. It seems antithetical to the main premise of this forum. If it was a poll of the subjective performance of the device under test by actual users of the device that would potentially add real value.
Has nothing to do with trust. It's just another set of measurement to compare to my capability. I know the gaps in my hearing because of annual audiograms. Around 4K I need a little boost.You don't trust that they can hear if they don't hear so well?
I think what you are trying to achieve is similar to the puzzle which has been baffling marketeers/researchers for decades, as to what makes a hit....To translate subjective descriptions into objective measurements...
I was being facetious.Has nothing to do with trust. It's just another set of measurement to compare to my capability. I know the gaps in my hearing because of annual audiograms. Around 4K I need a little boost.
TAKE MEASUREMENTS.
It's even worse than the chicken egg thing.
The point is, the measurements give us the vocabulary we need. Numbers. DONE.
I don't think these discussions even matter. Subjectivity is subjectivity. It cannot be agreed upon. Which is why we have measurements.
People should just go listen and buy. Some will realize their ears are the weakest links in the listening chain and some won't.
There simply is NO WAY to describe audio to another person when we live in a world where people think 3 foot lengths of AC cable makes your system sound different.
If it measures good it is good. Good specs are the definition of high fidelity. FULL STOP.
If this even would work, how would you get enough data to make relevant conclusions, and how would you set up a machine learning algorithm to do it. Surely there's not enough organised data in order for this to work. Let alone the different rooms that people listen in and other variables - but I get that if you have enough data then some of that could be averaged out. I just don't know how the system would work to gather the data in first place, and secondly would there be enough data, and thirdly how do you design the algorithm. I don't know much about machine learning, but it seems like there are some sticking points.Many of us throw up our hands when equipment is described as "fast", "slow", "crisp", "warm", etc. It seems impossible to relate these terms to measurable characteristics.
I have a slightly more optimistic view, in that subjective descriptions must be correlated with what people hear, and what people hear tends to be correlated (imperfectly) to measurable output.
To translate subjective descriptions into objective measurements (or the other way around, which might be more interesting), I propose that a machine learning model could be used.
The model would correlate subjective terms used (how often people say a speaker is "crisp", for example) with the measurements of the equipment relative to the median measurements of all equipment in its category.
The output would be a map of how semantically close certain audiophile words are to each other, and how use of those words correlates with measurable characteristics. Imagine a word cloud that groups words like "sharp, bright, tinny" together, and then displays measured characteristics that correlate with those terms - elevated response above 4khz, above-average H3 distortion in that range, etc.
This would be interesting in its own right, but if the ML was sophisticated enough, and fed enough data, it might also reveal trends in preference / subjective experience that go beyond the current preference score models. For example, you might unexpectedly find that some aspect of vertical directivity correlates with "warmth" or "speed", for example. I don't know.
I don't have anywhere near the skills to execute such a project, but it seems like a way to solve the "problem" of people using flowery language that, to many of us, is currently worse than useless. It might also reveal that things audiophiles consider to be "beyond science" are actually very well correlated with simple measurements. Which would be progress in and of itself.
If this even would work, how would you get enough data to make relevant conclusions, and how would you set up a machine learning algorithm to do it. Surely there's not enough organised data in order for this to work. Let alone the different rooms that people listen in and other variables - but I get that if you have enough data then some of that could be averaged out. I just don't know how the system would work to gather the data in first place, and secondly would there be enough data, and thirdly how do you design the algorithm. I don't know much about machine learning, but it seems like there are some sticking points.