smoking_man
Member
- Joined
- Jan 25, 2026
- Messages
- 31
- Likes
- 2
The 16bit 44.1khz standard looses musical information from the original performance, this cannot be recreated by an ordinary DAC. Will AI eventually fill in the gaps?
Any data to back this assertion?The 16bit 44.1khz standard looses musical information from the original performance, this cannot be recreated by an ordinary DAC. Will AI eventually fill in the gaps?
It should be obvious there is only one or two microphones recording the sound scape in a live performance and all digital music is frequency band limited. The resolution of cd means that the wave form is not exactly the same as when it was recorded because of the sample rate. Then there is reproduction in your listening room that needs optimising with AI.Any data to back this assertion?
The resolution of the CD is higher than the resolution of your earsIt should be obvious there is only one or two microphones recording the sound scape in a live performance and all digital music is frequency band limited. The resolution of cd means that the wave form is not exactly the same as when it was recorded because of the sample rate. Then there is reproduction in your listening room that needs optimising with AI.
If you capture a performance with only one microphone, you only have that amount of spatial information. No AI in the world could faithfully recreate the missing info from the other 2 or 10 mics you would have placed in the room to gather more info on how it sounds in the other corners of the room. You can just add something like a virtual sorround DSP to your output at home, which tries to upmix/"up-spatialise" the resulting sound to somehow sound "better" on your specific setup (stereo speakers, headphones, 5.1, whatever). You could also train a machine learning/"AI" model to perform that upmix and it would probably do it well. This does not re-create the lost information, though - the DSP or the AI model just alter the signal in a way that creates a specific illusion of how it might have sounded in a simulated room/sorround setup.a musical performance does not just create one wave, like cd. It produces a multitude of waves which are simplified down to one waveform when recorded.
The wave form is never the same as the event in the room—microphones are transducers, after all, subject to all the usual issues with transducers. It's not for nothing that recording engineers who are making money have a lot of microphones to work with as they all sound different. Nobody records with just two microphones for professional work with the rare exception of minimalist recordings of chamber groups. The resolution of the waveform is far more likely to be accurate with digital, Redbook standard recordings than analog recordings. Analog tape has inherent and quite audible flaws baked into the formula, like various speed variations and the loss of high frequency content as the level increases. Digital recording, at 16 bits, sample rate of 44.1 kHz, covers the dynamic range that would be audible in a domestic environment, the upper limit of 20.5 kHz covers the upper limit of what we humans can hear. I've done a lot of work transcribing LPs to reel to reel tape at 15 ips. Those transfers lost high-frequency energy, had audible hiss, simply weren't as good at transfers to 16/44.1 digital media.It should be obvious there is only one or two microphones recording the sound scape in a live performance and all digital music is frequency band limited. The resolution of cd means that the wave form is not exactly the same as when it was recorded because of the sample rate. Then there is reproduction in your listening room that needs optimising with AI.
A wave is not a field. Read carefully and follow the link about superposition. At any point in space and time, there is only one exact pressure level. I don't know what more to tell you.If there was only one sound wave, why would the sound wave change as I move a microphone in anechoic chamber?
Yes it can. You can create full songs and albums from just a simple prompt now. Therefore, you can create pretty much any audio effect using AI.AI can create near perfect video from its data set including voice and background sound, I think it would be able to manage to create near perfect audio from nothing which does not require video.
A "sound" (acoustic energy in the human frequency range) is a sum of many pressure disturbance waves in the air. The red book audio standard is enough to store these waves information in the digital domain for a human.a musical performance does not just create one wave, like cd. It produces a multitude of waves which are simplified down to one waveform when recorded.
There no "only one sound wave", as I said before, the sound is composed by a sum of waves in superposition. You might have seen a RTA spectrogram (FFT), the electric signal analyzed by the RTA is just one signal, but "inside" it there is a lot of waves which the FFT can decompose partially those waves. This change you talking about is related to the source and the microphone position, each point in scape there will be a different summation of waves. Supposing an ideal anechoic chamber with an ideal omnidirectional point source, an ideal microphone would only "perceive" the pressure level variation, not a frequency response change.If there was only one sound wave, why would the sound wave change as I move a microphone in anechoic chamber?
Nope. You can sum a bunch of sines together and create a complex waveform. There's no limit on how many.a musical performance does not just create one wave, like cd. It produces a multitude of waves which are simplified down to one waveform when recorded.