I think we commented past each other and my previous comment addresses a lot, but, are you sure?
Yeah I'm sure. What you cited was the DOJ position until the National Academy of Sciences report came out (2009?) which confirmed that it was junk science. As a result of the report, new standards were issued. If you want to take the deep dive into it I can post all of the reports, etc. Attached is the current draft standard and below is the significant language about how trashy this all was. The report, and quote below, are from 2023. To get a voice identification in now you have to be a certified and accredited in phonetics. You have to use both a spectogram and (wait for it) listen to the output and, based on education, training and experience, be able to discuss the difference in vowel sound pronunciation and other characteristics and give an opinion on whether it is same speaker or not. However, courts will not let that expert say "its a match." They now limit the opinion to "inconclusive, cannot be excluded as the speaker, or excluded as the speaker."
ASR, also mentioned in the standard, is Automated Speech Recognition and gives a numerical score on the probability of a match. As far as I know, it hasn't been admitted anywhere, it is not admissible in Texas.
The science is there to do it, and I'm sure the private sector if they have figured out a way to monetize, they have put the money in it to do so for voice identification. The was an article in Science (the world's premier peer-reviewed publication) related to the study and identification of the timbral properties of certain species of wood in order to identify which woods are better for musical instruments, and, to identify specific wood blanks as having superior timbral qualities for musical instruments. What they could tell you about wood was quite extraordinary.
From the Attached proposed standard (emphasis in bold in mine, the number refer to line numbers it he original).
5.3.1 The so-called "voiceprints" were the product of sound spectrography, a technology 85 carried forward even to the present day, which is still of great utility to speech scientists. 86 The most notable scientific failing of the voiceprint method was that it did not provide 87 examiners with the vocal output (i.e., the audio). This inevitably obscured the phonetic 88 nature of the patterns of acoustic energy
and reduced the analysis to a simple pattern-89 matching exercise. Nevertheless, that exercise was initially heralded with outsized claims 90 of success. The article that introduced the voiceprinting method in 1962 reported that 91 phonetically naïve examiners3
were able to identify a target voice with 99% accuracy, 92 even from a pool of a dozen speakers. Not surprisingly, this new methodology soon caught 93 the attention of law enforcement, and was presented as evidence in a number of criminal 94 prosecutions, in the US and elsewhere. 95 96 5.3.2 However, the scientific community remained skeptical. Well-known phoneticians 97 such as Peter Ladefoged and Harry Hollien reported that mere pattern matching (which is 98 all that the young voiceprint examiners were asked to do) was incapable of yielding the 99 astonishing results reported in the 1962 article.
Due to the variability present in speech 100 productions from sample to sample, spectrographic template matching is not effective and 101 it is inconsistent in speaker recognition work.
In time, phoneticians began to provide 102 expert testimony against the admissibility of voiceprint evidence, and in consequence, a 103 number of lower-court convictions were eventually overturned. In response to these 104 criticisms, another academic linguist, Oscar Tosi, initiated a more rigorous, and 105 procedurally transparent, voiceprint study (Tosi et al. 1972). This yielded less vertiginous, 106 but more scientifically reliable results – 6% false identification errors and 13% false 107 elimination errors, under laboratory conditions. Still, given the high stakes of introducing 108 a still largely unsupported procedure into courts of law, a report issued by the National 109 Research Council4 concluded that the voiceprint method lacked an adequate scientific 110 basis for estimating reliability in many practical situations, pointing out in addition that 111 laboratory evaluations of the voiceprint method showed increasing errors as the conditions 112 for evaluation moved toward real-life situations, such as poor signal-to-noise ratios and 113 dissimilar recording conditions. 114 115 5.3.3 The Federal Rules of Evidence, adopted in 1975, further challenged the voiceprint 116 methodology by shifting the standards for admissibility in favor of practitioners whose 117 "scientific, technical, and other specialized knowledge" can help the trier of fact "to 118 understand the evidence or to determine a fact in issue"5 Phonetically untrained voiceprint 119 examiners, who sought to identify speakers simply by looking at pictures of their voice 120 signals, were left at a marked disadvantage.