Voice Authentication for Classrooms? 28
USSJoin asks: "I am teaching a summer camp for gifted/talented children this summer, and one of my courses is an introduction to forensic science. One idea I had was to demonstrate voice printing and voice authentication. Using the magic Google, I was able to find software to get a visual representation of a voice print, but I didn't find anything that would allow me to demonstrate voice authentication. Ideally, I would like to be able to have students record their voices onto a cassette player, then speak into the computer, then try to fake out the computer using the tape recording. Does Slashdot have any ideas on how to demo this to brilliant young kids?"
Easy (Score:1, Funny)
Use it to protect the computer containing their final exam. They'll understand the technology backwards and forwards and have it broken by lunchtime.
Mac OS 9 login (Score:4, Informative)
I remember it being fairly good for a while, but having to re-record my passphrase as my pronunciation changed over a couple months or so. Nonetheless, it was popular with myself and my family simple because it was so freakin' cool to login via your voice.
An idea... but you'll need a midget (Score:1, Funny)
Perhaps the Whitehouse could help (Score:1)
Re:Perhaps the Whitehouse could help (Score:1)
Maybe this: (Score:1)
Re:Maybe this: (Score:3, Funny)
We'll just have to wait for Blu-Ray so we can just fit the voice-print of every person on the planet on the companion CD. Great idea though. Imagine a beowulf cluster of those..
Sigh..
First Cirlce (Score:1)
Re:First Cirlce (Score:1)
Re:First Cirlce (Score:1)
The russian authorities ask for help in matching a voice to a man, using the new technique being developed by Rubin - Voice prints.
Dear Slash Dot... (Score:1, Troll)
Maybe you shouldn't teach what you don't know? How about that?
Re:Dear Slash Dot... (Score:4, Funny)
Re:Dear Slash Dot... (Score:2)
Tell them to go to college... (Score:2)
I'm sorry, but I think it's a little bit out of the "Summer camp" league.
But please do post your findings, there have to be some "beginner" resources out there for people interested in this tech.
mm, Audacity and diff? (Score:2)
though any kind of static or background noise will automatically show. The idea still holds though, you're looking for a significant noise signature between the voice and the tape player. Now supposing your tape recorders are really good, the computer might get confused.
Oh, and make sure you've got
Re:mm, Audacity and diff? (Score:1)
Displaying it on an o-scope is slightly better, but for one it will fall victim to the same short-coming desc
Re:mm, Audacity and diff? (Score:2)
Re:mm, Audacity and diff? (Score:1)
Some more info on dynami [shef.ac.uk]
Diff? (Score:2)
To do what you suggest, you'd need to graph the 2 recordings and then use some sort of visual comparison program to determine "h
Re:mm, Audacity and diff? (Score:2)
Thats the problem with all the Linux freak fanatics. Every problem is simple to solve on a white box with the command line and the stuff you get from the gnu textutils. It must be so because its free, right??!
watch out for the quackery here (Score:5, Informative)
I think that you need to be very careful what you tell those kids. Most of what you see on TV about voice identification is nonsense. The images that they call "voiceprints" are spectrograms: that is, they're 3D plots of the spectrum over time, with frequency on the y axis, time on the x axis, and energy represented by darkness. Phoneticians like myself use them all the time.
In one sense every utterance, and therefore every spectrogram, is unique. The central problem in acoustic phonetics is the enormous variation in the physical signal for what in linguistic terms is the "same" utterance. The details of the signal depend on the speaker, the speaker's mood and state of health, the weather, rate of speech, choice of register (formal, casual, etc.), as well as on what other sounds the speaker is producing in the vicinity. There is a lot of contextual influence. If you compare, for example, the vowel /u/ in "tune" with that in "moose", you'll find a large difference. This one is so large you can see it just looking at the spectrogram.
Once spectrograms became available, in the late 1940s (using a machine called the sonagraph with analog filters), people started looking for the acoustic correlates of linguistic features. They thought that it would be simple. What they discovered was the tremendous amount of variation and the great difficulty of finding acoustic correlates of linguistic features that are invariant under changes in phonetic context and the various other factors I mentioned.
One result of this is that almost all of the research has been on abstracting away sources of variation such as speaker identity. As a result, not very much is known about the properties of the voice that are unique to individual speakers. In fact, we do not know whether voices are unique. It's clear, of course, that to some extent we can distinguish people by their voices, but we don't know that voices are truly unique, or how close they are to it.
The upshot of this is that there is no scientific basis for determining whether two recordings, or two "voiceprints", are of the same speaker. (If they're different enough we may be able to say that they are NOT from the same speaker.) Anybody who claims to be able to look at a couple of spectrograms and testify with confidence that the same person produced both utterances is a quack. I know people who've spent substantial time debunking this stuff in court. You won't find it supported by published research.
So, why can you login to your computer by voice? Systems like that rely on statistical "ignorance modelling". We don't know very much about what the relevant acoustic properties are, but we can make statistical models that are good enough at distinguishing one speaker from another for some applications. Even the better speaker identification systems don't work too well if they can't make a comparison between two instances of the same utterance, and as another poster mentioned from his own experience, changes in his own voice over a few months would throw off his voice login system.
The other relevant factor here is that for some purposes its okay to have systems that make a lot of mistakes as long as they are in the right direction. If you want to limit access to a lab, let's say, it will very likely be okay to have a system that produces a lot of false negatives, that is, that incorrectly denies that the person trying to enter is authorized to. So long as you have a very low rate of false positives, the system may be acceptable.
So, the real situation is that for some applications statistical voice recognition works well enough, but that such systems do not work well enough to be acceptable for such purposes as identifying a unique individual as a criminal. Speaker identification by visual comparison of spectrograms is junk science.
As for software for looking at speech, there are a number of free (as in beer and as in speech) programs available. This page [upenn.edu] has some links that you might find useful.
do fingerprints instead (Score:2)
I did a project on this... (Score:1)
Most modern voice identification systems use linear predictive coded (LPC) ceptra and either hidden Markov models (HMM) to evaluate how close a a given speaker is to a known user.
Having said that, I don't think it makes a very cool demo as the result is simply a number. In the case of speaker verification this number represents the proba