Cell Phones for the Deaf 270
nitzan writes "Quoting from the article: 'the software translates the voice on the other side of the line into a three dimensional animated face on the computer, whose lips move in real time synch with the voice allowing the receiver to lip read.' Unfortunately this only works with laptops, but a pda version is in the works." The company website has a demonstration.
Can you hear me now? (Score:5, Funny)
Ok, can you hear me now? Still no?
Ok....
Re:Can you hear me now? (Score:2)
Re:Can you hear me now? (Score:2)
You only assume that's what he meant because of your own frame of reference.
Whee! Nitpicking with high karma is fun isn't it
Oh, great... (Score:5, Funny)
Re:Oh, great... (Score:5, Insightful)
Yes, because the deaf person is bound to have the ringer turned way up...
Oy.
Re:Oh, great... (Score:5, Funny)
I believe this particular sign has already been standardized and is currently in use by 99% of the American driving population.
.!..
!!.. if you're from the other side of the pond.
Technology overkill (Score:5, Insightful)
Re:Technology overkill (Score:4, Informative)
Re:Technology overkill (Score:5, Interesting)
Text to speech:
1. person speaks
2. software interprets phonetics converts it into words
3. deaf person reads the words
versus
1. person speaks
2. software interprets phonetics into picture based lip movements
3. deaf person interprets picture based lip movements
Point of fact this is unbelievably dumb and is right up there with converting Russian to German for an English speaker to read.
Re:Technology overkill (Score:2, Insightful)
By doing the pictures, you're essentially leaving the last part (converting the phonemes into words and sentences) to the deaf person reading the lips, instead of a computer.
In order for the computer to do a more reliable job on this last part, it either has to take a long time or the processor has to be really fast. And even with that, the computer is still going to make a lot of mistakes.
This is certainly nowhere near as brain-dead as you make it out to be.
Re:Technology overkill (Score:5, Insightful)
"2. software interprets phonetics converts it into words"
Is a very different, much more complex problem than:
"2. software interprets phonetics into picture based lip movements"
Consider that for the first example, we need the computer to understand the language,
whereas in the second example, all the computer needs is a fourier transform and
Max Headroom anatomy.
Personally, I think it would be simpler and more effective to put a
camera on the phone and transmit an image of the speakers face.
Re:Technology overkill (Score:2)
Re:Technology overkill (Score:5, Interesting)
Very well put!
I've had deaf friends, one of whom attended Gallaudet University. (Famous liberal arts college for the deaf.) In addition, I lost most of my hearing for some years as a child -- fortunately, I got it back after surgery. I've thought about deafness, and dealt with it.
Lip-reading works best for people who were hearing at one point and lost some or all of their hearing. I went deaf after I learned to talk, and went deaf slowly, which means I relied heavily upon it. People who have always been deaf often find lip-reading very difficult, or even impossible. When you have no concept of hearing or sound, trying to figure out what meaning is associated with specific lip movements is tough.
This is true of learning to read, as well. A person who was already speaking, or could read, before going deaf has no real problem with reading. If you can't hear and never have heard, though, the concept of an alphabet and "sounding it out" makes no sense. A congenitally deaf person who wants to learn to read must learn each word as a whole, much as a Chinese or Japanese person who learns to read his/her language must learn each character separately.
Since a congenitally deaf person faces a humongous task regardless of whether he/she is learning to read lips, or read and write, just which one do you think he/she would rather have to learn? In most cases, learning to read and write is going to be a lot more useful.
From where I sit, speech to text would work better for most deaf people, congenitally deaf or not.
Re:Technology overkill (Score:2)
It's interesting that not only do the model's lips move, but there are visual cues in the cheeks, nose, and throat.
Re:Technology overkill (Score:2)
Finally a solution for illiterate deaf people! (Score:5, Funny)
combine lip-reading with speech2text (Score:3, Interesting)
Re:Technology overkill (Score:2, Interesting)
On a more serious note not everyone reads lips in english, if you develop this right its common for any language
I'll tell you. (Score:2)
People seem to be forgetting that speech to text is the back-end for this lip service anyway. In order for it to work, speach is interpreted by a computer which then maps the interpreted speech to canned lip movements. The canned lip movements require cpu horsepower to drive the graphics and they need a large screen for it to be readable. These two reasons are why it is only available on a laptop.
With the speech to text scenario; speech is interpreted by a computer and is matched to canned pieces of text. So far, pretty much the same. But, now the text is output to just about any screen, including the text screens of today's cell phones.
Basically the speech to text would be an automated TTY/TDD system. TTY/TDD has been in use and has proven highly effective for decades.
To answer your question, there is NOTHING wrong with speech to text. However, you won't draw too many VCs with it. Now, put a computerized talking head on it and extoll the greatness of its virtues and you may well be able to sucker in a few VCs. And afterall, isn't that what it's all about?
Re:I'll tell you. (Score:2)
Of course it is. It's also not current technology.
The phone is more of an oscilloscope than a speech translator.
If you want an apples-to-apples comparison, consider the alternatives
between an animated face or a spectrum analyser, not between an
animated face or a text display.
However ... (Score:2)
I know I'm generalizing, and not to be politically uncorrect, but don't most deaf people have difficulty speaking "clearly"? So how does the phone deal with that? Or does the deaf user need to type in their response?
It seems to me that speech-to-text for receiving and text-to-speech for sending is the way to go
In which case you've just re-invented the Blackberry.
Re:Technology overkill (Score:2)
To do speech-to-text well, you have to know the language being spoken, so that you can pick out the words and try to spell them right (since phoneme-to-letter mappings are not well-defined or predictable for most languages). On top of this, you have to somehow deal with slurrings (people on the phone are not necessarily the best enunciators in the world), slang, names, etc. etc. Then you have to do this for every language that you want to support.
Text-to-graphics, on the other hand, is comparitively simple. Humans the world over have a relatively small number of sounds that they use (probably on the order of 300, if you don't count tonal variations, and you're trying to count every distinction that's made in some language) and the mapping from these onto facial shapes is fairly well-understood. There is (in theory) no tweaking needed to make it understand other languages, so when your deaf Chinese friend borrows it to call home it'll work without trouble. Tonal languages could be an issue... but really, deaf people are going to have trouble with that anyway, and this could even help there, since extra cues (a raised or lowered chin, say) could be used to indicate tone.
It's not so much that there's anything wrong with speech-to-text as that this has the potential to be more right, esp. if used in combination w/ s2t. The fact that no word DB is needed makes it much more likely that s2g'll appear on a PDA that s2t, at least in the near future.
Re:Technology overkill (Score:2)
> What was wrong with speech to text?
Speech to text is a much harder problem!
You can map the articulation to an animated face without needing
to know the language at all. Now, whether you can do it well enough
to help a deaf person understand, is an open question.
All you'd need to do to represent someone saying "Oh" versus "Ooooh" is map the phoneme to a shape.
I'd imagine Fourier analysis would be oen of the more useful tools here.
Re:Technology overkill (Score:2)
Only reason I can see for this.. (Score:2)
How well do you actualy see? The software would require the speach to be converted into some form of phonetics before it could determine mouth position. And if it can be converted to phoenetics it's just as easy to convert to text.
The only possible reason I can see for this is if the software isn't really that accurate and so ti can get away with fudging the phonetics just a bit and/or the software was designed not so much for the pure deaf, but for those with some audio preception and the visuals of the mouth help them to better understand the audio.
Re:Technology overkill (Score:3, Insightful)
Kintanon
Re:Technology overkill (Score:3, Interesting)
If you have to say the sounds for the word "ow", what does that look like? There is a way for a computer to display this that would be pretty clear, and figuring this out would more or less require grabbing the "ow" picture group.
Now, what if you have to write something with a "ow" sound in it, but this "ow" sound might be in the middle, beginning or end of any word? The sounds all around it have an effect on it. It might be spelled "au", "ow", "ao", "ough", god-knows-what-else. There are dozens and dozens of situations where this sound might arise. Including ways you might not even consciously think about. Figuring all this out is really hard for a computer to do, because it has no AI. It doesn't know what is being talked about.
Probably mapping mouth movements isn't dead-easy, but I'd wager it is much easier than speech-to-text.
Re:Technology overkill (Score:2)
So I still have to put my vote in on the side of speech to text. There are too many subtleties to lipreading for me to believe a CELLPHONE will be able to display an image well enough for it using less resources than than speech to text takes. Plus, speech to text could theoretically be taken care of at some point between the two phones as far as the processing goes and then sent in final form to the recipient. The face HAS to be done on the cellphone itself...
Kintanon
Re:Technology overkill (Score:3, Interesting)
For my interpretation of how the lipreading worked, I was looking at the sample on the website. It appeared to be that certain sounds had certain pictures. Anyway, it also had something that touched on what you were talking about. It had a little bit of extra information you can't normally see, like red dots on the cheeks for g and k type sounds, and a red nose for nasal sounds, etc. Extras like these might make up for some of the lacking facial cues.
The other thing I was thinking about was that the lipreading could be part of the understanding process. A good number of people (most?) who are legally deaf are not truly 100% deaf. If a person is able to get a bit of auditory information, plus this lipsynching information, it might be enough to make things a lot easier for these people, even if the lip-reading by itself were too simplistic.
What about TTY? (Score:5, Interesting)
Re:What about TTY? (Score:4, Funny)
Re:What about TTY? (Score:2, Informative)
The new phones are TTY compatible, they do not have a TTY in them, but if you hook a TTY to them it actually works, whereas with the other digital phones that aren't TTY compatible (right now the majority) you get alot of garbage characters.
Analog cell phones are unaffected and with a TTY just fine without modification
WHy should they. (Score:4, Insightful)
OK, this might be a troll. Im not sure myself. Its definately a vent. Fucking sprint. Oh well.
Re:WHy should they. (Score:2)
Yes, mobile phone service sucks because they've pulled all their best engineers out of the field and made them write bowling games instead.
Negatory, good buddy. Building fancier handsets is trivial compared to actually improving coverage, and it's not like the two are mutually exclusive. Let's say the phone companies overcame all the various technical problems like interference, desensing, varying topography, etc. That still leaves political problems like unrealistic people in upscale neighborhoods who magically want 100% coverage without any visible antennas and dumb people who apparently think that cell phone towers will steal their souls.
It would also help if the phone companies would stop taking on massive numbers of new customers when they can't support the ones they have, but I guess the money has to come from somewhere.
Re:What about TTY? (Score:2)
But yeah, with just a minute effort... It's slightly surprising that there hasn't been anything mandated by law on this front.
Re:What about TTY? (Score:3, Interesting)
Case and point: I recently got a cellphone, so that someone else could have their phone back. I shopped around for a while and settled for one that would be free after rebate. I had it for a few days, and returned it for a more expensive one. First, the phone had an odd number layout, so I had to relearn the key mapping (the keys were part of a curve, instead of straight across). Second, I use a laptop to connect to the Internet, and I occasionally use a cell phone adapter to do it. The phone I bought was incompatible with the connector, and the phone's manufacturer had no immediate plans to make one. Those two reasons, as well as several other factors, prompted a return.
If the cell phone companies would agree on a single interface, it would make the compatibility much easier to implement. Not only that, but the TTY devices need the information to implement all the various brands and models of cell phones. The possibility's there, but there's not much of a chance it'll happen anytime soon.
Re:What about TTY? (Score:2)
Why didn't you just shop around a bit longer and make sure it would suit your needs?
One flaw ... (Score:3, Funny)
Faces (Score:2)
Or you could mix n match your own.
Yikes! (Score:2)
.
frightenly useful (Score:2)
"Sure looked good on paper...."
Crikey, they're already bad enough drivers (Score:2)
Re:Crikey, they're already bad enough drivers (Score:2)
Blind people driving though... That would be scary.
Kintanon
Re:Crikey, they're already bad enough drivers (Score:2)
There actually used to be laws against it. IIRC, correctly, the only thing needed in NYS is to have a larger rear-view mirror now, and I'm not even certain if that is a requirement.
Ugly (Score:2)
I hope to God they are not using Flash to deliver this product.
Uhhgg!
Complicated (Score:4, Insightful)
Re:Complicated (Score:2)
Converting phonemes into sentences requires context. Right now, speech recognition software "simulates" context by having large word banks and using probability. It tries to guess from sample text what the most likely string of words was. This is really not easy. This is why you have to train with speech recognition software. It tries to build up a database of likely things you'll say. It's like Tivo, sort of. And it can get a sentence totally wrong sometimes (often?). If you have a slow processor, going through all this data can take a REALLY long time.
Speech-to-text might be out of the scope of today's PDA to do, whereas the lip-synching stuff wouldn't be.
Good idea and good start but.... (Score:5, Informative)
Re:Good idea and good start but.... (Score:2)
Re:Good idea and good start but.... (Score:2)
"Ee Peach View. EIEIO."
Uhhh... (Score:5, Funny)
Re:Office space... (Score:4, Funny)
If only Helen Keller were still around... (Score:2)
Oh wait, she'd still need to see the cell phone. Never mind. I guess its a good thing she isn't still here.
Wouldn't it be simpler to translate to text? (Score:2)
speech->text->animated model
Wouldn't it be simpler to present the text to the user? I would have thought text->human is much higher bandwidth than animated-model->human.
Re:Wouldn't it be simpler to translate to text? (Score:2)
Speech->soundex (or equivalent)->animation.
Re:Wouldn't it be simpler to translate to text? (Score:2)
Speaking from experience (Score:5, Insightful)
Yeah, sure (Score:3, Interesting)
I sure as hell couldn't tell you what they were saying, even when I knew what words were coming out of their mouth. And this is not to mention cell phone static, distractions, contractions, mumbling, and lots of "ummm" and "uhhhh" that occurs during normal speech. I really don't see how this is a viable communication method.
Maybe it's because I'm not experienced with lip reading. Maybe people who are deaf are better at it than I am, but I can usually tell what Football coaches are saying on the sidelines of games (of course, that's limited to "Bull****" and "You've gotta be ****ing kidding me!", but still...)
Is it just me? (Score:2)
Re:Is it just me? (Score:2)
Ugh (Score:3, Insightful)
lip reading.. (Score:3, Funny)
'I want a fig newton'
IMHO:
too many flaws, the investors will back out
Re:lip reading.. (Score:2)
Dont even need to say that to a lip reader. Try mouthing it to the hot blonde across the room sometime. Everyone can lip-read that phrase.
Re:lip reading.. (Score:3, Insightful)
And unless you just randomly blurted out the sentence, you probably have context in the surrounding sentences (e.g. you are talking about fig newtons, food in general, newton's law, whatever).
I'd definitely put my money on the lip-reader, frankly.
Why Lip Reading instead of Signing? (Score:2)
In a related story... (Score:2)
This doesn't really do much... (Score:5, Informative)
Someone can say "Pot" and yet with the same lip movement, can also say "My". Men with bushy mustaches are a lip-reading disaster.
For me, I've adapted in my own way: I rely heavily on my hearing aids. That combination of both lip-reading and hearing the audio stream from your mouth enables me to achieve at least a 70% success rate (under ideal conditions, if it's a party atomosphere, fudgeddaboutit). I've had hearing aids since I was 1 1/2, and only with extensive speech therapy can I speak well. I'm one of the few deaf-from-birth people that can do it this well. So, from that perspective, I can speak on a phone (as long as I can understand that mangled audio coming out the receiver, which is 0%).
Why don't they just focus on speech recognition? A great speech recognition phone would enable deaf people that speak to use phones for near real-time conversations. In addition, such technology can also be (easily?) adapted to foreign language translators for tourists.
However, until such technology is available at the consumer level, I'm stuck with two-way text messaging devices like the T-Mobile SideKick.
-Cyc
Re:This doesn't really do much... (Score:2)
Actually, this is not so easily done.
You come across a number of ways to introduce errors. First, recognizing the speech and figuring out the phonemes, with some margin of error. Then, you have to convert these phonemes to text, which requires a lot of computing power to do in real time, and you introduce a lot more errors. Then you have to translate this text. We all know how babelfish is. So you'd end up with a very garbled (probably unintelligible) message.
I'm not saying this is impossible, just that this is no trivial task.
In the meantime, I think this technology we see in the article is a step in the right direction.
One other thing (Score:2)
With the accuracy of speech-to-text these days, the margin of error you get reading those lips might very well be smaller than if a computer tries to make those sounds into words and sentences.
Read My Lips (Score:5, Interesting)
I notice a lot of people complaining about improving text-to-speech, which is far more advanced than this technology. Speech sounds come out in a continuous flow. Getting a computer to recognize the breaks between words, properly spell them reliably, etc. is hard enough on a desktop system, much less a PDA. Especially considering in languages like English, where most vowels in unstressed syllables are rendered vocally as "uh".
This system simply has to hear a sound, and immediately display an associated... well, not "grapheme", since this isn't writing... maybe "pixeme". It is the graphical equivalent of attempting to spell perfectly phonetically.
Also, if you didn't notice it, "invisible" sounds that occur on the back of the tongue are indicated by circles on the cheeks (like hard 'g' and 'k'), and nasal sounds are indicated by a darkening of the nose.
All in all, I think this is an interesting idea. It will be even cooler when they can render different faces so the "avatar" resembles the person to whom you're speaking.
Re:Read My Lips (Score:2)
Hmm... I suppose in retrospect, what I was trying to get across and what I said are two different things.
What I was trying to get at is that lip reading is easier based on context. I know what the words are likely to be, so I can tell what's being said more easily. If I tell someone "Thank you", I expect a response from a fairly standard repertoire. Perhaps "you're welcome", "no problem", or something of the sort. If someone said "licking glass bonkers makes pencils freeze", I might be somewhat perplexed, but we usually don't have to deal with non sequiters of that nature. Lip reading the animated face with appropriate contextual cues would seem to be fairly easy to learn for someone who could already lip read.
The main problem is that this thing that the vast majority of the deaf community can not lip read.
I won't doubt it, but I tried to avoid singling out this technology as something solely for the deaf. A similar package might be used to add more life-like interfaces to computer systems someday. Here's an example:
AI systems can rival (if not surpass) human ability to do things like recognize faces, determine a person's gender, etc. A system specially designed to listen to speech phonemes and then represent them graphically might make speech-based interfaces easier to understand in noisy environments, because we'd be able to see what was being said as well as hear it. Being able to watch a mouth form the phonemes makes it immesurably easier to understand words that might otherwise be misheard.
The reasons this is better than speech-to-text (Score:5, Informative)
Speech to text isn't very good because its very hard to turn phonetics into words. Our ability to understand people is very reliant on context. Knowing what's been said helps you understand what's being said.
Some will say that speech to text is getting fairly good in English, which is somewhat true. Obviously, though, there are bigger markets in other languages.
So how does this thing work, if it doesn't do speech to text? It does speech to phonetics, and phonetics to lips.
For example, its relatively easy to understand when someone has said "h -ee- r", but knowing if that's supposed to be "here" or "hear" is quite difficult.
This is why the same software works across languages. "Th" is "Th" in any language, and your single algorithm doesn't have to care.
-Zipwow
Re:The reasons this is better than speech-to-text (Score:2)
I wonder why it wouldn't be appropriate to deliver the phonetics themselves?
In my university language classes, I'm expected to read French from a standard
phonetic alphabet. If the device in the article is truly mapping phonemes from
speech, then a representation of those phonemes would be very useful;
never mind representing them into a given language.
Re:The reasons this is better than speech-to-text (Score:2)
I'd guess that a text-based reprsentation of phonemes isn't as intuitive as one would like. At the very least, its another skill one would have to learn to use the phone, whereas the user presumably already can do some amount of lip reading.
In general, though, I've often wondered if sending the phonemes over 'the wire' and reconstructing them as sound on the other side wouldn't work well.
Granted, the phonemes would probably have to have a lot more detail than they do now to sound right, but it seems one could send a lot of that kind of detail and still take up less space than even a compressed sound wave.
One application of this approach would be chat programs, and game communication. Because the phonemes are being reconstructed, you could reconstruct them with different characteristics than they were recorded, making them more feminine, or with a scottish brogue, etc.
-Zipwow
Uuuuuhhhh... (Score:2)
The sprint guy has it bad enough... (Score:4, Funny)
Wife: "I told my husband 'get a gallon of milk from the store'."
Husband: "No, what I saw was 'kill my family and urinate in their lifeless armpits'."
SprintGuy: "Sir, it's the SpeechView. Here, Sprint built the..."
Nuff said (and signed)! (Score:2)
I thought we had this feature a long time ago! (Score:2)
I can see one advantage to this... (Score:5, Interesting)
I've spoken with a hearing impared person on a phone before through a TTY system and it is painfully slow. First you have to say your sentence and then they send it. Then the other end needs to read it, type in a response, and then send it at which point it is read back to you. Imagine having a conversation over an Instant Messenger except you're secretary was reading the screen and typing for you. (IM for the blind for example)
I agree that we need better voice to text and text to voice translation. That technology would give use better access for everyone. You could have "hearing" for the hearing impared (speech to text), "reading" for the vision impaired (text to speech), and you could even have "writing" for those with fine muscle control imparement or who are lacking the necessary limbs for various reasons.
But this is an interesting approach to solve one of the three problems.
no no no no no! (Score:2)
Can you imagine? I walk to work every day with my phone logged into AIM. I chat with people while I walk. I try not to step in potholes. The convenience of chatting and holding the cellphone at my side while waiting for the vibrating alert set me to thinking...
Iduno. Y'all want a portable SSH client that you don't have to look at in order to use? Without the requirement for a screen, I don't care how big the device is. It goes in my backpack. The input/output is all tactile.
I wonder how hard it is for sighted folks to learn braille. I wonder how hard it would be to mount braille-like output on a small handheld device. Dunno if that's possible, really.
I thought they had these already (Score:2, Funny)
At least I assumed that the folks speaking at 95 db into a highly compressed mic did so because they were deaf and unable to hear themselves.
I did this for karaoke (Score:2)
I decided to do La La land by green velvet one week and just for kicks I thought I would make a talking head ala max headroom.
http://www.zeromag.com/images/downloads/videos/
(Divx compressed BTW 6 megs)
Basically I just recorded a second track of my singing without the music, then pumped the wav through the facial animator in truespace 6.
What I found was it actually made it a bit easier for me to keep up with the words because I would watch how the lips on my on screen persona and mimic them myself.
Anyways, enjoy folks.
Lightbulbs for the Blind (Score:5, Funny)
What the hell ... (Score:2)
How about they just use a video phone? Or have the audio be displayed in a text output? It has to go through that step anyway.
3D modeled face
Ummm... (Score:2)
Phonemes, visemes, TTS, and lip synching (Score:3, Informative)
By way of introduction: I developed the core coarticulation and other algorithms for lip synching when I worked at a now-defunct company called...wait for it...LIPSinc. We thought the resulting lip synching was pretty damn convincing, so on my own I tested out our stuff with a hearing-impaired friend, with mixed results. Anyway, I don't know a little about this stuff, I know a *lot* about it.
What these guys have done is map phonemes onto exaggerated visemes (the pictures of the mouth). Not a bad idea at all! Bunch of problems, though. First, there's a data data reduction of about 3x in going from sound to video--there are 40-50 distinguishable phonemes, and 9-16 distinguishable visemes, depending on how you count each. This is because the visible part of the face only makes up the end of the vocal tract, a lot of distinctions between letters occurs without the involvement of the lips, like the difference between F and V, while others, like K, can be pronounced with the face in virtually any position. This is part of what makes lip reading so hard with a real person, and why they need a lot of context to pull it off. They also seem to be slowing down the timing, as if they recognized the phonemes and then synthesized each at the same length. This gives longer to recognize each one, but wrecks the visual prosody (rhthym) of the speech, which is a good cue for where the parts of speech are. Then there's the rest of the face. The eyebrows and head positions help you figure out key words, ends of clauses, tell if something is a question, etc.
Those who say that TTS is superior to lip reading have a point. Good TTS contains *more* accurate information than an uninterpreted stream of phonemes (itself 3x richer than a stream of visemes, as I said above), because the machine can do a Viterbi search to find the most likely sequence of words from a continuous stream of phonemes. Words also open up higher NLP functions, so you can do constraint relaxation to test whether "wreck a nice beach" or "recognize speech" fits better in the context.
Still, I'd like to see an experiment where the raw phonemes are fed, as text, to the recipient. I think with practice, your brain would start to decode the string (it manages with the sound, right?), despite the lack of word boundaries and the errors in phoneme detection (which is not all that high without text-I think seventy-something percent). Seems like an easier pattern recognition problem than lip reading. Who wants to go get funding?
Re:Phonemes, visemes, TTS, and lip synching (Score:2)
There are visual cues for "invisible" sounds, like the nose darkens, the throat turns blue and the dots appear on the cheeks for various sounds. Watch the demo.
It'll take practice to get used to, I'm sure, but no more than it takes to pick up Graffiti on PDAs, I imagine.
Wonderful! (Score:2)
Re: (Score:2)
Re:Wonderful! (Score:2)
Oops, you're absolutely right.
I was thinking of the *BRAILLE* cell phone.
Neat (Score:2)
Ive seen lots of suggestions for speech to text, but if you have had any experence with regular powerful PCs and speech->text you will see why that wont work on even a 2ghz intel system, let alone a pda/cel phone.
(Didnt we just have an ask slashdot about this?)
A wire frame of a face only requires slightly more CPU power than processing a WinAmp visual (No 32 bit color eyecandy here) and i have actually seen a visualization plugin (For ge-force, avail for winamp and itunes as well as some opensource packages) that has a module that draws a face and it in a way moves with the sound.
Granted that was not its design, it was made to look good, but obviously with a few changes it could be made to acuratly simulate a face and mouth for this very purpose.
Im all for any technology that makes interacting with a computer easier. While I personally would prefer direct sound into my ears over this, I also concider myself lucky to have that ability compared to those that dont.
Personally Im all for the direct brain connection, but i have a feeling thats a ways off yet
Experience talks (Score:2, Interesting)
It takes tremendous amounts of time, because not only does the translator have to interpret what the customer is saying, so that i can hear it, he then has to translate what i say back to the customer. It takes ages, and i'd imagine that with a cell phone, having a comptuer immediately translate, if slightly less accurate, would be preferable to having a human slowly (compared to the comptuer) enter it. Speed Vs Ease of Comprehension. Pretty common comparison. To each their own
DUH!!! (Score:2)
GSM phones. (Score:2)
SMS is very popular with the deaf (at least where I come from). It allows them to communicate just as easily with those with good hearing as other deaf people.
SMS also solves the problem of being globally accepted (just as long as you're not on the North-Western side of the Atlantic pond), and you don't need a special kind of GSM phone to be able to communicate with SMS. Another nice feature is that it works no matter how noisy it is surrounding the sender.
Better yet: (Score:3, Insightful)
If a deaf person wanted a "cell phone", they'll probably have one from Wynd Communications [wynd.com], a two-way pager with text/e-mail and other services built right into the damn thing. They're all the rage here [ntid.edu]. Screw lip reading over the phone. This technology is pure eye-candy. Nice, but how useful will it really be?
Re:Better yet: (Score:3, Insightful)
May you never go deaf then. May you never buy a product that breaks and have to call a phone number for customer services. May you never have to call an emergency services number. May you never have to call the pizza place to order a pizza.
The differnce between e-mail and TTY is the difference between push and pull technology. With e-mail, there's no guarantee that your e-mail is ever received, much less opened, read and processed.
Because of this, e-mail cannot (and does not!) qualify under the ADA soley as a reasonable accomodation.
I've worked as a secretary (in a school for the deaf, no less.) I know e-mails can take a long time to get delivered. There's still time between when it gets delivered and when it actually got read and processed by me. (Usually not long, but on some crazy hectic days, it could take some time.)
I hate feeding the trolls, but this one needed to be thwacked over the head.
a quote about the difficulty of lipreading (Score:2, Informative)
I'd also like to add that For Hearing People Only, ISBN 0-934016-1-0 is a great source of information about the complex and interesting world of Deaf people, and the language of ASL.
Re:And why isnt it just realtime text??? (Score:2)
Re:And why isnt it just realtime text??? (Score:3, Informative)
You have to record the speech and convert those sounds into phonemes. Now all you do is use the picture(s) that go with that phoneme, which is going to be more or less consistent.
With speech-to-text you have to use probability and word banks to figure out what the heck words those phonemes are supposed to go with, which is the hardest part by far, because spelling and grammar is so inconsistent. That requires a lot more time and computing power, and you are prone to a bunch more mistakes of course.
Re:and we thought drivers with cell phones were ba (Score:2)
AND I'm almost willing to bet you're one of those dumbasses who like to dial your cellphone with your phone out in front of you and not paying any attention to the road either..
Now whats the difference between this and that? None! dont use both on the road while driving!