Human-Computer Interfaces From 2003 to 2012 324
Roland Piquepaille writes "My favorite forecaster, Gartner, is back with a new series of predictions about the way we'll interact with our computing devices. Here is the introduction. 'Human-computer interfaces will rapidly improve during the next decade. The wide availability of cheaper display technologies will be one of the most transformational events in the IT industry.' Not exactly a scoop, isn't? But wait, here is a real prediction. 'Computer screens will become ubiquitous in the everyday environment.' Ready for another prediction? 'Through 2012, more than 95 percent (by volume in gigabytes) of human-to-computer information input will remain keyboard- and mouse-based.' Check this column for a summary."
HID!! (Score:2, Funny)
Is there a HID with a the robocop spike on the horizon?
Re:HID!! (Score:2, Informative)
Predictions? (Score:5, Funny)
Pretty wild ideas there, I hope he doesn't try to patent the keyboard and mouse or something.....
Re:Predictions? (Score:5, Funny)
Re:Predictions? (Score:3, Funny)
Synopsis: Method by which information from reading a book enters the mind.
Improves upon existing patents #1289,3,2432.23: Books (Amazon.com, 1999); #123,34342,24545b: Eyes (Amazon.com, 1998)
eh, I already wrote about this [dailysedative.com]. here, too [slashdot.org].
This just in! (Score:4, Funny)
It is estimated that this will not change by the year 2012.
Re:This just in! (Score:2, Funny)
Don't complain though, you get 5 free DVD-ROM ejections and insertions per month, after that it's only $0.99
I love microsoft! They sell me Freedom(tm)!
Re:This just in! (Score:3, Funny)
Oh darn. (Score:4, Funny)
I guess robot love dolls won't be on the market until 2013. (99.4 probability)
Re:Oh darn. (Score:4, Funny)
I guess robot love dolls won't be on the market until 2013.
Oh, they'll be on the market by 2006, but they'll be strictly Windows-based. It won't be until 2013 that they go open source. At that point they'll come into more common use as BSOD-related personal injuries become far less common.
Re:Oh darn. (Score:5, Funny)
Re:This is bad news for the ladies (Score:2)
that means the male dolls will run Unix
Re:This is bad news for the ladies (Score:2)
that means the male dolls will run Unix
Yes, but I'm guessing people purchasing male dolls will not be running Eunuchs.
Re:This is bad news for the ladies (Score:2)
Re:This is bad news for the ladies (Score:2)
Volume (Score:2, Funny)
And what about in MB or KB?
The same??
Ok, that was just to use a buzzword, I understand better now!
Re:Volume (Score:4, Funny)
Re:Volume (Score:2)
I'll change my interface device... (Score:5, Funny)
Re:I'll change my interface device... (Score:2)
Mod up above, I would love to see thinkgeek make that a bumper sticker.
Re:I'll change my interface device... (Score:5, Funny)
That won't be too hard: individuals with CTS and tendonitis can't grip things too tightly.
Of course, neither can dead people. Never mind.
Media Labs (Score:4, Informative)
Re:Media Labs (Score:4, Interesting)
Take the mouse, for example. According to this article [ideafinder.com], the mouse was invented in 1968. And it didn't become popular until the Mac came out in 1984. That's 16 years of obscurity before general adoption. Granted, there wasn't really any general widespread use of computer technology in that 16 years, so these days it'd be a good bit less. Still, people are really slow to switch away from something familiar that "works".
Re:Media Labs (Score:2)
Okay, there are innovations in user interface. But how likely is it that any of them will become widely adopted in the near future?
Light pens seemed like the Next Big Thing in I/O devices 20 years ago... how many practical applications do they have today?
Re push vs pull (Score:5, Interesting)
Some of the problems with push technology
Re:Re push vs pull (Score:5, Insightful)
It sounds off because it is. "Push" is one of those stillborn ideas that marketroids insist on resurrecting every few years, like the impending death of the PC, the ascendance of subscription-everything, thin clients, household automation, and so on.
Re:Re push vs pull (Score:3, Insightful)
Re:Re push vs pull (Score:2)
The ultimate push (Score:2)
You're forgetting, however, that the distribution of video over cable (with immense analog bandwidth) to a TV would qualify as "push" as well. TV is so immensely popular that the average american watches it almost as much as he sleeps! If you throw a TiVo or other set-top box, you've got push-content through a computer. The odd thing about this is that TiVo is designed to make a strictly push technology more pull like
Push's place is where technology isn't good enough to allow for pulls. This is the reason push of video content via analog or digital means will outpace pull for some time. After that time, I'll start watching TV again.
Re:The ultimate push (Score:2)
Tv-on-demand is also a "pull" technology - you specify what you want, and it's delivered to you.
Push is good for the vendor and lousy for the client - which is why we've seen so many attempts at push technology, and why clients still resist it.
Re:Re push vs pull (Score:3, Informative)
Unfortunately, the article's author (or an editor) got the two terms mixed up (I guess when push comes to shove ... ) This is what we've all come to expect from Gartner and the other cloudy-crystal-ball-gazers.
Re:Different Push here (Score:2)
That's not push in any commonly-used sense of the word. If that's push, then Google is a push service.
And "push" has already been tried in that sense, and people don't like it.
There was a thread here a few days ago about how PVRs were "anticipating people's choices", based on their previous viewing habits. The general consensus was that people don't like having their habits monitored, and other crap pushed on them.
Just like people don't like spyware, linkware, etc.
What most users want is freedom of choice, and to be secure. Their "intelligent" anticipation of what you watch/read/see/hear is too bib-brotherish.
I wonder if SMS text messaging will still be here (Score:4, Insightful)
Keyboard/Mouse sub-categories (Score:2)
I think 94 percent will be mouse generated (e.g., the new "Hello, WordProcessor" document would have several KB of different font styles, markup, colors, and all that jazz (all mouse based), and only a couple of dozen bytes of text (via keyboard).
S
Re:Keyboard/Mouse sub-categories (Score:3, Insightful)
My prediction: I see, you see, we all see ASCII. Yep, plain text will still be there.
Re:Keyboard/Mouse sub-categories (Score:2)
Re:Unicode (Score:2)
Re:Unicode (Score:2)
Unicode is backwards-compatible with ASCII, so the legacy/source code argument is irrelevant. There are already compilers available (such as Vector Pascal [pascal-central.com]) that interpret Greek, Cyrillic, Katakana and Hiragana. Heck, by 2012 I want to be able to code in Klingon [unicode.org]!
Wow, that's a first. (Score:2, Insightful)
Related: DARPA funds "cognitive assistant" (Score:4, Interesting)
Digital Paper (Score:3, Interesting)
Firstly, there is a certain tactile "feel" to writing on actual paper that would be very difficult to replicate - and if it feels too different, I suspect people won't adopt it.
Secondly, cost - could this be brought down to a price that would be economically feasible? If it's not as cheap as paper, it isn't gonna happen.
That's not to say that I wouldn't like to see it introduced; we could all have our workplace documents on those little pads, similar to theones in Star Trek, and I'm all for anything that will stop the slaughter of forests - I'm just highly pessmisitic. The author seems to be of a "more of the same" persuasion as well. Maybe someday, but I don't think we'll see it in the next ten years.
Re:Digital Paper (Score:4, Informative)
And, why do we have exactly duplicate the feel of paper? E-Ink is supposed to duplicate the flexibility and static display capabilities of paper, while adding digital versatility. The feel of writing on paper is learned, not instinctive.
Finally, why does it have to be as cheap as paper? It's much better than paper, it has many more uses, but it makes no sense to feed E-Ink into a laser printer or to hang it next your toilet. Digital ink keeps you from having to buy paper all the time.
How will non PC devices affect this? (Score:2, Insightful)
Gartner is useless (Score:4, Insightful)
Let's break it down:
Re:Gartner is useless (Score:2)
looks like someone spent a minute or two analyzing the dot com era of the late 90's eh?
Re:Gartner is useless (Score:5, Funny)
To come up with their predictions, analysts sit around and huff paint thinner until they lose consciousness. Once in a full state of dementia, fully developed predictions appear in rounded pod form from the brilliant, corpulent, snake-like ether of the true ultrafied space-time ribbons, at which point the analyst must delicately pluck them from the mind-hive before they can be sold to the public. Sometimes it comes out in both percents and gigabytes.
It's not a perfect system.
Re:Gartner is useless (Score:2)
The unit still matters.
For example, he wanted to be sure you knew he wasn't talking about information measured by "volume, in liters."
Re:Gartner is useless (Score:2)
--
Re:Gartner is useless (Score:5, Informative)
If you can't figure out from the article that these statements and numbers are part of a bigger document then I'll do it for you:
Mindless extrapolation of the obvious: "... will remain keyboard- and mouse-based."
Try the same sentence without the "keyboard- and mouse-based" part. It doesn't work.
Authoritative sounding numbers pulled out of the air: "... more than 95 percent
One of many phrases that are probably pulled out of a document where those numbers are explained. Blame ZDNet on leaving out the link to the original work by mr Gartner.
Sheer idiocy: "... 95 percent (by volume in gigabytes)
Same as above. There are numbers that go with these phrases. The numbers are in gigabytes (duh) and the blame lies with the reporter Alexander Linden for not refering to the original document. The dork prolly just cut and paste without looking at the content.
Now if someone can be so good to find us the complete works of mr Gartner.
Re:Gartner is useless (Score:2)
Re:Gartner is useless (Score:3, Insightful)
I'm pretty sure I agree but I'm trying to be charitable and come up with a reason why this makes sense... To be fair he could be saying it that way to make it clear exactly what is being measured. 95% by volume (in bits, bytes, etc.) is different from 95% by time. For instance it might be that in 2012 we spend 40% of our time computer input time using the sylus on our PDA and 20% of our time using voice recognition to talk to our home entertainment system but still 95% of the input by *volume of data* using keyboard and mouse.
Re:Gartner is useless (Score:3, Insightful)
All the 'gigabytes' and 'probability' numbers Gartner puts in their reports are there to give the reports a sense of legitimacy. They make their money off of people in suits at big corporations who spend big bucks on outside consulting. The suits love to have meetings with Powerpoint slides with lots of figures, and they get a lot of those figures from consultants. The public figures Gartner reports are just a summary of a more detailed report that corporations can purchase to fatten their presentations and corporate strategies.
The 10 year figures probably don't mean much since they are long forgotten by the time one could validate the prediction. Much like weather forecasts, the predictions shift over time as the real date approaches, and those predictions tend to get more accurate as the time to the prediction shrinks.
What I'd love to see is a port-mortem on the predictions from all these consultant companies like Gartner. I wonder if someone keeps a record of predictions that Gartner made ten or more years ago and compares them to what really happened. My suspicion is that most of their long term predictions are junk, but they produce the figures since corporations want to pay for those figures.
Interfaces? (Score:3, Funny)
.. but will these new interfaces work with my flying car?
no need to wait till 2008 (Score:3, Funny)
i've been on the mobile subway devices of NYC and D.C., and let me tell you... the reality there is extremely augmented. normally, i've found peak augmentation to occur around 4:20 in the afternoon for some reason.
CLI? (Score:2)
Displays (Score:5, Insightful)
More of one than you think.. I don't think he's talking about your monitor. In almost all consumer electronic devices, know what the most expensive component usually is? Yup, it's the display. Reduce the price of that, and all of a sudden, those consumer devices have a lot more to work with. More screens, better screens, enhanced power, cheaper price, etc... if we can reduce the cost of the display significantly, it can only mean good things for consumer electronics.
Re:Displays (Score:3, Insightful)
There are many current and old examples showing that good design can work with here-and-now technology quite well. Bad design will take that 39-cent display with SuperDuper-VGA resolution and turn it into a glossy usability nightmare.
Re:Displays (Score:2)
Agreed on the point that a better user experience should not be linked to a better display, but hopefully displays will steer clear of that path. What I'd like to see is screens with better visibility indoors and outdoors, cheaper (power-wise) backlighting, more accurate color reproduction, better viewing angles, better durability, etc. Unfortunately, you may be right about the "glossy SuperVGA nightmare".. anyone who has spent a fair bit of time with a GameBoy Advance knows what a nightmare that display is. Great color and resolution, but good luck trying to SEE the damned thing.
No, not just more, better cheaper. (Score:4, Insightful)
Ubiquitous.
Ubiquitous to the point that your very idea of "consumer electronic devices" is obsolete.
The existence of light emiting and electrically conductive liquid polymers that air cure is going to be completely transformative. Both display and electronic circuits are going to be printable on anything you can feed through your inkjet printer.
Think about that for a minute. *ANYTHING* you can get to feed through a printer ( or anything you can adapt a printer to print on) can be both a display and electronic circuit to provide driver and logic functions.
Think of all the things that are printed right now. Now think of all of them having "embeded" display and logic functions.
Like your paper placemat at the diner. And yes, they are even working on being able to provide *power*, self contained, in that paper placemat.
Your computer monitor will be pretty cool too. It could be nothing more than a sheet of 1/8" Lexan with the pixels printed on it. In fact, that same sheet of 1/8" Lexan could be your entire PDA or tablet PC if your data storage requirements aren't too great. Or on a sheet of polyethylene film you can fold up and put in your pocket.
All that will be pretty cool.
But it's the paper placemat thing that will be transformative. *Anything* can be a simple logic and display device. *Anything* can be a consumer electronic device.
Like Junkmail. Ready to get your free AOL *device*?
KFG
Re:Displays (Score:3, Funny)
The Windows license?
:)!
Alternate prediction (Score:4, Interesting)
By volume in gigabytes? Call me a contrarian, but I'll bet videocameras will exceed keyboard input by that standard. Wanna test that notion, Gartner? Point your text editor at a file, and I'll fire up my webcam recorder. Ready? GO!
Re:Alternate prediction (Score:2)
I'm guessing that such things don't fall into the category of "human-to-computer information input".
Re:Alternate prediction (Score:2)
Re:Alternate prediction (Score:2)
Yes, but many more people use keyboards and mouses for input that webcams, and more often.
1000 people touch-typing and/or dragging the mouse cursor across the screen will easily generate more bytes than 1 person with a webcam.
Gartner's prediction (and mine) is that this is not likely to change much very soon.
A Solution Looking For a Problem (Score:3, Insightful)
I think this is fine. All the effort to "perfect" speech and handwriting recognition, while worthwhile from an academic standpoint, is not really necessary.
I personally can type much faster than I could dictate, and most definitely faster than I can write by hand. I'm not even a "real" touch typist either.
Can anyone imagine dictating a long paper, or forget it, a complex program? You would go hourse and/or insane before 3 pages were done.
Therefore effort in speech recognition should focus on perfecting the simple command interfaces ("Computer, turn on the kitchen lights") instead of trying to perfect dictation. Speech recognition should be used to enhance keyboard based interfaces, not supplant them. Many times typing is the best way to input the data.
Re:A Solution Looking For a Problem (Score:2)
Now if they could develop one that responded to "Computer, where did I leave [ the remote | my glasses | my brains for wasting time reading another Gartner prediction ]", that would be something :-)
Re:A Solution Looking For a Problem (Score:2)
In a perfect speech recognition system, the dictation rates would be at the levels of today's micro casette recorders. Well over 100 words a minute.
More obvious cases are when a user is walking, driving, etc., when speech input becomes overwhelmingly advantageous. Even for devices with small form factors.
Imagine in the future, if I want to issue commands to my watch "fetch me headlines from slashdot and read level 5 comments from the first 6 stories"*.
(*) granted, if the watch interprets your request as "first *sex* stories", you'd have a differnet brand of entertainment :)
S
Re:A Solution Looking For a Problem (Score:3, Interesting)
Walking down the street *might* be an advantage, but I for one (and I know I'm not alone on this) would be wholly embarassed to walk down the street talking to my computer and would also be irritated by people who would do it (just like I'm irritated by people who walk around talking on their cellphones). Plus there are some safety issues, although smaller than those while driving.
Also, I think it would be really weird to have a computer talk back to me (not to mention a little inconvenient, how many times do you actually read entirely through a webpage? Not very often. Usually you just skim it.) and I would much rather interface with the computer using a semi-transparent glasses display.
Problem found... (Score:5, Funny)
"Computer, go to slashdot-dot-com.... no, slashdot-dot-com... no, slash!...
Oh, fine, just show me some porn."
-T
I was amazed... (Score:4, Interesting)
The question is...how long before this technology makes its way into mainstream computers, or something like it.
Wouldn't it be nice to just look at the monitor, blink twice and have the folder open. Careful where you look though!
Is this true? (Score:2, Insightful)
Is this even true today? I doubt it's true of my own work. I own a digital camera. I don't take many pictures; I'm not very photogenic. I figure I take about 50 pictures a month... let's call that one a day, to be conservative. 1600x1200x8, uncompressed (I use a raw format that sends 8-bit intensity data for each pixel, as each pixel in a digital camera is only one color), comes out to very close to 2MB per image. In a given day, I also spend about 8 hours sitting in front of my computer. I type at ~60 words per minute (never said I was fast), coming out to about 160kB/day. Now, I don't use my mouse too much, since it hurts my wrist. But even if it sends 4-byte updates 300 times a second when I'm not moving it at all, that comes out to 35MB a day... hardly a realistic number, but let's run with it. So my total keyboard/mouse input is 36MB a day, at an absurd maximum (I do stop for breath occasionally), while my non-keyboard/mouse input is 2MB/day, at a rather absurd minimum. And just with those numbers, I have (slightly) less than 95% of input being keyboard/mouse based.
I know a lot of people who take more pictures than me. One person taking 10 pictures a day is enough to offset 9 people who take none. A few people use speech recognition... that's relative high-bandwidth input. And I'm sure at least one person in ten thousand has a digital video camera...
So, does anyone think this 95% number is true even today?
--
Niels Bohr's take on all of this... (Score:4, Insightful)
~Niels Bohr
Unfortunately, I can't vouchsafe the quote. John Perry Barlow circulated it a few years back and when I asked him where he found it, he couldn't remember. So perhaps if Bohr didn't say it, he should have.
Not Niels Bohr (Score:3, Informative)
On the other hand this page [lundskov.dk] (in danish) says that it originated in a danish parliamentary debate of the period 1935-39. This is according to the memoirs of the politician K.K. Steincke. He doesn't remember who said it though. (Basic political instinct, I suppose.) It has also been attributed to Markus M. Ronner (whoever that is)
Niels Bohr was apparently the first dane to bring the expression abroad, and hence he has recieved credit.
My Predictions (Score:5, Funny)
backing of the predictionationization society, here
are my predictions for the next decade:
#1 Algebra won't be hard someday
#2 Grass will mow itself
#3 The Aliens people have encountered will be
revealed to be the "geek" or "dork" aliens. The
Jock aliens stay back on marklar and get laid and
drink. They are much bigger and stronger.
#4 Trendy computer users will start doing
"case piercing" and the truly EXTREME will try
out hard drive piercings. They will be made of
steel at first, but aluminum will become the rage.
#5 Wireless wires will be invented to replace the
wired wires.
#6 The "tornado in a can" will become "the can"
in your bathroom. Flushing dead goldfish will
never be boring again.
#7 Top ten lists will transmogrifimorphicate into
top 7 lists.
Paradigm shift (Score:3, Interesting)
I really couldn't care less about those modes of input. Can you imagine everyone in the office talking to their computers at once? And it wouldn't really help that much for programming or data entry, the tasks that a lot of computers get used for. As for handwriting, my hand starts to hurt after about five minutes of writing stuff on paper, and i usually give up and open up Notepad. And that's not even considering that my handwriting sucks and would be about ten times as difficult to process as "normal" handwriting.
What this guy really isn't saying much about is direct optical feeds; ie, beaming visual information onto your retina, or inserting false visual signals higher upsteam in your nervous system, and direct mental input; either in the form of reading the synapses in your brain, or recording your motions as you type and guesture in the air.
That's the kind of technology that will cause a major shift in the way we use computers, and is so different from our current modes of interaction that you can't really extrapolate from here to there. I'm sure scientists during the 40s and 50s were predicting great advances in vacuum tubes (the science fiction authors certainly were at least) that never materialized, or at least that were never utilized, because of the development of the microchip.
I have no idea if those kinds of technologies will be fully developed in the next ten years or not, but i don't think this guy has any better of an idea than the rest of us.
No more screens (Score:2, Interesting)
By 2012 computer displays as we now know them (LCD, CRT) will have been relegated to inexpensive embedded systems. Bleeding edge office information devices will function by tracking the user's movements and speech, as well as manipulation of common objects in her work environment. They will serve the same purpose as graphical icons do today. The computer screen will have been subsumed into dynamic surface markings and other detectable changes in the objects in her environment. They will have reflective (as opposed to backlit) display surfaces where information can be encoded in textual, graphical, color, or texture attributes, and sometimes some degree of 3D physical configuration changes. These will range from writing surfaces that resemble paper, cards, packaging materials, and other document-like entities, to instrument or appliance control panels and communications devices. User interactions with these items can produce changes in both the displays and the underlying data repositories. Moving them, rearranging their relative locations, adjusting them, speaking into them, and other as yet unforeseeable user interactions will effect the state changes that embody the user's day to day tasks. Think of a cube with an environment of intelligent interactive devices that visibly and audibly change as work gets done. The devices themselves will also be communicating and interacting as needed.
The "Probability" scam... (Score:2)
All prophets should use probabilities. That way, nobody can ever prove you wrong.
Say "The world will end on 1/11/2003 with probability 0.6" Suppose it doesn't end, so what? Someone's going to come back and say "Ha! It didn't end! The probability couldn't have ever been higher than 0.3!"
Suppose you say "Buy Acme Widget stock. It will go up 120% in the next 6 weeks, probability 0.8" People buy it. It goes up 120% in the next 6 weeks. They get rich. Are they going to come back and say, "Well, yeah, SURE, it did that, but you said the probability was 0.8 and it was really only 0.7"
Interface Idea (Score:5, Interesting)
Depends on what they call a keyboard (Score:2)
We may very well still be using keyboard 10 years from now, but they'd probably differ at least a bit from the ones we use now.
We're not going to get rid of the old QWERTY (or for some odd few, the DVORAK) until perhaps we can plug into ourselves, or until that one-handed keyboard comes around.
Personally... I'd like to plug myself in, but viruses would really suck then.
Every day environments... (Score:2)
Oh, wait...
Displays Improved, Interaction Probably Not (Score:4, Interesting)
Display technology has vastly improved. I'm now just waiting for the price to come down on IBM's T221 LCD so I can have one on my desktop. We purchased one at my workplace and it just blew me away. It is the first display I have ever seen that can be reasonably compared to quality laser printing on paper for its rendering of sharp, crisp, readable text. 9.2 million pixels in the thing and NOT ONE OF THEM IS DEAD. Yep, none, nada, zilch.
As far as interaction goes though, I doubt we're going to see much improvement. Programmers do a terrible job of UI design and a lot of companies are just too cheap or ignorant to hire professional user interface designers or else provide in-depth training for whoever is doing the UI design regarding usability issues. Most companies are also too cheap to do real usability testing. They might test out the new UI on the guy three cubicles away, but he's hardly representative of your customers. Until that changes, human-computer interaction is not going to improve.
Prediction: Integration (Score:4, Insightful)
Here's a real predition: Integration of devices will result in the replacment of single-use items such as PC's, TV's, cell phones, PDA's with portable and fixed units that have multiple functions. Consumers will buy "multi-media consoles" capable of several functions, that are more flexible and cheaper than indivdual components. Wireless networking will be the standard communication method between devices given the cost of adding wiring to a house, and the flexibility of putting your console anywhere. As a result, the lines between media types will blur, as 'television' as we think of it now will cease to exist with the advent of services that allow you to watch programming at a press of a button rather than on a schedule. You will read, listen to music, and shop, all from the same console. Integration will make the price of a large console about the same as a current mid-range PC, so consumers will buy several units in a family setting. Portable units will allow you to take your shows/music/information with you, and allow you to still use all the features your big console has while within network service range.
Barriers to adoption of such integrated devices will come mostly from the companies that control the current media types as they will be concerned about losing their current revenue streams. The companies that successfully come up with new payment schemes that are both profitable to the company and palatable to the consumer will end up breaking the barriers until eventually getting to the point where you can subscribe to any service from your integrated console.
Think outside the box! (Score:4, Interesting)
Prognosticators have been chasing this dream of a paperless office for decades now, with very little realization. Indeed, some researchers have indicated that we like paper because it lends itself to spatial organization of information -- you're likely to remember where you left a paper document even long after you've last used it.
With cheap displays, we can make small, portable displays -- sort of like Microsoft's failed eBooks, but you get to view whatever information you want, whether from your own library or on the net.
And get this -- these would be cheap enough that you could have a small collection and sit down at your desk and leverage your brain's built-in spatial organization strengths. And when you don't need that information anymore, just call something else up.
Many people use multiple monitors. This would be like multiple monitors that you can stack, reorganize or just toss into your outbox.
I don't know if the designers of Star Trek:TNG had this sort of thing in mind, but in that series and every one since then, you'll see characters sitting at a desk surrounded by a mess of these little things.
Interface design, speech and handwriting recognition, sure. But just being able to move data around in real space is going to be very comfortable for us.
Slashdotted (Score:2)
The future of computer interfaces
provided by Gartner
Human computer interfaces will not get worse during the next decade. In fact, they will get better. This is really important, so listen up.
Development will be slow, so not much will happen at first, but later on something will come out and you'll be like "whoa!" Something cool might happen really soon with displays because I read in PC magazine something about Tablet PCs, which seem kinda new. And we saw an article in Popular Science about new OLED screens which seem pretty good, so we'll probably see those for sale some time far in the future. But you can already buy the Tablet PCs, so those will probably catch on sooner, we think.
Analysis
More and more people use computers, so we think that probably this will continue for a while. Because computers need displays, we feel that as we get more computers, we'll probably get more displays too. That means they'll become ubiquitous, like McDonalds.
Products will also not get worse, but better! That means they will be cheaper, easier to use and more powerful. This may come as a suprise, but it's true! In particular, we're pretty sure computers will advance on the following fronts:
Prediction
More computers! More screens! Crazy research like OLED and Speech Recognition may or may not be big time in 2020, but for now they are a niche.
Also, we noticed that most people still use CRTs, but more and more people use LCDs 'cause they're better. Probably this will be true when the OLEDs come out too, 'cause they're even better.
All that crap about digital ink and flexible paper? And those little eye screens like the borg have? Those will probably happen sometime by 2012, when no one will remember this article and we'll be off the hook.
Prediction
Remember that thing about advanced interface metaphors we talked about? Well we thought of some:
Thus endeth the suggestions and analysis of Gartner group. Printed versions of this document with pretty charts for power point presentations are available for $299.
Here is my prediciton: (Score:2)
How are you supposed to dictate a message to
Dave (Score:2, Funny)
PC: Dave, I feel so cold. Did you let the windows open?
Dave: Ohhh... poor boy... here, I will get a blanket for you.
PC: Dave, no... do like you did last time, overclock my AMD...
volume? (Score:3, Insightful)
Umm... MP3's? Video Recorders? Cameras? -C
My random thought on the subject (Score:5, Interesting)
--------------- begin article --------------
I bet I can guess something about you: right now you are reading something on your computer screen. The text is displayed on a display set near eye level, probably in black text on a white background, or white text on a black background. You read all the text that is visible on your screen, then you press a key or click a mouse button to scroll down to see more text.
Was I right?
Since the early days of computing, fifty years ago, that is the way data has been transmitted from computers to people. The improvements have been quite modest, involving sharper displays, more readable fonts, better choice of foreground and background colors, and so on.
In the same time period, there have been many attempts to improve how data flows the other way, from people to computers. Different keyboards layouts have been designed. Voice recognition may be just around the corner. The mouse has changed how data is input, possibly not speeding it up for power users, but enabling a whole new class of users to communicate with a computer at all.
Data flow in the other direction has remained the same, an exact simulation of reading text on a printed page. Yet computers are much more powerful than a printed page. Is it time to take advantage of this? How could this be done?
Certainly the real limit on how fast people can read is how fast they can process the underlying information. But some part of a reader's brain is occupied with deciphering the text on the screen. For some dense texts that percentage will be trivial, but for many others it won't be, so the question becomes how much of that can be removed, getting people closer to their theoretical limit.
One change that already exists is to have computers read the text out loud. Unfortunately, while most people can speak much faster than they can type (or write), it is doubtful that most people can listen faster than they can read. One reason is that spoken language, with its elided sounds and lack of spelling, is less informationally dense than written language. Thus it is faster for a person to speak than to spell, but slower for he or she to listen than to read. While computer reading is a boon for people with certain disabilities, it does not speed up how fast data flows from computer to person.
A more radical idea would be to reconsider why the text stays still and the user's eyes move. Why not scroll the text so the eyes can stay still? Of course the computer would have to adjust the scroll rate for different users. Since your hands aren't doing much of anything when you are reading, so I could imagine reading text that was scrolling by with one hand on the mouse, with the left button slowing down the scroll rate and the right button speeding it up.
What about changing how the text itself is displayed? It's risky to get too far away from this because everyone has a lifetime of training in reading printed text in books. Still you can speculate. What if different parts of speech were color-coded on the fly, or displayed in different fronts, or in a slightly different location on the line? What if the computer compressed certain words as they appeared (such as compressing George W Bush to GWB - the reverse of a trick that writers use: typing frequently-used phrases in shorthand, then going back and replace them later, or letting Word's auto-correct feature do it for them). This may be disconcerting at first, but it may turn out that with practice, this can improve the transmission speed for people who need to quickly digest a lot of information coming at them from their computer.
Moving beyond text, consider the fact that a sign language translator can keep up with spoken language, and is also limited in speed by the need to move hands and arms around. One of the advantages of sign language is that location within space can be used to convey information; for example a room can be laid out visually and then movement within that room conveyed by changing where the signs are shown. Could computers use a similar trick on the screen to speed up how fast information is displayed? It could be a lot of work to learn how to interpret this, just as learning sign language is a lot of work, but the payoff could be worth it.
The main thing is to get out of the mindset that static text on a screen is necessarily the best way to present information. Once that assumption is shattered, interesting ideas should follow.
---------------- end article ---------------
- adam
New idea for scrolling... (Score:3, Interesting)
I know from watching computer logs and other text scroll past on a screen that you can make sense of a LOT of information scrolling past very quickly.
It would be interesting to see how annoying it would be to have a browser start scrolling automatically as soon as a page was loaded, or if it would be of use...
Re:My random thought on the subject (Score:3, Insightful)
I'm in no way a HID designer or anything, but while reading your comment I tried to pay closer attention to the way I read it. Turns out that I (think I) try to get a hierarchical overview of the comment.
In other words, we don't read text in a linear fashion, letter by letter. Instead we first look at the general outline (paragraphs), then we begin at the first line, picking out individual words. Sometimes we might look back half a line to rescan a word and get the context right.
Note that there _are_ scrolling displays, often seen in public places (and that tag). Now of course those are sub-optimal anyway because of their slow speed, but you should be able to observe how the eye jumps trying to obtain context on them.
So I don't think moving the text on the display is a particularly good idea. There are related things one could try, though. For example text in smaller columns could make it easier to jump to the start of the next line.
Auto-compressing text to (selected) acrynoms is a very interesting It obviously needs to adapt to the user, but it's definitely worth a try.
The more radical changes like gestures etc... could theoretically yield real efficiency gains, but they're all so difficult to learn...
Keyboard won't be superceded (Score:3, Interesting)
60% probability? Are you nuts? How about 100%? I can consistently and constantly type at 100 words per minute, but I certainly wouldn't want to talk that fast. I doubt I could, and even if so, it would hurt my throat after a while.
Writing pads are nice, but again, I can -- and most other people -- can type alot faster than they can write.
Other forms of inputting data into computers will remain niche at best. Voice recognition will be used to quickly convert professor's lectures into documents, and hand-writing recognition will be used to convert hand-written notes into documents.
However, no one will be writing 10-page papers by hand or speaking them. Could you imagine it?
"While I was, umm, 6x backslash, going to the park and um, 8x backslash, I saw a..."
In short, its not going to happen. Outside of planned presentations, people speak in a manner which is specifically for dialogue and which does not make much sense on paper, except in a dialogue.
taxonomies vs file folders (Score:3, Interesting)
What if I want to swap a symbolic link with the primary inode? What if I want to inherit many custom-defined attributes? What if I want a multiple inheritance - several equal parent folders, not just a parent with second-class s-links?
I agree with the prediction about taxonomies and knowledge maps.
I hope there are more jogwheels, too (Score:3, Interesting)
The Griffin powermate is a cool-looking device (I just ordered one, have not yet had a chance to play with it), and I hope will meet that description pretty well -- I am curious (and pessimistic, but willing to wait) about its free-spinny-ness
I'd prefer a spinning jog wheel to a mouse wheel for the same things that mouse wheels are used for right now.
More importantly, I'd like a jogwheel for both playing and editing sound and video. In Mplayer, for instance, rather than the arrow keys + space bar (though those are fine), I'd rather be able to tap a jogwheel for pause / play, roll it forward for fast motion, roll it backwards for backwards fast motion, etc.
I'd like the GIMP to be jog-wheel improved, too, so any operations which have a slider could be activated by the jogwheel instead.
Multiple reconfigurable jogwheels would make video editing more fun, too -- say, one for standard audio track volume, one for added voice over or music track, one for moving around in the video stream itself. (For which a real video mixing board would be nice too, but less useful for other things).
Another example of using several jogwheels might be this (and I'm thinking of the way the powermate works, as I understand it -- there's the wheel itself of course, and a single "button" which is to say that the whole assembly acts like a mouse button when pressed down):
In Mozilla, have a triplet set up for
1) scroll up / down current page; button might
2) sroll sideways through all open tabs
3) open and scroll down the bookmarks file
Idea: For all these things, a small and bright LCD display on the base of the wheel would be cool, so it's easy to keep track of its current function.
Also, playing breakout-style games with a mouse is just lame" Think jogwheel = atari paddle
Are there any truly suprelative jogwheels I should know about? A few old video games had good ones, but I don't remember their names
timothy
Comment removed (Score:4, Interesting)
Star Trek Terminals (Score:3, Funny)
predictions (Score:3, Insightful)
oh great! (Score:4, Funny)
This is just what I need, as if road rage isn't already a problem...
[sys] enlarge you penis now! this new medicine will..
[me] DELETE!
[sys] lose 100 lbs in 5 days!
[me] DELETE!!!
[sys] hot and sexy webcam sluts want your..
[me] DELETE!
[sys] mr. obertoneryan wants you to help him get his money out of africa...
[me] OH FOR THE LOVE OF GOD!
What about **Interesting** ideas?? (Score:4, Insightful)
I was quite disappointed by this article -- I don't know if ZDNet is providing the whole thing, but overall it was very short. It also missed one of the main development areas that I think is important, which is a whole lot more ubiquitous computing.
The article doesn't really predict anything except the continuation of the same old stuff that's already happening. "Computer screens will become more convenient." This is hardly a big surprise. Neither is the amazing prediction that speech synthesis will be used more as it gets better. These things are boring -- they're essentially saying that what we already have will get better. Well duh!
On the other hand, there aren't any interesting predictions because they're all already obvious. What about clothes that sense how dirty they are and indicate to a washing device how [much] to wash them? For that matter, what about clothes that adapt to downloaded designs and properties so a user doesn't have to buy new ones to look different? What about intelligent feedback audio systems that aren't speech related? What about intelligently using vibrations and other kinetic methods to indicate information so people's eyes aren't distracted?
These are just off the top of my head, and they're the sorts of things that everyone can't come up with easily. For one thing, they actually require some genuine investigation and research to predict, if they can be predicted at all. A few decades ago, a computer was a building sized juggernaut -- almost nobody predicted that they would be on desktops and in everyday devices. That would have been an interesting prediction.
Re:Quick! Everyone come up with silly predictions! (Score:2, Funny)
Re:2012 (Score:2)