There's no mistaking the significant influence of personal computers and the Internet on our modern way of life. Many of us have so quickly adapted to regular use of search engines and web surfing that it's difficult to imagine life without the Internet.
The Internet allow us to research products and companies, share ideas with the public, research nutritional supplements, find articles on historical figures, and do a million other things that simply weren't possible a mere two decades ago.
And yet our interface with the Internet remains the lowly personal computer. With its clumsy interface devices (keyboard and mouse, primarily), the personal computer is a makeshift bridge between the ideas of human beings and the world of information found on the Internet. These interface devices are clumsy and simply cannot keep pace with the speed of thought of which the human brain is capable.
Consider this: a person with an idea who wishes to communicate that idea to others must translate that idea into words, then break those words into individual letters, then direct her fingers to punch physical buttons (the keyboard) corresponding to each of those letters, all in the correct sequence. Not surprisingly, typing speed becomes a major limiting factor here: most people can only type around sixty words per minute. Even a fast typist can barely achieve 120 words per minute. Yet the spoken word approaches 300 words per minute, and the speed of "thought" is obviously many times faster than that.
Pushing thoughts through a computer keyboard is sort of like trying to put out a raging fire with a garden hose: there's simply not enough bandwidth to move things through quickly enough. As a result, today's computer / human interface devices are significant obstacles to breakthroughs in communicative efficiency.
The computer mouse is also severely limited. I like to think of the mouse as a clumsy translator of intention: if you look at your computer screen, and you intend to open a folder, you have to move your hand from your keyboard to your mouse, slide the mouse to a new location on your desk, watch the mouse pointer move across the screen in an approximate mirror of the mouse movement on your desk, then click a button twice. That's a far cry from the idea of simply looking at the icon and intending it to open, which would of course be the desired level of computer / human interface as I'll discuss below.
Today's interface devices are little more than rudimentary translation tools that allow us to access the world of personal computers and the Internet in a clumsy, inefficient way. Still, the Internet is so valuable that even these clumsy devices grant us immeasurable benefits, but a new generation of computer/human interface devices would greatly multiply those benefits and open up a whole new world of possibilities for exploiting the power of information and knowledge for the benefit of humanity. Let's take a closer look at those emerging technologies now.
Emerging Computer / Human Interface Technologies
The idea of eliminating the gap between human thought and computer responsiveness is an obvious one, and a number of companies are working hard on promising technologies
. One of the most obvious such technologies is voice recognition software that allows the computer to type as you speak, or allows users to control software applications by issuing voice commands.
The most advanced and accurate software in this category is Dragon Naturally Speaking, and I've spent a considerable number of hours with this software. Its accuracy is impressive, and the technology is far ahead of voice recognition technology from a mere decade ago, but it's still not at the point where people can walk up to their computer and start issuing voice commands without a whole lot of setup, training, and fine tuning of microphones and sound levels. For many people, that's just way too much configuration.
This situation is no doubt recognized by the developers of Dragon Naturally Speaking. Nevertheless, widespread, intuitive use of voice recognition technology still appears to be years away.
Another recent technology that represents a clever approach to computer / human interfaces is the iGesture Pad by a company called Fingerworks (http://www.FingerWorks.com). With the iGesture Pad, users place their hands on a touch sensitive pad (about the size of a mouse pad), then move their fingers in certain patterns (gestures) that are interpreted as application commands. For example, placing your fingers on the pad in a tight group, then rapidly opening and spreading your fingers is interpreted as an Open command.
This technology represents a leap in intuitive interface devices, and it promises a whole new dimension of control versus the one-dimensional mouse click, but it's still a somewhat clumsy translation of intention through physical limbs.
For more intuitive control of software interfaces, what's needed is a device that tracks eye movements and accurately translates them into mouse movements: so you could just look at an icon on the screen and the mouse would instantly move there. Interestingly, some of the best technology in this area comes from companies building systems for people with physical disabilities. For people who can't move their limbs, computer control through alternate means is absolutely essential.
Head movement tracking technology
One approach to this is tracking the movement of a person's head and translating that into mouse movements. One device, the HeadMouse (http://www.orin.com/access/hme/index.htm), does exactly that. You stick a reflective dot on your forehead, put the sensor on top of your monitor, then move your head to move your mouse. I haven't tried the technology, so I can't say how well it works, but the company (Origin Instruments) has a reputation for providing assistive technologies to physically disabled persons, and the HeadMouse is their latest technology.
Another company called Madentec (http://www.Madentec.com) offers a similar technology called Tracker One. Place a dot on your forehead, then you can control the mouse simply by moving your head.
In terms of affordable head tracking products for widespread use, a company called NaturalPoint (http://www.NaturalPoint.com) seems to have the best head tracking technology at the present: a product called SmartNav, priced at a mere $199, allows for hands-free mouse control via head movement. Add a foot switch and you can click with your feet. I've used this product myself, and while it definitely presents a learning curve for new users, it works as promised.
Tracking eye movements
While tracking head movement is in many ways better than tracking mouse movement, a more intuitive approach, it seems, would be to track actual eye movements. A company called LC Technologies, Inc. is doing precisely that with their EyeGaze systems (http://www.lctinc.com/PRODUCTS.htm). By mounting one or two cameras under your monitor and calibrating the software to your screen dimensions, you can control your mouse by simply looking at the desired position on the screen.
Once again, this technology was originally developed for people with physical disabilities, yet the potential application of it is far greater. In time, I believe that eye tracking systems will become the preferred method of cursor control for users of personal computers.
Eye tracking technology is quickly emerging as a technology with high potential for widespread adoption by the computing public. Companies such as Tobii Technology (http://www.tobii.se), Seeing Machines (http://www.SeeingMachines.com), SensoMotoric Instruments (http://www.smi.de), Arrington Research (http://www.ArringtonResearch.com), and EyeTech Digital Systems (http://www.eyetechds.com) all offer eye tracking technology with potential for computer / human interface applications. The two most promising technologies in this list, in terms of widespread consumer-level use, appear to be Tobii Technology and EyeTech Digital Systems.
Mind control for your PC
Moving to the next level of computer / human interface technology, the ability to control your computer with your thoughts alone seems to be an obvious goal. The technology is called Brain Computer Interface technology, or BCI.
Although the idea of brain-controlled computers has been around for a while, it received a spike of popularity in 2004 with the announcement that nerve-sensing circuitry was implanted in a monkey's brain, allowing it to control a robotic arm by merely thinking. This Washington Post article gives a fascinating account of the breakthrough and training required by the monkey to learn how to use the brain implant:
The lead researchers in the monkey experiment are now involved in a commercial venture to develop the technology for use in humans. The company, Cyberkinetics Inc. (http://www.cyberkineticsinc.com) hopes to someday implant circuits in the brains of disabled humans, then allow those people to control robotic arms, wheelchairs, computers or other devices through nothing more than brain behavior.
A key obstacle to widespread use is, of course, the requirement that circuitry be surgically implanted in the brain. If the technology can take a quantum leap and work its magic without needing the surgery -- by wearing a sensing helmet, for example -- it will suddenly be a lot more interesting to the population at large, and not just those with severe physical disabilities.
Imagine the limitless applications of direct brain control. People could easily manipulate cursors on the screen or control electromechanical devices. They could direct software applications, enter text on virtual keyboards, or even drive vehicles on public roads. Today, all these tasks are accomplished by our brains moving our limbs, but the limbs, technically speaking, don't have to be part of the chain of command.
Another promising area of computer / human interface technology is being explored by companies like Immersion Corporation (http://www.Immersion.com), which offers tactile feedback hardware that allows users to "feel" their computer interfaces.
Slide on Immersion's CyberGlove, and your computer can track and translate detailed hand and finger movements. Add their CyberTouch accessory, and tiny force feedback generators mounted on the glove deliver the sensation of touch or vibration to your fingers. With proper software translation, these technologies give users the ability to manipulate virtual objects using their hands. It's an intuitive way to manipulate objects in virtual space, since nearly all humans have the natural ability to perform complex hand movements with practically no training whatsoever.
Another company exploring the world of tactile feedback technologies is SensAble Technologies (http://www.sensable.com). Their PHANTOM devices allow users to construct and "feel" three-dimensional objects in virtual space. Their consumer-level products include a utility for gamers that translates computer game events into tactile feedback (vibrations, hitting objects, gun recoil, etc.).
On a consumer level, Logitech makes a device called the IFeel Mouse that vibrates or thumps when your mouse cursor passes over certain on-screen features. Clickable icons, for example, feel like "bumps" as you mouse over them. The edges of windows can also deliver subtle feedback. The mouse sells for around $40, but it hasn't seen much success in the marketplace. Reviews from users reveal that the vibrating mouse is considered more annoying than helpful, so don't expect to see this technology taking over the world of computer mice.
But tactile feedback has potential for making human / computer interfaces more intuitive and efficient, even if today's tactile technologies are clunky first attempts. The more senses we can directly involve in our control of computers, the broader the bandwidth of information and intention between human beings and machines.
The long-promised 3D computer monitor finally seems to be close to reality. Manipulating complex windows, documents and virtual objects on a two-dimensional display -- as is standard today -- is rather limiting. With a 3D monitor, we could work in layers or position documents and objects in 3D space rather than squeezing them down to a tiny toolbar at the bottom of one screen.
For human beings, 3D space is intuitive. We get it without training. That's because we live in a world of 3D objects and space, and our perception is hard-wired to understand spatial relationships. That's why gamers who play first-person shooters like Quake can mentally retrace their way through enormous maps (levels) in their heads, eyes closed, without even trying: the human brain was built to remember and navigate 3D space.
Recent breakthroughs in 3D displays promise to make computing more intuitive and powerful. Companies like LightSpace Technologies (http://www.lightspacetech.com) are already selling desktop 3D display monitors that display true 3D images without the need for special glasses.
The trouble is, Windows and Mac operating systems weren't written with 3D displays in mind. So there's no capability to stack windows or view the depth of objects. It's a classic chicken-and-egg conundrum: who's going to buy 3D displays if the software can't support them, and why would software makers write 3D layering logic if nobody owns the displays?
In time, thanks to the "cool" factor of 3D displays, the technology will eventually receive enough attention to warrant the necessary R&D investment by operating system developers like Microsoft and Apple. No doubt, future generations will conduct all their computing with the aid of 3D displays, and the very idea of 2D displays will seem as outdated as black & white movies do to us today.
Another new 3D display device is the Perspecta Spatial 3D globe, seen at:
http://www.actuality-systems.com/index.php/actuality . This device displays 3D objects or animations inside a globe. Users can walk around the globe and view the objects from any angle. It's a rather expensive item, of course, so early applications for this product focus on medical and research tasks. In time, however, the technology will drop in price, bringing it within reach of more consumers.
In the category of the more familiar, a German company called SeeReal Technologies (http://www.SeeReal.com) offers a 20" LCD 3D display that uses eye tracking combined with unique left/right display technology to create a true 3D image on a flat panel monitor without the need for special viewing glasses. These monitors are typically used in the CAD/CAM industry where the visualization of 3D objects is especially helpful. The lack of support for 3D space in the Windows operating system, however, makes these monitors useless for everyday users... at least for the moment.
What would 3D displays do for us?
So what should a flat panel 3D display actually do for a typical Windows or Mac user? At the most basic level, operating systems would need to support fundamental 3D features like:
- Layering of windows: Background windows would appear further away, while foreground windows appear closer.
- Pop out elements: Certain elements of a document or page could appear to "pop out" of the screen a half inch or so. This might be used similarly to bolding or italicizing.
- Floating cursors: the mouse cursor appears to float above the screen and then, when clicked, it actually buries itself in the button being clicked, then quickly returns to its hover status.
Note, however, that a 3D flat panel monitor is not the same as a true 3D display system: you can't walk to the side of the monitor and see the windows behind it. It's still essentially a 2D system in that it can't display true volumetric shapes and objects that are viewable from multiple angles.
Tabletop 3D displays
For that, we'll ultimately need a tabletop 3D display system that lays flat on your desk (like an LCD monitor laying down) and projects 3D images into the space above the panel. This would be a true volumetric 3D display system, and it's here that the technology truly represents a breakthrough. Program application windows could literally be stacked from the rear to the front, and if you peeked around the side of the display, you could see a side view of all the windows at once.
With proper software control, objects or documents could be placed in true 3D space: desktop icons, for example, could be lined up along the very back row. Games could display true 3D scenes as if you're actually in them, and CAD engineers would have the ability to observe their designs in true 3D space.
Better yet, if coupled with a motion tracking glove or similar technology, users could use their hands to grasp, move, resize or otherwise manipulate elements in 3D space. This, of course, opens up an unlimited universe of possibilities for computer / human interaction.
Closing the gap
This brief tour of computer / human interface technologies is really only a glimpse of what's possible. It's all about closing the gap between human intention and computing systems. Today, the gap is very large: a typical keyboard and mouse setup is essentially a two-channel interface system. But tomorrow, the gap could be very small: add a head tracking system, hand-sensing glove, foot pedal switches, voice recognition system, 3D display and a brainwave-sensing helmet, and you've created layers of multi-channel interface technologies that allow infinite expression.
In time, as this technology is developed and adopted by mainstream users, the gap will continue to shrink. This has enormous positive implications in the workplace, medicine, science, education, social interaction, entertainment and many other areas, which is why it earns such a lengthy discussion in this report. And it's not technology that's "way out there," either: it's technology that's emerging now and will continue to be developed in the years ahead.
Next: Part 9 - Vibrational medicine
This article has been adapted from, The Ten Most Important Emerging Technologies For Humanity, an ebook by futurist Mike Adams.
About the author: Mike Adams is a natural health researcher, author and award-winning journalist with a strong interest in personal health, the environment and the power of nature to help us all heal He is a prolific writer and has published thousands of articles, interviews, reports and consumer guides, and he is well known as the creator of popular downloadable preparedness programs on financial collapse, emergency food storage, wilderness survival and home defense skills. Adams is an honest, independent journalist and accepts no money or commissions on the third-party products he writes about or the companies he promotes. In 2010, Adams created TV.NaturalNews.com, a natural living video sharing site featuring thousands of user videos on foods, fitness, green living and more. He also founded an environmentally-friendly online retailer called BetterLifeGoods.com that uses retail profits to help support consumer advocacy programs. He's also the founder and CEO of a well known email mail merge software developer whose software, 'Email Marketing Director,' currently runs the NaturalNews email subscriptions. Adams also serves as the executive director of the Consumer Wellness Center, a non-profit consumer protection group, and practices nature photography, Capoeira, martial arts and organic gardening. He's also author of numerous health books published by Truth Publishing and is the creator of several consumer-oriented grassroots campaigns, including the Spam. Don't Buy It! campaign, and the free downloadable Honest Food Guide. He also created the free reference sites HerbReference.com and HealingFoodReference.com. Adams believes in free speech, free access to nutritional supplements and the ending of corporate control over medicines, genes and seeds. Known by his callsign, the 'Health Ranger,' Adams posts his missions statements, health statistics and health photos at www.HealthRanger.org
Have comments on this article? Post them here:
people have commented on this article.