by Mark
Ollig
We process information using touch,
smell, hearing, taste, and sight.
Technology giant IBM predicts we will
soon be using advanced computing technology to augment these five senses.
IBM recently released their “5-in-5”
technology prognostications for the next five years.
These predictions focus on developing
advanced, cognitive computing sensory systems.
During the 1980s, AT&T ran a
commercial about using your telephone to “reach out and touch someone.”
Within five years, our mobile devices
and smartphone touch display screens may allow us to actually feel what those
Egyptian cotton sheets we are looking at online are like, by “touching” the
picture on the display screen.
Digital image correlation processing,
and infrared and haptic technologies producing calibrated vibrations on the
display screen will allow the capturing of the texture qualities our sense of
touch will be able to distinguish.
“By matching variable-frequency
patterns of vibration to physical objects, so that when a shopper touches what
the webpage says is a silk shirt, the screen will emit vibrations that match
what our skin mentally translates to the feel of silk,” the IBM 5-in-5 report
said.
IBM Research says they are bringing
together the virtual (online) and real-world shopping experiences to our
smartphones.
Soon, we may be able to use our sense
of touch to get physical feedback via the display screen on our smartphones.
To watch an IBM video on the future
of touch, go to http://tinyurl.com/cdy22ra.
Smartphones in the future will also
possess the sense of smell. IBM says tiny sensors having the ability to “sniff”
will be contained within our smart mobile devices.
These future smartphones will detect
certain chemical biomarkers that can be found in the air.
The biomarkers from our breath will
be analyzed (via cognitive computing processing) through our smartphones. This
type of application could be used to identify any existing health conditions.
Also, within the next five years,
when in the grocery store, we might be able to check the freshness and quality
of the meat, fruit, and produce by having our smartphones “sniff” the item.
To view the IBM video on smell
technology, go to http://tinyurl.com/cwxuydb.
Within five years, we may be
downloading a hearing application onto our smartphones.
IBM predicts the sounds emanating
from various sources; human, nature, animals, and even the weather will be
received and coded into our smartphones and processed into meaningful
information for us.
Hearing sensor technology using
auditory signaling processing to more efficiently extract and transform sound
into information the human brain can comprehend would be applicable to hearing
aids, or cochlear implants.
An IBM video on futuristic hearing is
at http://tinyurl.com/cn9sb88.
How will IBM Research develop
computing cognitive systems having the ability to taste?
By having a professionally trained
chef, now a computer engineer, on their team.
IBM is developing a cognitive
computing system which analyzes the chemical compounds in food, and how they
react with each other.
By using this information, along with
psychophysical data modeling on specific chemicals in food which creates
perceptions of “pleasantness, familiarity, and enjoyment,” IBM is hoping the
result will be new, unique food recipes containing precise ingredient
combinations that are “scientifically flavorful.”
IBM’s cognitive taste technology
video is at http://tinyurl.com/dyg8ajx.
Seeing is believing, and IBM research
is working on cognitive computer vision systems that, according to IBM, will
“help us understand the 500 billion photos we’re taking every year.”
Although we know how the human eye
processes images, today’s computing technology is unable to exactly replicate
human sight.
IBM is trying to get the computer to
“reason out” what a particular image is, by having it detect patterns in a
photo, or determine what it is “seeing” from video taken on a smartphone.
For example, by providing photos of
various sandy beach scenes to a cognitive computing program, it is thought the
program would eventually be able to distinguish the individual features, color
distributions, texture patterns, surroundings, and other information.
Motion-related information data would be used if the source came from a video.
The cognitive programming would “learn”
the distinct features of various beach scenes and their surroundings.
It is hoped that with enough
learning, the cognitive computer would be able to distinguish one sandy beach
from another, and thus determine if a shoreline it “sees” is located in France
or in California.
The idea is to have a cognitive
computer learn various outdoor environments, including outdoor structures,
plants, trees, animals, birds – you get the idea. Thus, the next time we are in
France, and hold our smartphone’s camera in front of say, the Eiffel Tower (or
an outdoor location we are unfamiliar with), the phone will “recognize” it and
provide us with detailed information.
IBM’s sight technology video is at
http://tinyurl.com/ccmbp2f.
This columnist would like to say
Merry Christmas to his mom, and to all his faithful readers.