Selker is an Associate Professor at the MIT Media and
Arts Technology Laboratory and the Director of the Context
Aware Computing Lab (www.media.mit.edu/context). Prior
to joining MIT faculty, in November 1999, Selker was
an IBM Fellow and ran a large research group called
User Systems Ergonomics Research at IBM Research. He
has served as a consulting professor at Stanford University,
taught at Hampshire University of Massachusetts, Amherst
and Brown Universities. Selker also worked at Xerox
PARC and Atari Research Labs.
research has borne fruit in several product lines and
has contributed to products ranging from notebook computers
to operating systems. He is known for the design of
the "TrackPoint" in-keyboard pointing device found in
many notebook computers. His PhD work "COACH" demonstrated
that an adaptive agent can improves user performance.
work is documented in dozens of industry awards, papers
and invention patents .
scientist's answers are available below.Please see our resources
page for the scientists home page and other related infomation.
a special education student at ETSU and I am interested
in how this technology could be used as communication
devices for non verbal students. Any chance that this
could become affordable enough (and user friendly enough)
that it could be put to actual use by non-communicative
students instead of stuck on a shelf like most of the
first and second generation communication devices? Particularly
the eye gaze technology for kids that have no outward
signs of pleasure responses? Super work! You give us
The eye gaze work in Eye aRe is very cheap, the
circuit board and battery Are the most expensive things
in it! This system could be developed to Recognize stare
(interest) blinking rapidly (nervousness, discomfort),
winking, eyes open and eyes closed. Such a set of things
could be used by people to control their environment.
Someone would have to want to make the product and deploy
it... how about you!
the floor tell male from female? What other kind of
information could the floor be programmed to get?
Some work has been done on gait. It seems that we
can recognized various things by how people move and
walk. Still, the best things for the floor are, noticing
that things have moved, making paths for where they
have moved to (finding keys by showing you the paths
for where you have been) showing where people are associating,
making games on the floor, weighing food ingredients
or other things.