Future of eReading: Following your eyes?


Text 2.0 uses infrared light and a camera to track eye movement across a screen.
Text 2.0 uses infrared light and a camera to track eye movement across a screen.

As eReading devices and the software that runs them become more advanced in an increasingly competitive market, researchers are creating applications that could take reading to a whole new level, with tools such as Text 2.0—a reading technology that personalizes the user’s experience by tracking eye movements.

Created by researchers Ralf Biedert, Georg Buscher, and Andreas Dengal at the German Research Center for Artificial Intelligence (DFKI), Text 2.0 uses eye-tracking technology from Tobii Technology (a Stockholm-based startup that just closed $21.5 million in Series B funding from venture capitalists), along with HTML, CSS, and JavaScript, to customize reading based on signals sent by eye movement.

Text 2.0 uses infrared light and a camera to track eye movement across a screen, and it uses this information to infer a user’s intentions during the course of reading.

For example, taking more time to read certain words, phrases, or names could trigger the appearance of sound effects, footnotes, translations, biographies, definitions, or animations. If the user begins skimming the text, the tracker will begin fading out words it deems less important to the text. If the reader glances away, a bookmark automatically appears, pointing to where the user stopped reading.

Though many new technologies now beginning to surface are experimenting with hands-free controlling of software and devices—such as electrocorticography (ECoG), in which a sheet of electrodes is laid directly on the surface of the brain to allow for “mind typing” and performing computer activities based solely on brain stimuli—some observers say Text 2.0 is a just-right futuristic technology that’s already generating interest from major companies.

One of these is reported to be Apple Inc., a company known to take risks on highly-profitable technologies, such as the computer mouse in 1984 and the touch screen in 2007.

In a 2007 patent filing by Wayne Westerman and John Elias, co-founders of the Fingerworks firm acquired by Apple during the development of the original iPhone, the patent details a handful of these newly proposed inputs under the title “Multi-Touch Data Fusion.”

The pair of engineers note in their filing that while the touch technology gives users more control, fusing additional information from other “sensing modalities” can enhance a device or improve its overall ease of use.

These sensing modalities can include voice fusion, finger ID fusion, facial expression fusion, biometrics fusion, and Gaze Vector fusion—a technology from Tobii Technology.

Microsoft Corp. also recently has been backing Tobii’s technology.

At this year’s Computex, an information technology show based in Taipei, Microsoft demonstrated eye control from Tobii. Interested participants were placed in front of a screen and were told simply to “interact by using your eyes.”

Meris Stansbury

Want to share a great resource? Let us know at submissions@eschoolmedia.com.