Home > Research > Publications & Outputs > Multi-Level Sensory Interpretation and Adaptati...

Electronic data

View graph of relations

Multi-Level Sensory Interpretation and Adaptation in a Mobile Cube

Research output: Contribution to conference Conference paper

Published
Publication date2003
Number of pages7
Pages111-117
<mark>Original language</mark>English

Conference

ConferenceProc. of the third workshop on Artificial Intelligence in Mobile Systems (AIMS) at Ubicomp 2003
CitySeattle, WA, US
Period1/01/00 → …

Abstract

Signals from sensors are often analyzed in a sequence of steps, starting with the raw sensor data and eventually ending up with a classification or abstraction of these data. This paper will give a practical example of how the same information can be trained and used to initiate multiple interpretations of the same data on different, application-oriented levels. Crucially, the focus is on expanding embedded analysis software, rather than adding more powerful, but possibly resource-hungry, sensors. Our illustration of this approach involves a tangible input device the shape of a cube that relies exclusively on lowcost accelerometers. The cube supports calibration with user supervision, it can tell which of its sides is on top, give an estimate of its orientation relative to the user, and recognize basic gestures.