Contemporary visualization systems often make use of large monitors or projection screens to display complex information. Even very sophisticated visualization systems, that offer a wide variety of interaction possibilities and exhibit complex user interfaces, do usually not make use of additional advanced input and output devices. The interaction is typically limited to the computer mouse and a keyboard. One of the reasons for the lack of advanced interaction devices is the high cost of special hardware.
This thesis introduces the idea of Augmented Visualization. The aim of the project is to develop a novel interaction solution for projection walls as well as PC monitors using cheap hardware such as mobile phones or tablets. Several features of mobile devices will be exploited to improve the interaction experience. The main technical challenge of the project is to implement a solution for markerless visual tracking of the changing visualized scene. In the proposed setup, this also requires real-time wireless video streaming between the mobile device and the PC. The real-time tracking of the visualized scene will allow to estimate the six-degrees-of-freedom pose of the mobile device. The calculated position and orientation information can be used for advanced interaction metaphors like magic lenses. Moreover, for a group of experts who are analyzing the data in front of the same screen, we can provide a personal augmented view of the visualized scene, for each user on his/her personal device.
The thesis discusses the design questions and the implementation steps of an Augmented Visualization System, describes the prototype setup and presents the experimental results.