004 Datenverarbeitung; Informatik
Refine
Document Type
- Report (16)
- Doctoral Thesis (15)
- Article (7)
- Preprint (3)
- Part of a Book (2)
- Master's Thesis (2)
- Conference Proceeding (1)
Institute
- Junior-Professur Augmented Reality (16)
- Professur Computergestütztes kooperatives Arbeiten (9)
- Professur Content Management und Webtechnologien (5)
- F. A. Finger-Institut für Baustoffkunde (FIB) (2)
- Professur Informatik in der Architektur (2)
- Professur Medieninformatik (2)
- Professur Systeme der Virtuellen Realität (2)
- In Zusammenarbeit mit der Bauhaus-Universität Weimar (1)
- Medienkunst/Mediengestaltung (1)
- Professur Bauphysik (1)
Keywords
- CGI <Computergraphik> (9)
- Angewandte Informatik (8)
- Ubiquitous Computing (7)
- Maschinelles Sehen (6)
- Virtuelle Realität (6)
- Awareness (5)
- Association for Computing Machinery / Special Interest Group on Graphics (4)
- Gruppengewahrsein (4)
- Projektor-Kamera Systeme (4)
- Bildverarbeitung (3)
- Computer-Supported Cooperative Work (3)
- Computergraphik (3)
- Erweiterte Realität <Informatik> (3)
- Handy (3)
- Mensch-Maschine-Kommunikation (3)
- Museumsführer (3)
- Neuronales Netz (3)
- Objekterkennung (3)
- Projector-Camera Systems (3)
- projector-camera systems (3)
- Anpassung (2)
- Bildbasiertes Rendering (2)
- Bilderkennung (2)
- Bildsegmentierung (2)
- CAD (2)
- Content Management (2)
- Contrast (2)
- Editor (2)
- Group Awareness (2)
- High Dynamic Range (2)
- Hoher Dynamikumfang (2)
- Human-Computer Interaction (2)
- Information Retrieval (2)
- Konfiguration (2)
- Kontrast (2)
- Multi-User Virtual Reality (2)
- Museum (2)
- Privacy (2)
- Privatsphaere (2)
- Privatsphäre (2)
- Projektion (2)
- Projektionsverfahren (2)
- Radiometrische Kompensation (2)
- Rasterelektronenmikroskopie (2)
- Reciprocity (2)
- Rendering (2)
- Reziprozitaet (2)
- Reziprozität (2)
- Zementklinker (2)
- configuration (2)
- information sharing (2)
- museum guidance (2)
- museum guidance system (2)
- neural networks (2)
- object recognition (2)
- privacy (2)
- radiometric compensation (2)
- reciprocity (2)
- 1997 (1)
- 3D Interaction Techniques (1)
- AES-192 (1)
- AES-256 (1)
- ARIA (1)
- Ad-hoc-Netz (1)
- Advanced Encryption Standard (1)
- Aufsatz (1)
- Augmented Audio Reality (1)
- Automatisierungssystem (1)
- Bauplanung (1)
- Benutzerfreundlichkeit (1)
- Benutzerorientierung (1)
- Benutzerschnittstellen (1)
- Beteiligung (1)
- Bildkorrektur (1)
- Blende <Optik> (1)
- Blockchiffre (1)
- Bluetooth tracking (1)
- CAMShift (1)
- City (1)
- Collaboration (1)
- Collaborative Virtual Environments (1)
- Computer Graphics (1)
- Computer Supported Cooperative Learning (1)
- Computer Supported Cooperative Work (1)
- Computersimulation (1)
- Computertomographie (1)
- Computervermitteltes Arbeiten (1)
- Computervermitteltes Lernen (1)
- Content <Internet> (1)
- Datenbank (1)
- Datensammlung (1)
- Desktop-VR-Interface (1)
- Differentielle Kryptoanalyse (1)
- Display (1)
- EDX (1)
- Experimentalplattform (1)
- Facebook (1)
- Fallbasiertes Schließen (1)
- Freac (1)
- GPU Programming (1)
- Gebäude (1)
- Generierung (1)
- Geoinformationssystem (1)
- Gestaltungsrichtlinien (1)
- Graphische Benutzeroberfläche (1)
- Grindahl (1)
- Group Navigation (1)
- Gruppenbewusstsein (1)
- HAS-160 (1)
- Hash-Algorithmus (1)
- Hologramm (1)
- Human-Computer-Interaction (1)
- IKM (1)
- Immersion <Virtuelle Realität> (1)
- Industrie 4.0 (1)
- Informatik (1)
- Information (1)
- Inhalt (1)
- Interaction Fidelity (1)
- Interaction Techniques (1)
- Internet (1)
- Interoperabilität (1)
- Inverse Light Transport (1)
- Kalibrierung (1)
- Katastrophe (1)
- Katastrophenmanagement (1)
- Kernel-Based Tracking (1)
- Knowledge modelling (1)
- Kontextbezogenes System (1)
- Kryptoanalyse (1)
- Kryptologie (1)
- Kulturerbe (1)
- Kybernetik (1)
- Licht Transport (1)
- Localised Sound (1)
- Media Space (1)
- Mensch-Maschine-Interaktion (1)
- Metamodell (1)
- Metapher (1)
- Metaphors (1)
- Microscopy (1)
- Mikroskopie (1)
- Mobile Computing (1)
- Mobile phones (1)
- Model Predictive Control (1)
- Motivation (1)
- Multi-Frame Rate Composition (1)
- Multi-Frame Rate Rendering (1)
- Multi-Projektor Systeme (1)
- Museum Guidance (1)
- MySQL (1)
- Navigation (1)
- Neue Medien (1)
- Neural Networks (1)
- Nutzerpartizipation (1)
- OWL <Informatik> (1)
- Objektverfolgung (1)
- Ontologie (1)
- PHP (1)
- Parallel Rendering Methods (1)
- Peer-to-Peer-Netz (1)
- Pfadanalyse (1)
- Phone-to-phone communication (1)
- PhoneGuide (1)
- Positionsbestimmung (1)
- Problemlösen (1)
- Produktmodell (1)
- Projector Camera System (1)
- Projekor-Kamera System (1)
- Projektionsapparat (1)
- Projektionssystem (1)
- RadioGatun (1)
- Radiometric Compensation (1)
- Raum (1)
- Raumklang (1)
- Reconstruction (1)
- Rich Media Content (1)
- Räumliche Beziehungen (1)
- SHACAL-2 (1)
- Scanning (1)
- Sensor-Based Infrastructure (1)
- Sensor-basierte Infrastrukture (1)
- Service-Oriented Platform (1)
- Service-orientierte Plattform (1)
- Simulation (1)
- Smartphone (1)
- Social Media (1)
- Softwareentwicklung (1)
- Soziale Medien (1)
- Spatial Relationships (1)
- Stadt (1)
- Stadtinformationssystem (1)
- Stadtstruktur (1)
- Stereofonie (1)
- Stereophonie (1)
- Strukturvergleich (1)
- Subobject Detection (1)
- Subobjekterkennung (1)
- Suchmaschine (1)
- Suchverfahren (1)
- Suchverhalten (1)
- Textproduktion (1)
- Tiger (1)
- Usability (1)
- User Generated Content (1)
- User created content (1)
- User-Created Content (1)
- Userinterface-Design (1)
- VRML (1)
- Virtual Reality (1)
- Visual Quality (1)
- Visually Impaired (1)
- Visuelle Wahrnehmung (1)
- Web archive (1)
- Wegrouten (1)
- Weimar (1)
- Wissensmodellierung (1)
- adaption (1)
- adaptive image classification (1)
- block ciphers (1)
- calibration (1)
- cement clinker (1)
- cement microstructure (1)
- comparison of structure (1)
- computer grafik (1)
- computer graphics (1)
- computerenhanced (1)
- content production (1)
- contribution management (1)
- differential cryptanalysis (1)
- digital media (1)
- digitales Gebäudemodell (1)
- editor (1)
- focussed ion beam (1)
- game mechanic (1)
- hash function (1)
- image correction (1)
- image segmentation (1)
- imperceptible pattern projection (1)
- micro-CT (1)
- mobile ad-hoc networks (1)
- mobile phones (1)
- multi-projector systems (1)
- pathway awareness (1)
- pervasive tracking (1)
- projection (1)
- radiometrische Kompensation (1)
- simulation (1)
- slide attacks (1)
- social game (1)
- spatial augmented reality (1)
- superpixel (1)
- temporal adaptation (1)
- tomography (1)
- unsichtbare Muster Projektion (1)
- urban structures (1)
- visual perception (1)
Burning of clinker is the most influencing step of cement quality during the production process. Appropriate characterisation for quality control and decision-making is therefore the critical point to maintain a stable production but also for the development of alternative cements. Scanning electron microscopy (SEM) in combination with energy dispersive X-ray spectroscopy (EDX) delivers spatially resolved phase and chemical information for cement clinker. This data can be used to quantify phase fractions and chemical composition of identified phases.
The contribution aims to provide an overview of phase fraction quantification by semi-automatic phase segmentation using high-resolution backscattered electron (BSE) images and lower-resolved EDX element maps. Therefore, a tool for image analysis was developed that uses state-of-the-art algorithms for pixel-wise image segmentation and labelling in combination with a decision tree that allows searching for specific clinker phases. Results show that this tool can be applied to segment sub-micron scale clinker phases and to get a quantification of all phase fractions. In addition, statistical evaluation of the data is implemented within the tool to reveal whether the imaged area is representative for all clinker phases.
Schwerpunkt der Arbeit ist die Auseinandersetzung mit den Möglichkeiten und Grenzen der Desktop-VR als neue Generation der Benutzerschnittstellen. Besondere Bedeutung bei dieser Art des Interface-Designs kommt den Metaphern zu. Ein großer Teil der Arbeit beschäftigt sich mit der Klassifikation, der Auswahl und dem Einsatz passender Metaphern unter Berücksichtigung der in der Applikation darzustellenden Informationsinhalte. Aus der Kombination dieser beiden Merkmale (Art der Metapher, Informationsinhalt) ergeben sich vier verschiedene virtuelle Umgebungen, deren Eigenschaften und Besonderheiten konkretisiert und an Beispielen aus dem Anwendungsgebiet der Stadtinformationssysteme vorgestellt werden. Als praktischer Untersuchungsgegenstand dient das Anwendungsgebiet der Stadtinformationssysteme. Die theoretisch basierten Erkenntnisse und Schlußfolgerungen werden durch statistische Untersuchungen, in Form von Fragebögen zu Stadtinformationssystemen, überprüft und konkretisiert.
Im vorliegenden Beitrag wird ein Framework für ein verteiltes dynamisches Produktmodell (FREAC) vorgestellt, welches der experimentellen Softwareentwicklung dient. Bei der Entwicklung von FREAC wurde versucht, folgende Eigenschaften umzusetzen, die bei herkömmlichen Systemen weitgehend fehlen: Erstens eine hohe Flexibilität, also eine möglichst hohe Anpassbarkeit für unterschiedliche Fachdisziplinen; Zweitens die Möglichkeit, verschiedene Tools nahtlos miteinander zu verknüpfen; Drittens die verteilte Modellbearbeitung in Echtzeit; Viertens das Abspeichern des gesamten Modell-Bearbeitungsprozesses; Fünftens eine dynamische Erweiterbarkeit sowohl für Softwareentwickler, als auch für die Nutzer der Tools. Die Bezeichnung FREAC umfasst sowohl das Framework zur Entwicklung und Pflege eines Produktmodells (FREAC-Development) als auch die entwickelten Tools selbst (FREAC-Tools).
Superimposing Dynamic Range
(2008)
We present a simple and cost-efficient way of extending contrast, perceived tonal resolution, and the color space of static hardcopy images, beyond the capabilities of hardcopy devices or low-dynamic range displays alone. A calibrated projector-camera system is applied for automatic registration, scanning and superimposition of hardcopies. We explain how high-dynamic range content can be split for linear devices with different capabilities, how luminance quantization can be optimized with respect to the non-linear response of the human visual system as well as for the discrete nature of the applied modulation devices; and how inverse tone-mapping can be adapted in case only untreated hardcopies and softcopies (such as regular photographs) are available. We believe that our approach has the potential to complement hardcopy-based technologies, such as X-ray prints for filmless imaging, in domains that operate with high quality static image content, like radiology and other medical fields, or astronomy.
Superimposing Dynamic Range
(2008)
Replacing a uniform illumination by a high-frequent illumination enhances the contrast of observed and captured images. We modulate spatially and temporally multiplexed (projected) light with reflective or transmissive matter to achieve high dynamic range visualizations of radiological images on printed paper or ePaper, and to boost the optical contrast of images viewed or imaged with light microscopes.
We present a novel image classification technique for detecting multiple objects (called subobjects) in a single image. In addition to image classifiers, we apply spatial relationships among the subobjects to verify and to predict locations of detected and undetected subobjects, respectively. By continuously refining the spatial relationships throughout the detection process, even locations of completely occluded exhibits can be determined. Finally, all detected subobjects are labeled and the user can select the object of interest for retrieving corresponding multimedia information. This approach is applied in the context of PhoneGuide, an adaptive museum guidance system for camera-equipped mobile phones. We show that the recognition of subobjects using spatial relationships is up to 68% faster than related approaches without spatial relationships. Results of a field experiment in a local museum illustrate that unexperienced users reach an average recognition rate for subobjects of 85.6% under realistic conditions.
Das Ziel dieser Arbeit war es, durch Verwendung geeigneter vorhandener CAD-Pläne die Bearbeitung neuer CAD-Pläne zu unterstützen. Entstanden ist ein generischer Ansatz zum fallbasierten Schließens. Da in CAD-Plänen die räumliche Struktur eine wichtige Rolle spielt, ist das Konzept auf strukturorientierte Anwendungen ausgerichtet. Deshalb bezeichne ich es als ein Konzept zum " strukturorientierten fallbasierten Schließen". Die Arbeit spezifiziert das Minimum an Wissen, welches zur Suche und Wiederverwendung von Fällen benötigt wird, wie das darüber hinausgehende Wissen verarbeitet wird, welche Zusammenhänge es zum Beispiel zwischen Vergleichs- und Anpassungswissen gibt und wie man das Wissen modellieren kann. Zur Erläuterung wird das benötigte Wissen anhand verschiedener Anwendungen dargestellt. Das in der Arbeit vorgestellte Konzept erlaubt die Ergänzung, Detaillierung und Korrektur einer Anfrage. Die beiden entscheidenden Algorithmen dienen dem Vergleich von Anfrage und Fall und der Anpassung der Information des Falles zur Modifikation der Anfrage.
In today’s information society the vast technical progress and the sinking cost of information and communication technology provide new opportunities for information supply, and new technical support for communication and cooperation over distance. These trends also entail challenges such as supplying information that is adequate for a particular person in a specific situation as well as managing communication among geographically distributed parties efficiently. Context-aware systems that use sensors in order to analyse their environment and to adapt their behaviour. Yet, adequate tools for developing sensor-based infrastructures are missing. We have designed and developed Sens-ation, an open and generic service-oriented platform, which provides powerful, yet easy-to-use, tools to software developers who want to develop context-aware, sensor-based infrastructures. The service-oriented paradigm of Sens-ation enables standardised communication within individual infrastructures, between infrastructures and their sensors, but also among distributed infrastructures. On a whole, Sens-ation facilitates the development allowing developers to concentrate on the semantics of their infrastructures, and to develop innovative concepts and implementations of context-aware systems.
Die Verbreitung mobiler Smartphones und besonders deren allgegenwärtige Lokalisierungstechnologien verändern das Navigationsverhalten im Raum nachhaltig. Parallel zur schnell voranschreitenden Entwicklung alltäglicher Geräte, die mitgeführt werden, setzt der Übergang der bereits länger dauernden Entwicklung von Virtual-Reality-Technik in eine erweiterte und augmentierte Mixed Reality ein. In diesem Spannungsfeld untersucht die vorliegende Arbeit, inwieweit richtungsgebundene und binaural wiedergegebene Stereofonie die menschliche Bewegung im Raum beeinflussen kann und versucht zu erörtern, welche Potenziale in der Wiederentdeckung einer relativ lange bekannten Technik liegen. Der Autor hat im Rahmen dieser Arbeit eine binaurale mobile Applikation für richtungsgebundene Stereofonie entwickelt, mit der virtuelle bewegte oder statische Audio-Hotspots im Raum platziert werden können. So kann links, rechts oder 30 Meter vor einer Person ein virtueller oder tatsächlicher Klang im Raum verortet sein. Durch die in Echtzeit berechnete binaurale Wiedergabe der Klangquellen mit einem Stereo-Kopfhörer können diese räumlich verorteten Klänge mit zwei Ohren dreidimensional wahrgenommen werden, ähnlich dem räumlichen Sehen mit zwei Augen. Durch den Einsatz mehrerer lokalisierter Klangquellen als Soundscape entsteht eine augmentierte auditive Realität, die die physische Realität erweitert. Die Position und Navigation des Nutzers wird durch binaurale Lautstärkenmodulation (die Lautstärke nimmt bei abnehmender Distanz zur Quelle zu) und Stereopanning mit Laufzeitmodulation (die Richtung wird über ein Stereosignal auf beiden Ohren räumlich links-rechts-vorne verortet) interaktiv und kybernetisch beeinflusst. Die Nutzer navigieren — durch ihr Interesse an den hörbaren virtuellen Klangquellen geleitet — durch einen dynamisch erzeugten, dreidimensionalen akustischen Raum, der gleichzeitig ein virtueller und kybernetischer Raum ist, da die Repräsentation der Klänge an die Bewegung und Ausrichtung der Nutzer im Raum angepasst wird. Diese Arbeit untersucht, ob die Bewegung von Menschen durch (virtuelle) Klänge beeinflusst werden kann und wie groß oder messbar dieser Einfluss ist. Dabei können nicht alle künstlerischen, architektonischen und philosophischen Fragen im Rahmen der vorliegenden Schrift erörtert werden, obwohl sie dennoch als raumtheoretische Fragestellung von Interesse sind. Hauptgegenstand der vorliegenden Arbeit liegt in der Erforschung, ob richtungsgebundene Stereofonie einen relevanten Beitrag zur menschlichen Navigation, hauptsächlich zu Fuß, in urbanen Gebieten — vorwiegend im Außenraum — leisten kann. Der erste Teil gliedert sich in »Raum und Klang«, es werden raumtheoretische Überlegungen zur menschlichen Bewegung im Raum, Raumvorstellungen, räumliche Klänge und Klangwahrnehmung sowie die Entwicklung stereofoner Apparaturen und Aspekte der Augmented Audio Reality besprochen. Im zweiten Teil werden drei Demonstratoren als Anwendungsszenarien und drei Evaluierungen im Außenraum vorgestellt. Die Tests untersuchen, ob sich das Verfahren zur Navigation für Fußgänger eignet und inwieweit eine Einflussnahme auf das Bewegungsverhalten von Nutzern getroffen werden kann. Die Auswertungen der Tests zeigen, dass sich stereofone Klänge grundsätzlich als Navigationssystem eignen, da eine große Mehrzahl der Teilnehmer die akustisch markierten Ziele leicht gefunden hat. Ebenso zeigt sich ein klarer Einfluss auf die Bewegungsmuster, allerdings ist dieser abhängig von individuellen Interessen und Vorlieben. Abschließend werden die Ergebnisse der Untersuchungen im Kontext der vorgestellten Theorien diskutiert und die Potenziale stereofoner Anwendungen in einem Ausblick behandelt. Bei der Gestaltung, Erzeugung und Anwendung mobiler Systeme sind unterschiedliche mentale und räumliche Modelle und Vorstellungen der Entwickler und Anwender zu beachten. Da eine umfassende transdisziplinäre Betrachtung klare Begrifflichkeiten erfordert, werden Argumente für ein raumtheoretisches Vokabular diskutiert. Diese sind für einen gestalterischen Einsatz von richtungsgebundener Stereofonie — besonders im Kontext mobiler Navigation durch akustisch augmentierte Räume — äußerst relevant.
The task-based view of web search implies that retrieval should take the user perspective into account. Going beyond merely retrieving the most relevant result set for the current query, the retrieval system should aim to surface results that are actually useful to the task that motivated the query.
This dissertation explores how retrieval systems can better understand and support their users’ tasks from three main angles: First, we study and quantify search engine user behavior during complex writing tasks, and how task success and behavior are associated in such settings. Second, we investigate search engine queries formulated as questions, and explore patterns in a large query log that may help search engines to better support this increasingly prevalent interaction pattern. Third, we propose a novel approach to reranking the search result lists produced by web search engines, taking into account retrieval axioms that formally specify properties of a good ranking.
This study demonstrates the application and combination of multiple imaging techniques [light microscopy, micro-X-ray computer tomography (μ-CT), scanning electron microscopy (SEM) and focussed ion beam – nano-tomography (FIB-nT)] to the analysis of the microstructure of hydrated alite across multiple scales. However, by comparing findings with mercury intrusion porosimetry (MIP), it becomes obvious that the imaged 3D volumes and 2D images do not sufficiently overlap at certain scales to allow a continuous quantification of the pore size distribution (PSD). This can be overcome by improving the resolution and increasing the measured volume. Furthermore, results show that the fibrous morphology of calcium-silicate-hydrates (C-S-H) phases is preserved during FIB-nT. This is a requirement for characterisation of nano-scale porosity. Finally, it was proven that the combination of FIB-nT with energy-dispersive X-ray spectroscopy (EDX) data facilitates the phase segmentation of a 11 × 11 × 7.7 μm3 volume of hydrated alite.
Recent radiometric compensation techniques make it possible to project images onto colored and textured surfaces. This is realized with projector-camera systems by scanning the projection surface on a per-pixel basis. With the captured information, a compensation image is calculated that neutralizes geometric distortions and color blending caused by the underlying surface. As a result, the brightness and the contrast of the input image is reduced compared to a conventional projection onto a white canvas. If the input image is not manipulated in its intensities, the compensation image can contain values that are outside the dynamic range of the projector. They will lead to clipping errors and to visible artifacts on the surface. In this article, we present a novel algorithm that dynamically adjusts the content of the input images before radiometric compensation is carried out. This reduces the perceived visual artifacts while simultaneously preserving a maximum of luminance and contrast. The algorithm is implemented entirely on the GPU and is the first of its kind to run in real-time.
Projector-based displays have been evolving tremendously in the last decade. Reduced costs and increasing capabilities have let to a widespread use for home entertainment and scientific visualization. The rapid development is continuing - techniques that allow seamless projection onto complex everyday environments such as textured walls, window curtains or bookshelfs have recently been proposed. Although cameras enable a completely automatic calibration of the systems, all previously described techniques rely on a precise mapping between projector and camera pixels. Global illumination effects such as reflections, refractions, scattering, dispersion etc. are completely ignored since only direct illumination is taken into account. We propose a novel method that applies the light transport matrix for performing an image-based radiometric compensation which accounts for all possible lighting effects. For practical application the matrix is decomposed into clusters of mutually influencing projector and camera pixels. The compensation is modeled as a linear equation system that can be solved separately for each cluster. For interactive compensation rates this model is adapted to enable an efficient implementation on programmable graphics hardware. Applying the light transport matrix's pseudo-inverse allows to separate the compensation into a computational expensive preprocessing step (computing the pseudo-inverse) and an on-line matrix-vector multiplication. The generalized mathematical foundation for radiometric compensation with projector-camera systems is validated with several experiments. We show that it is possible to project corrected imagery onto complex surfaces such as an inter-reflecting statuette and glass. The overall sharpness of defocused projections is increased as well. Using the proposed optimization for GPUs, real-time framerates are achieved.
Projector-Based Augmentation
(2006)
Projector-based augmentation approaches hold the potential of combining the advantages of well-establishes spatial virtual reality and spatial augmented reality. Immersive, semi-immersive and augmented visualizations can be realized in everyday environments – without the need for special projection screens and dedicated display configurations. Limitations of mobile devices, such as low resolution and small field of view, focus constrains, and ergonomic issues can be overcome in many cases by the utilization of projection technology. Thus, applications that do not require mobility can benefit from efficient spatial augmentations. Examples range from edutainment in museums (such as storytelling projections onto natural stone walls in historical buildings) to architectural visualizations (such as augmentations of complex illumination simulations or modified surface materials in real building structures). This chapter describes projector-camera methods and multi-projector techniques that aim at correcting geometric aberrations, compensating local and global radiometric effects, and improving focus properties of images projected onto everyday surfaces.
Presence, Privacy, and PRIMIFaces: Towards Selective Information Disclosure in Instant Messaging
(2008)
Efficient distant cooperation often requires spontaneous ad-hoc social interaction, which is only possible with adequate information on the prospective communication partner. This often requires disclosing and sharing personal information via tools such as instant messaging systems and can conflict with the users’ wishes for privacy. In this paper we present an initial study investigating this trade-off and discuss implications for the design of instant messaging systems. We present the functionality and design of the PRIMIFaces instant messaging prototype supporting flexible identity management and selective information disclosure.
Few studies have investigated how search behavior affects complex writing tasks. We analyze a dataset of 150 long essays whose authors searched the ClueWeb09 corpus for source material, while all querying, clicking, and writing activity was meticulously recorded. We model the effect of search and writing behavior on essay quality using path analysis. Since the boil-down and build-up writing strategies identified in previous research have been found to affect search behavior, we model each writing strategy separately. Our analysis shows that the search process contributes significantly to essay quality through both direct and mediated effects, while the author's writing strategy moderates this relationship. Our models explain 25–35% of the variation in essay quality through rather simple search and writing process characteristics alone, a fact that has implications on how search engines could personalize result pages for writing tasks. Authors' writing strategies and associated searching patterns differ, producing differences in essay quality. In a nutshell: essay quality improves if search and writing strategies harmonize—build-up writers benefit from focused, in-depth querying, while boil-down writers fare better with a broader and shallower querying strategy.
We present PhoneGuide – an enhanced museum guidance approach that uses camera-equipped mobile phones and on-device object recognition. Our main technical achievement is a simple and light-weight object recognition approach that is realized with single-layer perceptron neuronal networks. In contrast to related systems which perform computational intensive image processing tasks on remote servers, our intention is to carry out all computations directly on the phone. This ensures little or even no network traffic and consequently decreases cost for online times. Our laboratory experiments and field surveys have shown that photographed museum exhibits can be recognized with a probability of over 90%. We have evaluated different feature sets to optimize the recognition rate and performance. Our experiments revealed that normalized color features are most effective for our method. Choosing such a feature set allows recognizing an object below one second on up-to-date phones. The amount of data that is required for differentiating 50 objects from multiple perspectives is less than 6KBytes.
In this paper, we present a novel technique for adapting local image classifiers that are applied for object recognition on mobile phones through ad-hoc network communication between the devices. By continuously accumulating and exchanging collected user feedback among devices that are located within signal range, we show that our approach improves the overall classification rate and adapts to dynamic changes quickly. This technique is applied in the context of PhoneGuide – a mobile phone based museum guidance framework that combines pervasive tracking and local object recognition for identifying a large number of objects in uncontrolled museum environments.
Compiling and disseminating information about incidents and disasters are key to disaster management and relief. But due to inherent limitations of the acquisition process, the required information is often incomplete or missing altogether. To fill these gaps, citizen observations spread through social media are widely considered to be a promising source of relevant information, and many studies propose new methods to tap this resource. Yet, the overarching question of whether and under which circumstances social media can supply relevant information (both qualitatively and quantitatively) still remains unanswered. To shed some light on this question, we review 37 disaster and incident databases covering 27 incident types, compile a unified overview of the contained data and their collection processes, and identify the missing or incomplete information. The resulting data collection reveals six major use cases for social media analysis in incident data collection: (1) impact assessment and verification of model predictions, (2) narrative generation, (3) recruiting citizen volunteers, (4) supporting weakly institutionalized areas, (5) narrowing surveillance areas, and (6) reporting triggers for periodical surveillance. Furthermore, we discuss the benefits and shortcomings of using social media data for closing information gaps related to incidents and disasters.
Multi-Frame Rate Rendering
(2008)
Multi-frame rate rendering is a parallel rendering technique that renders interactive parts of a scene on one graphics card while the rest of the scene is rendered asynchronously on a second graphics card. The resulting color and depth images of both render processes are composited, by optical superposition or digital composition, and displayed. The results of a user study confirm that multi-frame rate rendering can significantly improve the interaction performance. Multi-frame rate rendering is naturally implemented on a graphics cluster. With the recent availability of multiple graphics cards in standalone systems the method can also be implemented on a single computer system where memory bandwidth is much higher compared to off-the-shelf networking technology. This decreases overall latency and further improves interactivity. Multi-frame rate rendering was also investigated on a single graphics processor by interleaving the rendering streams for the interactive elements and the rest of the scene. This approach enables the use of multi-frame rate rendering on low-end graphics systems such as laptops, mobile phones, and PDAs. Advanced multi-frame rate rendering techniques reduce the limitations of the basic approach. The interactive manipulation of light sources and their parameters affects the entire scene. A multi-GPU deferred shading method is presented that splits the rendering task into a rasterization and lighting pass and assigns the passes to the appropriate image generators such that light manipulations at high frame rates become possible. A parallel volume rendering technique allows the manipulation of objects inside a translucent volume at high frame rates. This approach is useful for example in medical applications, where small probes need to be positioned inside a computed-tomography image. Due to the asynchronous nature of multi-frame rate rendering artifacts may occur during migration of objects from the slow to the fast graphics card, and vice versa. Proper state management allows to almost completely avoid these artifacts. Multi-frame rate rendering significantly improves the interactive manipulation of objects and lighting effects. This leads to a considerable increase of the size for 3D scenes that can be manipulated compared to conventional methods.
We present a novel multi-step technique for imperceptible geometry and radiometry calibration of projector-camera systems. Our approach can be used to display geometry and color corrected images on non-optimized surfaces at interactive rates while simultaneously performing a series of invisible structured light projections during runtime. It supports disjoint projector-camera configurations, fast and progressive improvements, as well as real-time correction rates of arbitrary graphical content. The calibration is automatically triggered when mis-registrations between camera, projector and surface are detected.
Among all imaging techniques that have been invented throughout the last decades, computer graphics is one of the most successful tools today. Many areas in science, entertainment, education, and engineering would be unimaginable without the aid of 2D or 3D computer graphics. The reason for this success story might be its interactivity, which is an important property that is still not provided efficiently by competing technologies – such as holography. While optical holography and digital holography are limited to presenting a non-interactive content, electroholography or computer generated holograms (CGH) facilitate the computer-based generation and display of holograms at interactive rates [2,3,29,30]. Holographic fringes can be computed by either rendering multiple perspective images, then combining them into a stereogram [4], or simulating the optical interference and calculating the interference pattern [5]. Once computed, such a system dynamically visualizes the fringes with a holographic display. Since creating an electrohologram requires processing, transmitting, and storing a massive amount of data, today’s computer technology still sets the limits for electroholography. To overcome some of these performance issues, advanced reduction and compression methods have been developed that create truly interactive electroholograms. Unfortunately, most of these holograms are relatively small, low resolution, and cover only a small color spectrum. However, recent advances in consumer graphics hardware may reveal potential acceleration possibilities that can overcome these limitations [6]. In parallel to the development of computer graphics and despite their non-interactivity, optical and digital holography have created new fields, including interferometry, copy protection, data storage, holographic optical elements, and display holograms. Especially display holography has conquered several application domains. Museum exhibits often use optical holograms because they can present 3D objects with almost no loss in visual quality. In contrast to most stereoscopic or autostereoscopic graphics displays, holographic images can provide all depth cues—perspective, binocular disparity, motion parallax, convergence, and accommodation—and theoretically can be viewed simultaneously from an unlimited number of positions. Displaying artifacts virtually removes the need to build physical replicas of the original objects. In addition, optical holograms can be used to make engineering, medical, dental, archaeological, and other recordings—for teaching, training, experimentation and documentation. Archaeologists, for example, use optical holograms to archive and investigate ancient artifacts [7,8]. Scientists can use hologram copies to perform their research without having access to the original artifacts or settling for inaccurate replicas. Optical holograms can store a massive amount of information on a thin holographic emulsion. This technology can record and reconstruct a 3D scene with almost no loss in quality. Natural color holographic silver halide emulsion with grain sizes of 8nm is today’s state-of-the-art [14]. Today, computer graphics and raster displays offer a megapixel resolution and the interactive rendering of megabytes of data. Optical holograms, however, provide a terapixel resolution and are able to present an information content in the range of terabytes in real-time. Both are dimensions that will not be reached by computer graphics and conventional displays within the next years – even if Moore’s law proves to hold in future. Obviously, one has to make a decision between interactivity and quality when choosing a display technology for a particular application. While some applications require high visual realism and real-time presentation (that cannot be provided by computer graphics), others depend on user interaction (which is not possible with optical and digital holograms). Consequently, holography and computer graphics are being used as tools to solve individual research, engineering, and presentation problems within several domains. Up until today, however, these tools have been applied separately. The intention of the project which is summarized in this chapter is to combine both technologies to create a powerful tool for science, industry and education. This has been referred to as HoloGraphics. Several possibilities have been investigated that allow merging computer generated graphics and holograms [1]. The goal is to combine the advantages of conventional holograms (i.e. extremely high visual quality and realism, support for all depth queues and for multiple observers at no computational cost, space efficiency, etc.) with the advantages of today’s computer graphics capabilities (i.e. interactivity, real-time rendering, simulation and animation, stereoscopic and autostereoscopic presentation, etc.). The results of these investigations are presented in this chapter.
With the growing importance of the World Wide Web, the major challenges our society faces are also increasingly affecting the digital areas of our lives. Some of the associated problems can be addressed by computer science, and some of these specifically by data-driven research. To do so, however, requires to solve open issues related to archive quality and the large volume and variety of the data contained.
This dissertation contributes data, algorithms, and concepts towards leveraging the big data and temporal provenance capabilities of web archives to tackle societal challenges. We selected three such challenges that highlight the central issues of archive quality, data volume, and data variety, respectively:
(1) For the preservation of digital culture, this thesis investigates and improves the automatic quality assurance of the web page archiving process, as well as the further processing of the resulting archive data for automatic analysis.
(2) For the critical assessment of information, this thesis examines large datasets of Wikipedia and news articles and presents new methods for automatically determining quality and bias.
(3) For digital security and privacy, this thesis exploits the variety of content on the web to quantify the security of mnemonic passwords and analyzes the privacy-aware re-finding of the various seen content through private web archives.
Multi-user virtual reality systems enable collocated as well as distributed users to perform collaborative activities in immersive virtual environments. A common activity in this context is to move from one location to the next as a group to explore the environment together. The simplest solution to realize these multi-user navigation processes is to provide each participant with a technique for individual navigation. However, this approach entails some potentially undesirable consequences such as the execution of a similar navigation sequence by each participant, a regular need for coordination within the group, and, related to this, the risk of losing each other during the navigation process.
To overcome these issues, this thesis performs research on group navigation techniques that move group members together through a virtual environment. The presented work was guided by four overarching research questions that address the quality requirements for group navigation techniques, the differences between collocated and distributed settings, the scalability of group navigation, and the suitability of individual and group navigation for various scenarios. This thesis approaches these questions by introducing a general conceptual framework as well as the specification of central requirements for the design of group navigation techniques. The design, implementation, and evaluation of corresponding group navigation techniques demonstrate the applicability of the proposed framework.
As a first step, this thesis presents ideas for the extension of the short-range teleportation metaphor, also termed jumping, for multiple users. It derives general quality requirements for the comprehensibility of the group jumping process and introduces a corresponding technique for two collocated users. The results of two user studies indicate that sickness symptoms are not affected by user roles during group jumping and confirm improved planning accuracy for the navigator, increased spatial awareness for the passenger, and reduced cognitive load for both user roles.
Next, this thesis explores the design space of group navigation techniques in distributed virtual environments. It presents a conceptual framework to systematize the design decisions for group navigation techniques based on Tuckman's model of small-group development and introduces the idea of virtual formation adjustments as part of the navigation process. A quantitative user study demonstrates that the corresponding extension of Multi-Ray Jumping for distributed dyads leads to more efficient travel sequences and reduced workload. The results of a qualitative expert review confirm these findings and provide further insights regarding the complementarity of individual and group navigation in distributed virtual environments.
Then, this thesis investigates the navigation of larger groups of distributed users in the context of guided museum tours and establishes three central requirements for (scalable) group navigation techniques. These should foster the awareness of ongoing navigation activities as well as facilitate the predictability of their consequences for all group members (Comprehensibility), assist the group with avoiding collisions in the virtual environment (Obstacle Avoidance), and support placing the group in a meaningful spatial formation for the joint observation and discussion of objects (View Optimization). The work suggests a new technique to address these requirements and reports on its evaluation in an initial usability study with groups of five to ten (partially simulated) users. The results indicate easy learnability for navigators and high comprehensibility for passengers. Moreover, they also provide valuable insights for the development of group navigation techniques for even larger groups.
Finally, this thesis embeds the previous contributions in a comprehensive literature overview and emphasizes the need to study larger, more heterogeneous, and more diverse group compositions including the related social factors that affect group dynamics.
In summary, the four major research contributions of this thesis are as follows:
- the framing of group navigation as a specific instance of Tuckman's model of small-group development
- the derivation of central requirements for effective group navigation techniques beyond common quality factors known from single-user navigation
- the introduction of virtual formation adjustments during group navigation and their integration into concrete group navigation techniques
- evidence that appropriate pre-travel information and virtual formation adjustments lead to more efficient travel sequences for groups and lower workloads for both navigators and passengers
Overall, the research of this thesis confirms that group navigation techniques are a valuable addition to the portfolio of interaction techniques in multi-user virtual reality systems. The conceptual framework, the derived quality requirements, and the development of novel group navigation techniques provide effective guidance for application developers and inform future research in this area.
Im Rahmen der Arbeit wird untersucht, welche Awarenessinformationen in kooperativen Situation benötigt werden. Herangezogen wird dazu das Denver Modell von Salvador et al., das fünf Dimensionen bereit stellt, anhand derer jedwede kooperative Situation klassifiziert werden kann: Abhängigkeit, Zeit, Gruppengröße, Ort und Timing. Bei der untersuchten Situation handelt es sich um eine eng gekoppelte, synchrone, örtlich verteilte geplante Kooperationen in einer Kleingruppe. Als konkrete Ausprägung der so bezeichneten Modellinstanz wird das Chat-basierte Rollenspiel betrachtet. In der Arbeit wird untersucht, welche Awarenessinformationen in einer so charakterisierten Situation benötigt werden und wie diese Awarenessinformationen benutzerfreundlich dargestellt werden können, so dass sie den Benutzer dabei unterstützen, sein kooperatives Ziel bzw. seine Teilziele zu erreichen. Dazu wurde eine Analyse des situativen Bedarfs durchgeführt und tragende Awarenessinformationssäulen identifiziert. Um zu Richtlinien zur Darstellung dieser benötigten Awarenessinformationen zu gelangen, wurden Annahmen aufgestellt, zur Überprüfung eine Studie konzipiert und mit drei Designvarianten durchgeführt. Aus der Datenanalyse wurden Schlussfolgerungen gezogen. Für den erstellten Katalog mit Gestaltungsrichtlinien wurden neben diesen Schlussfolgerungen auch Normen und andere Empirie (Vorerfahrung, Literatur) berücksichtigt. Die eingesetzte Evaluationsform hat sich in der Studie bewährt und sie kann für zukünftige Studien verwendet werden. Auch die Anpassung klassischer Usability-Kriterien für die Präsentation von Awarenessinformationen hat sich bewährt. Durch die Analyse des spezifischen Bedarfs in einer kooperativen Situation wird gewährleistet, dass die Benutzer die Unterstützung erhalten, die erforderlich ist, um die Aufgabe effektiv und effizient bearbeiten zu können, ohne dabei mit unnötigen Informationen überlastet zu werden. Durch Bereitstellung empirisch belegter und theoretisch fundierter Gestaltungsrichtlinien schafft diese Arbeit eine Basis für die zukünftige Entwicklung von Groupwareanwendungen, insbesondere solcher, die Werkzeuge für die Computer-vermittelte, synchrone, eng gekoppelte, geplante Kooperationen in Kleingruppen bereitstellen. Die Arbeit liefert wichtige Erkenntnisse, wie Awarenessinformationen dargestellt werden sollen, d.h. in welcher Platzierung bzw. Gruppierung und in welcher Form, so dass der Benutzer alle wesentlichen Awarenessinformationen schnell entdecken und interpretieren kann, ohne dabei überfordert oder von der Primäraufgabe abgelenkt zu werden.
CAMShift is a well-established and fundamental algorithm for kernel-based visual object tracking. While it performs well with objects that have a simple and constant appearance, it is not robust in more complex cases. As it solely relies on back projected probabilities it can fail in cases when the object's appearance changes (e.g. due to object or camera movement, or due to lighting changes), when similarly colored objects have to be re-detected or when they cross their trajectories. We propose extensions to CAMShift that address and resolve all of these problems. They allow the accumulation of multiple histograms to model more complex object appearance and the continuous monitoring of object identi- ties to handle ambiguous cases of partial or full occlusion. Most steps of our method are carried out on the GPU for achieving real-time tracking of multiple targets simultaneously. We explain an ecient GPU implementations of histogram generation, probability back projection, im- age moments computations, and histogram intersection. All of these techniques make full use of a GPU's high parallelization.
Die heute erhältlichen Web-Content Management-Systeme (WCMS) verfügen über ein umfangreiches und breit gefächertes Angebot an Funktionen, die weit über die, zur Redaktion und zum Management von Internetpräsentationen, not-wendigen Grundanforderungen hinausgehen. Das macht diese Systeme in ih-ren Einsatz sehr flexibel und deckt vielfältige Anforderungen der Endanwender ab. Andererseits steigt durch die dadurch bedingte Komplexität der Arbeitsauf-wand erheblich und die Bedien- und Benutzerfreundlichkeit sinkt. Gerade für kleinere Internetpräsentationen, die ohne aufwendige Interaktionsmöglichkeiten aber auf häufig wechselndem Informationsangeboten aufwarten, wäre dies in seiner Grundfunktionalität reduziertes System vorteilhaft. Ein solches reduziertes Web-Content Management-System soll während der Diplomarbeit entworfen und beispielhaft implementiert werden. Als Ausgangs- und Orientierungspunkt soll hierzu die Internetpräsentation der Professur Informations- und Wissensverarbeitung dienen. Zur softwaretechnischen Umsetzung sind PHP und MySQL in Verbindung mit regulären HTML und CSS zu be-nutzen. Für das weitere Vorgehen müssen zunächst die Struktur und der Aufbau der Internetpräsentation der Professur analysiert, strukturiert und formalisiert werden. Anschließend sind die am häufigsten professionell genutzten Webcontent-Managementsysteme (TYPO3 und weitere siehe www.opensourcecms.com) hinsichtlich der durch sie angebotenen Grundfunktionalitäten und der verwen-deten Templates und Vorlagen zu untersuchen. Die aus dieser Analyse resultierenden Ergebnisse sind Ausgangspunkt für die Anforderungsdefinition des zu erstellenden Mini-WCMS. Anschließend ist eine prototypische Implementierung des theoretisch entstan-denen Systems, zugeschnitten auf die speziellen Bedürfnisse der Professur, vorzunehmen und hinsichtlich seiner Eignung zu diskutieren.
Die zu beobachtenden kürzeren Produktlebenszyklen und eine schnellere Marktdurchdringung von Produkttechnologien erfordern adaptive und leistungsfähige Produktionsanlagen. Die Adaptivität ermöglicht eine Anpassung der Produktionsanlage an neue Produkte, und die Leistungsfähigkeit der Anlage stellt sicher, dass ausreichend Produkte in kurzer Zeit und zu geringen Kosten hergestellt werden können. Durch eine Modularisierung der Produktionsanlage kann die Adaptivität erreicht werden. Jedoch erfordert heutzutage jede Adaption manuellen Aufwand, z.B. zur Anpassung von proprietären Signalen oder zur Anpassung übergeordneter Funktionen. Dadurch sinkt die Leistungsfähigkeit der Anlage.
Das Ziel dieser Arbeit ist es, die Interoperabilität in Bezug auf die Informationsverwendung in modularen Produktionsanlagen zu gewährleisten. Dazu werden Informationen durch semantische Modelle beschrieben. Damit wird ein einheitlicher Informationszugriff ermöglicht, und übergeordnete Funktionen erhalten Zugriff auf alle Informationen der Produktionsmodule, unabhängig von dem Typ, dem Hersteller und dem Alter des Moduls. Dadurch entfällt der manuelle Aufwand bei Anpassungen des modularen Produktionssystems, wodurch die Leistungsfähigkeit der Anlage gesteigert und Stillstandszeiten reduziert werden.
Nach dem Ermitteln der Anforderungen an einen Modellierungsformalismus wurden potentielle Formalismen mit den Anforderungen abgeglichen. OWL DL stellte sich als geeigneter Formalismus heraus und wurde für die Erstellung des semantischen Modells in dieser Arbeit verwendet. Es wurde exemplarisch ein semantisches Modell für die drei Anwendungsfälle Interaktion, Orchestrierung und Diagnose erstellt. Durch einen Vergleich der Modellierungselemente von unterschiedlichen Anwendungsfällen wurde die Allgemeingültigkeit des Modells bewertet. Dabei wurde gezeigt, dass die Erreichung eines allgemeinen Modells für technische Anwendungsfälle möglich ist und lediglich einige Hundert Begriffe benötigt.
Zur Evaluierung der erstellten Modelle wurde ein wandlungsfähiges Produktionssystem der SmartFactoryOWL verwendet, an dem die Anwendungsfälle umgesetzt wurden. Dazu wurde eine Laufzeitumgebung erstellt, die die semantischen Modelle der einzelnen Module zu einem Gesamtmodell vereint, Daten aus der Anlage in das Modell überträgt und eine Schnittstelle für die Services bereitstellt. Die Services realisieren übergeordnete Funktionen und verwenden die Informationen des semantischen Modells. In allen drei Anwendungsfällen wurden die semantischen Modelle korrekt zusammengefügt und mit den darin enthaltenen Informationen konnte die Aufgabe des jeweiligen Anwendungsfalles ohne zusätzlichen manuellen Aufwand gelöst werden.
Die Arbeit beschäftigt sich mit der komplexen Aufgabe Stadtstrukturen mit einem Simulationsmodell im Rechner erzeugen zu können. Dafür wird ein Vierebenenmodell als Untermodellebene eingeführt, um mit allgemeinen Informationen, Parzellierung, Gebäuden und Optimierung der Struktur arbeiten zu können.
This thesis focuses on the cryptanalysis and the design of block ciphers and hash func- tions. The thesis starts with an overview of methods for cryptanalysis of block ciphers which are based on differential cryptanalysis. We explain these concepts and also sev- eral combinations of these attacks. We propose new attacks on reduced versions of ARIA and AES. Furthermore, we analyze the strength of the internal block ciphers of hash functions. We propose the first attacks that break the internal block ciphers of Tiger, HAS-160, and a reduced round version of SHACAL-2. The last part of the thesis is concerned with the analysis and the design of cryptographic hash functions. We adopt a block cipher attack called slide attack into the scenario of hash function cryptanalysis. We then use this new method to attack different variants of GRINDAHL and RADIOGATUN. Finally, we propose a new hash function called TWISTER which was designed and pro- posed for the SHA-3 competition. TWISTER was accepted for round one of this com- petition. Our approach follows a new strategy to design a cryptographic hash function. We also describe several attacks on TWISTER and discuss the security issues concern- ing these attack on TWISTER.
Digital storytelling of remote social interaction, where the situation of a remote group distributed over two locations is captured and a story is generated for later retrieval, can provide valuable insight into the structure and processes in a group. Yet, capturing these situations is a challenge—both from a technical perspective, and from a social perspective. In this paper we present CoLocScribe: a concept and prototype of an advanced media space featuring ubiquitous computing technology for capturing remote social interaction as well as a study of its use providing valuable feedback for the captured persons as well as input for the authors.
Early sensor-based infrastructures were often developed by experts with a thorough knowledge of base technology for sensing information, for processing the captured data, and for adapting the system’s behaviour accordingly. In this paper we argue that also end-users should be able to configure Ubiquitous Computing environments. We introduce the CollaborationBus application: a graphical editor that provides abstractions from base technology and thereby allows multifarious users to configure Ubiquitous Computing environments. By composing pipelines users can easily specify the information flows from selected sensors via optional filters for processing the sensor data to actuators changing the system behaviour according to the users’ wishes. Users can compose pipelines for both home and work environments. An integrated sharing mechanism allows them to share their own compositions, and to reuse and build upon others’ compositions. Real-time visualisations help them understand how the information flows through their pipelines. In this paper we present the concept, implementation, and early user feedback of the CollaborationBus application.
Capturing the interaction of users in a room based on real-world and electronic sensors provides valuable input for their interactive stories. However, in such complex scenarios there is a gap between the huge amount of rather fine-grained data that is captured and the story summarising and representing the most significant aspects of the interaction. In this paper we present the CollaborationBus Aqua editor that provides an easy to use graphical editor for capturing, authoring, and sharing stories based on mixed-reality scenarios.
Virtual reality systems offer substantial potential in supporting decision processes based purely on computer-based representations and simulations. The automotive industry is a prime application domain for such technology, since almost all product parts are available as three-dimensional models. The consideration of ergonomic aspects during assembly tasks, the evaluation of humanmachine interfaces in the car interior, design decision meetings as well as customer presentations serve as but a few examples, wherein the benefit of virtual reality technology is obvious. All these tasks require the involvement of a group of people with different expertises. However, current stereoscopic display systems only provide correct 3D-images for a single user, while other users see a more or less distorted virtual model. This is a major reason why these systems still face limited acceptance in the automotive industry. They need to be operated by experts, who have an advanced understanding of the particular interaction techniques and are aware of the limitations and shortcomings of virtual reality technology. The central idea of this thesis is to investigate the utility of stereoscopic multi-user systems for various stages of the car development process. Such systems provide multiple users with individual and perspectively correct stereoscopic images, which are key features and serve as the premise for the appropriate support of collaborative group processes. The focus of the research is on questions related to various aspects of collaboration in multi-viewer systems such as verbal communication, deictic reference, embodiments and collaborative interaction techniques. The results of this endeavor provide scientific evidence that multi-viewer systems improve the usability of VR-applications for various automotive scenarios, wherein co-located group discussions are necessary. The thesis identifies and discusses the requirements for these scenarios as well as the limitations of applying multi-viewer technology in this context. A particularly important gesture in real-world group discussions is referencing an object by pointing with the hand and the accuracy which can be expected in VR is made evident. A novel two-user seating buck is introduced for the evaluation of ergonomics in a car interior and the requirements on avatar representations for users sitting in a car are identified. Collaborative assembly tasks require high precision. The novel concept of a two-user prop significantly increases the quality of such a simulation in a virtual environment and allows ergonomists to study the strain on workers during an assembly sequence. These findings contribute toward an increased acceptance of VR-technology for collaborative development meetings in the automotive industry and other domains.
Coded Aperture Projection
(2008)
In computer vision, optical defocus is often described as convolution with a filter kernel that corresponds to an image of the aperture being used by the imaging device. The degree of defocus correlates to the scale of the kernel. Convolving an image with the inverse aperture kernel will digitally sharpen the image and consequently compensate optical defocus. This is referred to as deconvolution or inverse filtering. In frequency domain, the reciprocal of the filter kernel is its inverse, and deconvolution reduces to a division. Low magnitudes in the Fourier transform of the aperture image, however, lead to intensity values in spatial domain that exceed the displayable range. Therefore, the corresponding frequencies are not considered, which then results in visible ringing artifacts in the final projection. This is the main limitation of previous approaches, since in frequency domain the Gaussian PSF of spherical apertures does contain a large fraction of low Fourier magnitudes. Applying only small kernel scales will reduce the number of low Fourier magnitudes (and consequently the ringing artifacts) -- but will also lead only to minor focus improvements. To overcome this problem, we apply a coded aperture whose Fourier transform has less low magnitudes initially. Consequently, more frequencies are retained and more image details are reconstructed.
In ubiquitous environments an increasing number of sensors capture information on users and at the same time an increasing number of actuators are available to present information to users. This vast capturing of information potentially enables the system to adapt to the users. At the same time the system might violate the users' privacy by capturing information that the users do not want to share, and the system might disrupt the users by being too obtrusive in its adaptation or information supply. In this paper we present CoDaMine - a novel approach for providing users with system - generated feedback and control in ubiquitous environments giving them the freedom they need while reducing their effort. Basically, CoDaMine captures and analyses the users' online communication to learn about their social relationships in order to provide them with recommendations for inter-personal privacy and trust management.
The effective and efficient cooperation in communities and groups requires that the members of the community or group have adequate information about each other and the environment. In this paper, we outline the basic challenges of managing awareness information. We analyse the management of awareness information in face-to-face situations, and discuss challenges and requirements for the support of awareness management in distributed settings. Finally, after taking a look at related work, we present a simple, yet powerful framework for awareness management based on constraint pattern named COBRA.
Visually impaired is a common problem for human life in the world wide. The projector-based AR technique has ability to change appearance of real object, and it can help to improve visibility for visually impaired. We propose a new framework for the appearance enhancement with the projector camera system that employed model predictive controller. This framework enables arbitrary image processing such as photo-retouch software in the real world and it helps to improve visibility for visually impaired. In this article, we show the appearance enhancement result of Peli's method and Wolffshon's method for the low vision, Jefferson's method for color vision deficiencies. Through experiment results, the potential of our method to enhance the appearance for visually impaired was confirmed as same as appearance enhancement for the digital image and television viewing.
Die technische Entwicklung, insbesondere auf dem Gebiet der Digitaltechnik eröffnet heute neue und sehr weitreichende Möglichkeiten für die Automatisierung in Zweck- und Wohnbauten. Die zur Verfügung stehenden technischen Komponenten (intelligente Sensoren und Aktoren sowie ein hausinternes Netz für die Datenübertragung -Feldbus-) unterscheiden sich für diese Einsatzfälle kaum. Die Zielstellungen sind jedoch gänzlich andere. Intelligenz im Wohnbau bedeutet vor allem intelligente Alltagsbewältigung (z.B. Zeiteinsparung), Komfort und Wohlbefinden. Daß im Heimbereich nichtfunktionale Faktoren (Human Interface, Ästhetik, Preis, Attraktivität) eine große Rolle spielen, ist in das Problembewußtsein der Gerätehersteller und Käufer getreten. Im Bereich der Heimautomatisierung werden zunehmend moderne, die Möglichkeiten der konventioellen Steuerungs- und Regelungstechnik ergänzende Technologien wie Fuzzy- Steuerungen zur Optimierung der internen Arbeitsweise von Geräten eingesetzt. Die informatorische Vernetzung im Wohnbau unterstützt darüberhinaus wichtige Anliegen des Gebäudemanagements (energetische, ergonomische und ökologische Betrachtungen der Gebäudenutzung unter wirtschaftlichen Gesichtspunkten).
In a historical perspective, the relationship between digital media and the museum environment is marked by the role of museums as example use cases for the appli- cation of digital media. Today, this exceptional use as an often technology oriented application has changed and instead digital media have turned into an integral part of mediation strategies in the museum environment. Alongside with this shift not only an increasing professionalization of application development but also a grow- ing demand for new content can be observed. Comparable to its role as the main cost factor in the media industry, the production of content rises to a challenge for museums. In particular small and medium scale european museums with limited funding and an often low level of staff coverage face this new demand and strive therefore for alternative production resources. While productive user contributions can be seen as such an alternative resource, user contributions are at the same time a manifestation for a different mode of in- teracting with content. In contrast to the dominantly passive role of audiences as re- ceivers of information, productive contributions emerge as a mode of content ex- ploration and become in this regard influential for museum mediation strategies. As applications of user contributions in museums and cultural heritage are currently rather seldom, a broader perspective towards user contributions becomes necessary to understand its specific challenges, opportunities and limitations. Productive user contributions can be found in a growing number of applications on the Internet where they either complement or fully substitute corporate content production processes. While the Wikipedia1, an online encyclopedia written entirely by a group of users and open to contributions by all its users, is one of the most prominent examples for this practice, several more applications emerged or are be- ing developed. In consequence user contributions are about to become a powerful source for the production of content in digital media environments.
We present an enhancement towards adaptive video training for PhoneGuide, a digital museum guidance system for ordinary camera–equipped mobile phones. It enables museum visitors to identify exhibits by capturing photos of them. In this article, a combined solution of object recognition and pervasive tracking is extended to a client–server–system for improving data acquisition and for supporting scale–invariant object recognition.
The advent of high-performance mobile phones has opened up the opportunity to develop new context-aware applications for everyday life. In particular, applications for context-aware information retrieval in conjunction with image-based object recognition have become a focal area of recent research. In this thesis we introduce an adaptive mobile museum guidance system that allows visitors in a museum to identify exhibits by taking a picture with their mobile phone. Besides approaches to object recognition, we present different adaptation techniques that improve classification performance. After providing a comprehensive background of context-aware mobile information systems in general, we present an on-device object recognition algorithm and show how its classification performance can be improved by capturing multiple images of a single exhibit. To accomplish this, we combine the classification results of the individual pictures and consider the perspective relations among the retrieved database images. In order to identify multiple exhibits in pictures we present an approach that uses the spatial relationships among the objects in images. They make it possible to infer and validate the locations of undetected objects relative to the detected ones and additionally improve classification performance. To cope with environmental influences, we introduce an adaptation technique that establishes ad-hoc wireless networks among the visitors’ mobile devices to exchange classification data. This ensures constant classification rates under varying illumination levels and changing object placement. Finally, in addition to localization using RF-technology, we present an adaptation technique that uses user-generated spatio-temporal pathway data for person movement prediction. Based on the history of previously visited exhibits, the algorithm determines possible future locations and incorporates these predictions into the object classification process. This increases classification performance and offers benefits comparable to traditional localization approaches but without the need for additional hardware. Through multiple field studies and laboratory experiments we demonstrate the benefits of each approach and show how they influence the overall classification rate.
We propose a novel method that applies the light transport matrix for performing an image-based radiometric compensation which accounts for all possible types of light modulation. For practical application the matrix is decomposed into clusters of mutually influencing projector and camera pixels. The compensation is modeled as a linear system that can be solved with respect to the projector patterns. Precomputing the inverse light transport in combination with an efficient implementation on the GPU makes interactive compensation rates possible. Our generalized method unifies existing approaches that address individual problems. Based on examples, we show that it is possible to project corrected images onto complex surfaces such as an inter-reflecting statuette, glossy wallpaper, or through highly-refractive glass. Furthermore, we illustrate that a side-effect of our approach is an increase in the overall sharpness of defocused projections.
Multi-user projection systems provide a coherent 3D interaction space for multiple co-located users that facilitates mutual awareness, full-body interaction, and the coordination of activities. The users perceive the shared scene from their respective viewpoints and can directly interact with the 3D content.
This thesis reports on novel interaction patterns for collaborative 3D interaction for local and distributed user groups based on such multi-user projection environments. A particular focus of our developments lies in the provision of multiple independent interaction territories in our workspaces and their tight integration into collaborative workflows. The motivation for such multi-focus workspaces is grounded in research on social cooperation patterns, specifically in the requirement for supporting phases of loose and tight collaboration and the emergence of dedicated orking territories for private usage and public exchange. We realized independent interaction territories in the form of handheld virtual viewing windows and multiple co-located hardware displays in a joint workspace. They provide independent views of a shared virtual environment and serve as access points for the exploration and manipulation of the 3D content. Their tight integration into our workspace supports fluent transitions between individual work and joint user engagement. The different affordances of various displays in an exemplary workspace consisting of a large 3D wall, a 3D tabletop, and handheld virtual viewing windows, promote different usage scenarios, for instance for views from an egocentric perspective, miniature scene representations, close-up views, or storage and transfer areas. This work shows that this versatile workspace can make the cooperation of multiple people in joint tasks more effective, e.g. by parallelizing activities, distributing subtasks, and providing mutual support.
In order to create, manage, and share virtual viewing windows, this thesis presents the interaction technique of Photoportals, a tangible interface based on the metaphor of digital photography. They serve as configurable viewing territories and enable the individual examination of scene details as well as the immediate sharing of the prepared views. Photoportals are specifically designed to complement other interface facets and provide extended functionality for scene navigation, object manipulation, and for the creation of temporal recordings of activities in the virtual scene.
A further objective of this work is the realization of a coherent interaction space for direct 3D input across the independent interaction territories in multi-display setups. This requires the simultaneous consideration of user input in several potential interaction windows as well as configurable disambiguation schemes for the implicit selection of distinct interaction contexts. We generalized the required implementation structures into a high-level software pattern and demonstrated its versatility by means of various multi-context 3D interaction tools.
Additionally, this work tackles specific problems related to group navigation in multiuser projection systems. Joint navigation of a collocated group of users can lead to unintentional collisions when passing narrow scene sections. In this context, we suggest various solutions that prevent individual collisions during group navigation and discuss their effect on the perceived integrity of the travel group and the 3D scene. For collaboration scenarios involving distributed user groups, we furthermore explored different configurations for joint and individual travel.
Last but not least, this thesis provides detailed information and implementation templates for the realization of the proposed interaction techniques and collaborative workspaces in scenegraph-based VR systems. These contributions to the abstraction of specific interaction patterns, such as group navigation and multi-window interaction, facilitate their reuse in other virtual reality systems and their adaptation to further collaborative scenarios.
For efficient distant cooperation the members of workgroups need information about each other. This need for information disclosure often conflicts with the users' wishes for privacy. In the literature often reciprocity is suggested as a solution to this trade-off. Yet, this conception of reciprocity and its enforcement by systems does not match reality. In this paper we present our study's major findings investigating the role of reciprocity among which we found that participants greatly disregarded the above conception. Additionally we discuss their significant implications for the design of systems seeking to disclose personal information.