TY - CHAP A1 - Theiler, Michael A1 - Tauscher, Eike A1 - Tulke, Jan A1 - Riedel, Thomas ED - von Both, Petra ED - Koch, Volker T1 - Visualisierung von IFC-Objekten mittels Java3D T2 - Forum Bauinformatik 2009 N2 - Die Planung komplexer Bauwerke erfolgt zunehmend mit rechnergestützten Planungswerkzeugen, die den Export von Bauwerksinformationen im STEP-Format auf Grundlage der Industry Foundation Classes (IFC) ermöglichen. Durch die Verfügbarkeit dieser Schnittstelle ist es möglich, Bauwerksinformationen für eine weiterführende applikationsübergreifende Verarbeitung bereitzustellen. Ein großer Teil der bereitgestellten Informationen bezieht sich auf die geometrische Beschreibung der einzelnen Bauteile. Um den am Bauprozess Beteiligten eine optimale Auswertung und Analyse der Bauwerksinformationen zu ermöglichen, ist deren Visualisierung unumgänglich. Das IFC-Modell stellt diese Daten mit Hilfe verschiedener Geometriemodelle bereit. Der vorliegende Beitrag beschreibt die Visualisierung von IFC-Objekten mittels Java3D. Er beschränkt sich dabei auf die Darstellung von Objekten, deren Geometrie mittels Boundary Representation (Brep) oder Surface-Model-Repräsentation beschrieben wird. KW - IFC KW - Brep KW - STEP KW - Visualisierung KW - Schnittstelle Y1 - 2009 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:gbv:wim2-20130107-18149 SN - 978-3-86644-396-9 SP - 149 EP - 159 PB - Universitätsverlag CY - Karlsruhe ER - TY - THES A1 - Willenbacher, Susanne T1 - Untersuchungen zu räumlichen Benutzerschnittstellen am Beispiel der Präsentation von Stadtinformationen T1 - Examinations to spatial user-interfaces on the example of city-informationd N2 - Schwerpunkt der Arbeit ist die Auseinandersetzung mit den Möglichkeiten und Grenzen der Desktop-VR als neue Generation der Benutzerschnittstellen. Besondere Bedeutung bei dieser Art des Interface-Designs kommt den Metaphern zu. Ein großer Teil der Arbeit beschäftigt sich mit der Klassifikation, der Auswahl und dem Einsatz passender Metaphern unter Berücksichtigung der in der Applikation darzustellenden Informationsinhalte. Aus der Kombination dieser beiden Merkmale (Art der Metapher, Informationsinhalt) ergeben sich vier verschiedene virtuelle Umgebungen, deren Eigenschaften und Besonderheiten konkretisiert und an Beispielen aus dem Anwendungsgebiet der Stadtinformationssysteme vorgestellt werden. Als praktischer Untersuchungsgegenstand dient das Anwendungsgebiet der Stadtinformationssysteme. Die theoretisch basierten Erkenntnisse und Schlußfolgerungen werden durch statistische Untersuchungen, in Form von Fragebögen zu Stadtinformationssystemen, überprüft und konkretisiert. N2 - Topic of the paper is a discussion about the possibilities and boundaries of a new age in interface-design - the age of Desktop-VR interfaces. The important basis of this approach of interface-design is the use of a metaphor. One part of this paper deals with the classification of metaphors and gives a guideline which kind of metaphor fits to which kind of information / application. If you combine this two features (kind of metaphor, kind of information) you can get four different kinds of virtual environments. The features and characteristics of this four special virtual environments will be presented. Examples from the field of city-information-systems will be discussed. The field of city-information-system-application will used for a practical examine. Therefore a statistical evaluation of questionnaire about city-information-systems was realised. KW - Virtuelle Realität KW - Geoinformationssystem KW - Graphische Benutzeroberfläche KW - Benutzerorientierung KW - Metapher KW - Stadt KW - Benutzerschnittstellen KW - Stadtinformationssystem KW - Desktop-VR-Interface KW - VRML KW - City KW - Userinterface-Design KW - Metaphors KW - VRML KW - Desktop-VR-Interface Y1 - 2000 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:gbv:wim2-20040218-363 ER - TY - CHAP A1 - König, Reinhard ED - Martens, Bob ED - Wurzer, G, Gabriel ED - Grasl, Tomas ED - Lorenz, Wolfgang ED - Schaffranek, Richard T1 - CPlan: An Open Source Library for Computational Analysis and Synthesis T2 - 33rd eCAADe Conference N2 - Some caad packages offer additional support for the optimization of spatial configurations, but the possibilities for applying optimization are usually limited either by the complexity of the data model or by the constraints of the underlying caad system. Since we missed a system that allows to experiment with optimization techniques for the synthesis of spatial configurations, we developed a collection of methods over the past years. This collection is now combined in the presented open source library for computational planning synthesis, called CPlan. The aim of the library is to provide an easy to use programming framework with a flat learning curve for people with basic programming knowledge. It offers an extensible structure that allows to add new customized parts for various purposes. In this paper the existing functionality of the CPlan library is described. KW - Architektur KW - Computer KW - CAAD KW - cplan KW - CAD Y1 - 2015 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:gbv:wim2-20160118-25037 SP - 245 EP - 250 PB - Vienna University of Technology CY - Vienna ER - TY - CHAP A1 - Bimber, Oliver T1 - HOLOGRAPHICS: Combining Holograms with Interactive Computer Graphics T2 - New Directions in Holography and Speckles N2 - Among all imaging techniques that have been invented throughout the last decades, computer graphics is one of the most successful tools today. Many areas in science, entertainment, education, and engineering would be unimaginable without the aid of 2D or 3D computer graphics. The reason for this success story might be its interactivity, which is an important property that is still not provided efficiently by competing technologies – such as holography. While optical holography and digital holography are limited to presenting a non-interactive content, electroholography or computer generated holograms (CGH) facilitate the computer-based generation and display of holograms at interactive rates [2,3,29,30]. Holographic fringes can be computed by either rendering multiple perspective images, then combining them into a stereogram [4], or simulating the optical interference and calculating the interference pattern [5]. Once computed, such a system dynamically visualizes the fringes with a holographic display. Since creating an electrohologram requires processing, transmitting, and storing a massive amount of data, today’s computer technology still sets the limits for electroholography. To overcome some of these performance issues, advanced reduction and compression methods have been developed that create truly interactive electroholograms. Unfortunately, most of these holograms are relatively small, low resolution, and cover only a small color spectrum. However, recent advances in consumer graphics hardware may reveal potential acceleration possibilities that can overcome these limitations [6]. In parallel to the development of computer graphics and despite their non-interactivity, optical and digital holography have created new fields, including interferometry, copy protection, data storage, holographic optical elements, and display holograms. Especially display holography has conquered several application domains. Museum exhibits often use optical holograms because they can present 3D objects with almost no loss in visual quality. In contrast to most stereoscopic or autostereoscopic graphics displays, holographic images can provide all depth cues—perspective, binocular disparity, motion parallax, convergence, and accommodation—and theoretically can be viewed simultaneously from an unlimited number of positions. Displaying artifacts virtually removes the need to build physical replicas of the original objects. In addition, optical holograms can be used to make engineering, medical, dental, archaeological, and other recordings—for teaching, training, experimentation and documentation. Archaeologists, for example, use optical holograms to archive and investigate ancient artifacts [7,8]. Scientists can use hologram copies to perform their research without having access to the original artifacts or settling for inaccurate replicas. Optical holograms can store a massive amount of information on a thin holographic emulsion. This technology can record and reconstruct a 3D scene with almost no loss in quality. Natural color holographic silver halide emulsion with grain sizes of 8nm is today’s state-of-the-art [14]. Today, computer graphics and raster displays offer a megapixel resolution and the interactive rendering of megabytes of data. Optical holograms, however, provide a terapixel resolution and are able to present an information content in the range of terabytes in real-time. Both are dimensions that will not be reached by computer graphics and conventional displays within the next years – even if Moore’s law proves to hold in future. Obviously, one has to make a decision between interactivity and quality when choosing a display technology for a particular application. While some applications require high visual realism and real-time presentation (that cannot be provided by computer graphics), others depend on user interaction (which is not possible with optical and digital holograms). Consequently, holography and computer graphics are being used as tools to solve individual research, engineering, and presentation problems within several domains. Up until today, however, these tools have been applied separately. The intention of the project which is summarized in this chapter is to combine both technologies to create a powerful tool for science, industry and education. This has been referred to as HoloGraphics. Several possibilities have been investigated that allow merging computer generated graphics and holograms [1]. The goal is to combine the advantages of conventional holograms (i.e. extremely high visual quality and realism, support for all depth queues and for multiple observers at no computational cost, space efficiency, etc.) with the advantages of today’s computer graphics capabilities (i.e. interactivity, real-time rendering, simulation and animation, stereoscopic and autostereoscopic presentation, etc.). The results of these investigations are presented in this chapter. KW - Erweiterte Realität KW - CGI KW - Hologramm KW - Projektionsapparat KW - Rendering KW - Scanning KW - Reconstruction KW - computer grafik KW - computer graphics Y1 - 2005 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:gbv:wim2-20111215-7365 ER - TY - RPRT A1 - Bimber, Oliver A1 - Iwai, Daisuke T1 - Superimposing Dynamic Range N2 - We present a simple and cost-efficient way of extending contrast, perceived tonal resolution, and the color space of static hardcopy images, beyond the capabilities of hardcopy devices or low-dynamic range displays alone. A calibrated projector-camera system is applied for automatic registration, scanning and superimposition of hardcopies. We explain how high-dynamic range content can be split for linear devices with different capabilities, how luminance quantization can be optimized with respect to the non-linear response of the human visual system as well as for the discrete nature of the applied modulation devices; and how inverse tone-mapping can be adapted in case only untreated hardcopies and softcopies (such as regular photographs) are available. We believe that our approach has the potential to complement hardcopy-based technologies, such as X-ray prints for filmless imaging, in domains that operate with high quality static image content, like radiology and other medical fields, or astronomy. KW - Bildverarbeitung KW - CGI KW - Computergraphik KW - Kontrast KW - Projektor-Kamera Systeme KW - Hoher Dynamikumfang KW - Contrast KW - Projector-Camera Systems KW - High Dynamic Range Y1 - 2008 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:gbv:wim2-20080422-13585 ER - TY - CHAP A1 - Bimber, Oliver T1 - Projector-Based Augmentation T2 - Emerging Technologies of Augmented Reality: Interfaces & Design N2 - Projector-based augmentation approaches hold the potential of combining the advantages of well-establishes spatial virtual reality and spatial augmented reality. Immersive, semi-immersive and augmented visualizations can be realized in everyday environments – without the need for special projection screens and dedicated display configurations. Limitations of mobile devices, such as low resolution and small field of view, focus constrains, and ergonomic issues can be overcome in many cases by the utilization of projection technology. Thus, applications that do not require mobility can benefit from efficient spatial augmentations. Examples range from edutainment in museums (such as storytelling projections onto natural stone walls in historical buildings) to architectural visualizations (such as augmentations of complex illumination simulations or modified surface materials in real building structures). This chapter describes projector-camera methods and multi-projector techniques that aim at correcting geometric aberrations, compensating local and global radiometric effects, and improving focus properties of images projected onto everyday surfaces. KW - Erweiterte Realität KW - Virtuelle Realität KW - Projektionsverfahren KW - CGI KW - Bildbasiertes Rendering KW - Rendering KW - Projektor-Kamera Systeme KW - Multi-Projektor Systeme KW - projector-camera systems KW - multi-projector systems KW - spatial augmented reality Y1 - 2006 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:gbv:wim2-20111215-7353 ER - TY - JOUR A1 - Grundhöfer, Anselm A1 - Seeger, Manja A1 - Häntsch, Ferry A1 - Bimber, Oliver T1 - Coded Projection and Illumination for Television Studios N2 - We propose the application of temporally and spatially coded projection and illumination in modern television studios. In our vision, this supports ad-hoc re-illumination, automatic keying, unconstrained presentation of moderation information, camera-tracking, and scene acquisition. In this paper we show how a new adaptive imperceptible pattern projection that considers parameters of human visual perception, linked with real-time difference keying enables an in-shot optical tracking using a novel dynamic multi-resolution marker technique KW - Association for Computing Machinery / Special Interest Group on Graphics KW - CGI KW - Maschinelles Sehen KW - Virtuelle Studios KW - Erweiterte Realität KW - Kamera Tracking KW - Projektion KW - Virtual Studios KW - Augmented Reality KW - Camera Tracking KW - Projection Y1 - 2007 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:gbv:wim2-20111215-8005 ER - TY - INPR A1 - Zollmann, Stefanie A1 - Bimber, Oliver T1 - Imperceptible Calibration for Radiometric Compensation N2 - We present a novel multi-step technique for imperceptible geometry and radiometry calibration of projector-camera systems. Our approach can be used to display geometry and color corrected images on non-optimized surfaces at interactive rates while simultaneously performing a series of invisible structured light projections during runtime. It supports disjoint projector-camera configurations, fast and progressive improvements, as well as real-time correction rates of arbitrary graphical content. The calibration is automatically triggered when mis-registrations between camera, projector and surface are detected. KW - Association for Computing Machinery / Special Interest Group on Graphics KW - CGI KW - Maschinelles Sehen KW - unsichtbare Muster Projektion KW - Projektor-Kamera Systeme KW - Kalibrierung KW - Radiometrische Kompensation KW - imperceptible pattern projection KW - projector-camera systems KW - calibration KW - radiometric compensation Y1 - 2007 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:gbv:wim2-20111215-8094 ER - TY - INPR A1 - Grundhöfer, Anselm A1 - Bimber, Oliver T1 - Real-Time Adaptive Radiometric Compensation N2 - Recent radiometric compensation techniques make it possible to project images onto colored and textured surfaces. This is realized with projector-camera systems by scanning the projection surface on a per-pixel basis. With the captured information, a compensation image is calculated that neutralizes geometric distortions and color blending caused by the underlying surface. As a result, the brightness and the contrast of the input image is reduced compared to a conventional projection onto a white canvas. If the input image is not manipulated in its intensities, the compensation image can contain values that are outside the dynamic range of the projector. They will lead to clipping errors and to visible artifacts on the surface. In this article, we present a novel algorithm that dynamically adjusts the content of the input images before radiometric compensation is carried out. This reduces the perceived visual artifacts while simultaneously preserving a maximum of luminance and contrast. The algorithm is implemented entirely on the GPU and is the first of its kind to run in real-time. KW - Maschinelles Sehen KW - CGI KW - Bildbasiertes Rendering KW - Display KW - Projektionsverfahren KW - Radiometrische Kompensation KW - Projektion KW - Projekor-Kamera System KW - Bildkorrektur KW - Visuelle Wahrnehmung KW - radiometric compensation KW - projection KW - projector-camera systems KW - image correction KW - visual perception Y1 - 2006 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:gbv:wim2-20111215-7848 ER - TY - RPRT A1 - Kurz, Daniel A1 - Häntsch, Ferry A1 - Grosse, Max A1 - Schiewe, Alexander A1 - Bimber, Oliver T1 - Laser Pointer Tracking in Projector-Augmented Architectural Environments N2 - We present a system that applies a custom-built pan-tilt-zoom camera for laser-pointer tracking in arbitrary real environments. Once placed in a building environment, it carries out a fully automatic self-registration, registrations of projectors, and sampling of surface parameters, such as geometry and reflectivity. After these steps, it can be used for tracking a laser spot on the surface as well as an LED marker in 3D space, using inter-playing fisheye context and controllable detail cameras. The captured surface information can be used for masking out areas that are critical to laser-pointer tracking, and for guiding geometric and radiometric image correction techniques that enable a projector-based augmentation on arbitrary surfaces. We describe a distributed software framework that couples laser-pointer tracking for interaction, projector-based AR as well as video see-through AR for visualizations with the domain specific functionality of existing desktop tools for architectural planning, simulation and building surveying. KW - Association for Computing Machinery / Special Interest Group on Graphics KW - CGI KW - Architektur KW - Maschinelles Sehen KW - Laserpointer Tracking KW - Erweiterte Realität KW - Interaktion KW - Projektion KW - Verteilte Systeme KW - Laser Pointer Tracking KW - Augmented Reality KW - Interaction KW - Projection KW - Distributed Systems Y1 - 2007 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:gbv:wim2-20111215-8183 ER - TY - RPRT A1 - Grundhöfer, Anselm A1 - Seeger, Manja A1 - Häntsch, Ferry A1 - Bimber, Oliver T1 - Dynamic Adaptation of Projected Imperceptible Codes N2 - In this paper we present a novel adaptive imperceptible pattern projection technique that considers parameters of human visual perception. A coded image that is invisible for human observers is temporally integrated into the projected image, but can be reconstructed by a synchronized camera. The embedded code is dynamically adjusted on the fly to guarantee its non-perceivability and to adapt it to the current camera pose. Linked with real-time flash keying, for instance, this enables in-shot optical tracking using a dynamic multi-resolution marker technique. A sample prototype is realized that demonstrates the application of our method in the context of augmentations in television studios. KW - Association for Computing Machinery / Special Interest Group on Graphics KW - CGI KW - Maschinelles Sehen KW - Erweiterte Realität KW - Kamera Tracking KW - Projektion KW - Augmented Reality KW - Camera Tracking KW - Projection Y1 - 2007 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:gbv:wim2-20111215-8168 ER - TY - INPR A1 - Wetzstein, Gordon A1 - Bimber, Oliver T1 - A Generalized Approach to Radiometric N2 - We propose a novel method that applies the light transport matrix for performing an image-based radiometric compensation which accounts for all possible types of light modulation. For practical application the matrix is decomposed into clusters of mutually influencing projector and camera pixels. The compensation is modeled as a linear system that can be solved with respect to the projector patterns. Precomputing the inverse light transport in combination with an efficient implementation on the GPU makes interactive compensation rates possible. Our generalized method unifies existing approaches that address individual problems. Based on examples, we show that it is possible to project corrected images onto complex surfaces such as an inter-reflecting statuette, glossy wallpaper, or through highly-refractive glass. Furthermore, we illustrate that a side-effect of our approach is an increase in the overall sharpness of defocused projections. KW - Association for Computing Machinery / Special Interest Group on Graphics KW - CGI KW - Maschinelles Sehen Y1 - 2006 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:gbv:wim2-20111215-7625 ER - TY - RPRT A1 - Wetzstein, Gordon A1 - Bimber, Oliver T1 - Radiometric Compensation through Inverse Light Transport N2 - Radiometric compensation techniques allow seamless projections onto complex everyday surfaces. Implemented with projector-camera systems they support the presentation of visual content in situations where projection-optimized screens are not available or not desired - as in museums, historic sites, air-plane cabins, or stage performances. We propose a novel approach that employs the full light transport between a projector and a camera to account for many illumination aspects, such as interreflections, refractions and defocus. Precomputing the inverse light transport in combination with an efficient implementation on the GPU makes the real-time compensation of captured local and global light modulations possible. KW - Association for Computing Machinery / Special Interest Group on Graphics KW - CGI KW - Maschinelles Sehen KW - Projektionssystem KW - radiometrische Kompensation KW - Licht Transport KW - Projector-Camera Systems KW - Radiometric Compensation KW - Inverse Light Transport Y1 - 2007 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:gbv:wim2-20111215-8126 ER - TY - THES A1 - Salzmann, Holger T1 - Collaboration in Co-located Automotive Applications N2 - Virtual reality systems offer substantial potential in supporting decision processes based purely on computer-based representations and simulations. The automotive industry is a prime application domain for such technology, since almost all product parts are available as three-dimensional models. The consideration of ergonomic aspects during assembly tasks, the evaluation of humanmachine interfaces in the car interior, design decision meetings as well as customer presentations serve as but a few examples, wherein the benefit of virtual reality technology is obvious. All these tasks require the involvement of a group of people with different expertises. However, current stereoscopic display systems only provide correct 3D-images for a single user, while other users see a more or less distorted virtual model. This is a major reason why these systems still face limited acceptance in the automotive industry. They need to be operated by experts, who have an advanced understanding of the particular interaction techniques and are aware of the limitations and shortcomings of virtual reality technology. The central idea of this thesis is to investigate the utility of stereoscopic multi-user systems for various stages of the car development process. Such systems provide multiple users with individual and perspectively correct stereoscopic images, which are key features and serve as the premise for the appropriate support of collaborative group processes. The focus of the research is on questions related to various aspects of collaboration in multi-viewer systems such as verbal communication, deictic reference, embodiments and collaborative interaction techniques. The results of this endeavor provide scientific evidence that multi-viewer systems improve the usability of VR-applications for various automotive scenarios, wherein co-located group discussions are necessary. The thesis identifies and discusses the requirements for these scenarios as well as the limitations of applying multi-viewer technology in this context. A particularly important gesture in real-world group discussions is referencing an object by pointing with the hand and the accuracy which can be expected in VR is made evident. A novel two-user seating buck is introduced for the evaluation of ergonomics in a car interior and the requirements on avatar representations for users sitting in a car are identified. Collaborative assembly tasks require high precision. The novel concept of a two-user prop significantly increases the quality of such a simulation in a virtual environment and allows ergonomists to study the strain on workers during an assembly sequence. These findings contribute toward an increased acceptance of VR-technology for collaborative development meetings in the automotive industry and other domains. N2 - Virtual-Reality-Systeme sind ein innovatives Instrument, um mit Hilfe computerbasierter Simulationen Entscheidungsprozesse zu unterstützen. Insbesondere in der Automobilbranche spielt diese Technologie eine wichtige Rolle, da heutzutage nahezu alle Fahrzeugteile in 3D konstruiert werden. Im Entwicklungsbereich der Automobilindustrie werden Visualisierungssysteme darüber hinaus bei der Untersuchung ergonomischer Aspekte von Montagevorgängen, bei der Bewertung der Mensch-Maschine-Schnittstelle im Fahrzeuginterieur, bei Designentscheidungen sowie bei Kundenpräsentationen eingesetzt. Diese Entscheidungsrunden bedürfen der Einbindung mehrerer Experten verschiedener Fachbereiche. Derzeit verfügbare stereoskopische Visualisierungssysteme ermöglichen aber nur einem Nutzer eine korrekte Stereosicht, während sich für die anderen Teilnehmer das 3D-Modell verzerrt darstellt. Dieser Nachteil ist ein wesentlicher Grund dafür, dass derartige Systeme bisher nur begrenzt im Automobilbereich anwendbar sind. Der Fokus dieser Dissertation liegt auf der Untersuchung der Anwendbarkeit stereoskopischer Mehrbenutzer-Systeme in verschiedenen Stadien des automobilen Entwicklungsprozesses. Derartige Systeme ermöglichen mehreren Nutzern gleichzeitig eine korrekte Stereosicht, was eine wesentliche Voraussetzung für die Zusammenarbeit in einer Gruppe darstellt. Die zentralen Forschungsfragen beziehen sich dabei auf die Anforderungen von kooperativen Entscheidungsprozessen sowie den daraus resultierenden Aspekten der Interaktion wie verbale Kommunikation, Gesten sowie virtuelle Menschmodelle und Interaktionstechniken zwischen den Nutzern. Die Arbeit belegt, dass stereoskopische Mehrbenutzersysteme die Anwendbarkeit virtueller Techniken im Automobilbereich entscheidend verbessern, da sie eine natürliche Kommunikation zwischen den Nutzern fördern. So ist die Unterstützung natürlicher Gesten beispielsweise ein wichtiger Faktor und es wird dargelegt, welche Genauigkeit beim Zeigen mit der realen Hand auf virtuelle Objekte erwartet werden kann. Darüber hinaus werden Anforderungen an virtuelle Menschmodelle anhand einer Zweibenutzer-Sitzkiste identifiziert und untersucht. Diese Form der Simulation, bei der die Nutzer nebeneinander in einem Fahrzeugmodell sitzen, dient vor allem der Bewertung von Mensch-Maschine-Schnittstellen im Fahrzeuginterieur. Des Weiteren wird das neue Konzept eines Mehrbenutzer-Werkzeugs in die Arbeit mit einbezogen, da hier verdeutlicht wird wie die Simulation von Montagevorgängen in virtuellen Umgebungen mit passivem haptischem Feedback zu ergonomischen Verbesserungen entsprechender Arbeitsvorgänge in der Realität beitragen kann. Diese Konzepte veranschaulichen wie VR-Systeme zur Unterstützung kollaborativer Prozesse in der Automobilbranche und darüber hinaus eingesetzt werden können. T2 - Zusammenarbeit in virtuellen Gruppenszenarien in der automobilen Entwicklung KW - Virtuelle Realität KW - Immersion KW - Simulation KW - Computergraphik KW - Virtual Reality KW - Computer Graphics KW - Interaction Techniques KW - Collaboration Y1 - 2010 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:gbv:wim2-20100712-15102 ER - TY - THES A1 - Lux, Christopher T1 - A Data-Virtualization System for Large Model Visualization N2 - Interactive scientific visualizations are widely used for the visual exploration and examination of physical data resulting from measurements or simulations. Driven by technical advancements of data acquisition and simulation technologies, especially in the geo-scientific domain, large amounts of highly detailed subsurface data are generated. The oil and gas industry is particularly pushing such developments as hydrocarbon reservoirs are increasingly difficult to discover and exploit. Suitable visualization techniques are vital for the discovery of the reservoirs as well as their development and production. However, the ever-growing scale and complexity of geo-scientific data sets result in an expanding disparity between the size of the data and the capabilities of current computer systems with regard to limited memory and computing resources. In this thesis we present a unified out-of-core data-virtualization system supporting geo-scientific data sets consisting of multiple large seismic volumes and height-field surfaces, wherein each data set may exceed the size of the graphics memory or possibly even the main memory. Current data sets fall within the range of hundreds of gigabytes up to terabytes in size. Through the mutual utilization of memory and bandwidth resources by multiple data sets, our data-management system is able to share and balance limited system resources among different data sets. We employ multi-resolution methods based on hierarchical octree and quadtree data structures to generate level-of-detail working sets of the data stored in main memory and graphics memory for rendering. The working set generation in our system is based on a common feedback mechanism with inherent support for translucent geometric and volumetric data sets. This feedback mechanism collects information about required levels of detail during the rendering process and is capable of directly resolving data visibility without the application of any costly occlusion culling approaches. A central goal of the proposed out-of-core data management system is an effective virtualization of large data sets. Through an abstraction of the level-of-detail working sets, our system allows developers to work with extremely large data sets independent of their complex internal data representations and physical memory layouts. Based on this out-of-core data virtualization infrastructure, we present distinct rendering approaches for specific visualization problems of large geo-scientific data sets. We demonstrate the application of our data virtualization system and show how multi-resolution data can be treated exactly the same way as regular data sets during the rendering process. An efficient volume ray casting system is presented for the rendering of multiple arbitrarily overlapping multi-resolution volume data sets. Binary space-partitioning volume decomposition of the bounding boxes of the cube-shaped volumes is used to identify the overlapping and non-overlapping volume regions in order to optimize the rendering process. We further propose a ray casting-based rendering system for the visualization of geological subsurface models consisting of multiple very detailed height fields. The rendering of an entire stack of height-field surfaces is accomplished in a single rendering pass using a two-level acceleration structure, which combines a minimum-maximum quadtree for empty-space skipping and sorted lists of depth intervals to restrict ray intersection searches to relevant height fields and depth ranges. Ultimately, we present a unified rendering system for the visualization of entire geological models consisting of highly detailed stacked horizon surfaces and massive volume data. We demonstrate a single-pass ray casting approach facilitating correct visual interaction between distinct translucent model components, while increasing the rendering efficiency by reducing processing overhead of potentially invisible parts of the model. The combination of image-order rendering approaches and the level-of-detail feedback mechanism used by our out-of-core data-management system inherently accounts for occlusions of different data types without the application of costly culling techniques. The unified out-of-core data-management and virtualization infrastructure considerably facilitates the implementation of complex visualization systems. We demonstrate its applicability for the visualization of large geo-scientific data sets using output-sensitive rendering techniques. As a result, the magnitude and multitude of data sets that can be interactively visualized is significantly increased compared to existing approaches. KW - Computer Graphics KW - Visualisation KW - Volume Rendering KW - Large Data Y1 - 2013 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:gbv:wim2-20130725-19855 ER - TY - THES A1 - Springer, Jan P. T1 - Multi-Frame Rate Rendering N2 - Multi-frame rate rendering is a parallel rendering technique that renders interactive parts of a scene on one graphics card while the rest of the scene is rendered asynchronously on a second graphics card. The resulting color and depth images of both render processes are composited, by optical superposition or digital composition, and displayed. The results of a user study confirm that multi-frame rate rendering can significantly improve the interaction performance. Multi-frame rate rendering is naturally implemented on a graphics cluster. With the recent availability of multiple graphics cards in standalone systems the method can also be implemented on a single computer system where memory bandwidth is much higher compared to off-the-shelf networking technology. This decreases overall latency and further improves interactivity. Multi-frame rate rendering was also investigated on a single graphics processor by interleaving the rendering streams for the interactive elements and the rest of the scene. This approach enables the use of multi-frame rate rendering on low-end graphics systems such as laptops, mobile phones, and PDAs. Advanced multi-frame rate rendering techniques reduce the limitations of the basic approach. The interactive manipulation of light sources and their parameters affects the entire scene. A multi-GPU deferred shading method is presented that splits the rendering task into a rasterization and lighting pass and assigns the passes to the appropriate image generators such that light manipulations at high frame rates become possible. A parallel volume rendering technique allows the manipulation of objects inside a translucent volume at high frame rates. This approach is useful for example in medical applications, where small probes need to be positioned inside a computed-tomography image. Due to the asynchronous nature of multi-frame rate rendering artifacts may occur during migration of objects from the slow to the fast graphics card, and vice versa. Proper state management allows to almost completely avoid these artifacts. Multi-frame rate rendering significantly improves the interactive manipulation of objects and lighting effects. This leads to a considerable increase of the size for 3D scenes that can be manipulated compared to conventional methods. N2 - Multi-Frame Rate Rendering ist eine parallele Rendertechnik, die interaktive Teile einer Szene auf einer separaten Graphikkarte berechnet. Die Abbildung des Rests der Szene erfolgt asynchron auf einer anderen Graphikkarte. Die resultierenden Farb- und Tiefenbilder beider Darstellungsprozesse werden mittels optischer Überlagerung oder digitaler Komposition kombiniert und angezeigt. Die Ergebnisse einer Nutzerstudie zeigen, daß Multi-Frame Rate Rendering die Interaktion für große Szenen deutlich beschleunigt. Multi-Frame Rate Rendering ist üblicherweise auf einem Graphikcluster zu implementieren. Mit der Verfügbarkeit mehrerer Graphikkarten für Einzelsysteme kann Multi-Frame Rate Rendering auch für diese realisiert werden. Dies ist von Vorteil, da die Speicherbandbreite um ein Vielfaches höher ist als mit üblichen Netzwerktechnologien. Dadurch verringern sich Latenzen, was zu verbesserter Interaktivität führt. Multi-Frame Rate Rendering wurde auch auf Systemen mit einer Graphikkarte untersucht. Die Bildberechnung für den Rest der Szene muss dazu in kleine Portionen aufgeteilt werden. Die Darstellung erfolgt dann alternierend zu den interaktiven Elementen über mehrere Bilder verteilt. Dieser Ansatz erlaubt die Benutzung von Multi-Frame Rate Rendering auf einfachen Graphiksystemen wie Laptops, Mobiltelefonen and PDAs. Fortgeschrittene Multi-Frame Rate Rendering Techniken erweitern die Anwendbarkeit des Ansatzes erheblich. Die interaktive Manipulation von Lichtquellen beeinflußt die ganze Szene. Um diese Art der Interaktion zu unterstützen, wurde eine Multi-GPU Deferred Shading Methode entwickelt. Der Darstellungsvorgang wird dazu in einen Rasterisierungs- und Beleuchtungsschritt zerlegt, die parallel auf den entsprechenden Grafikkarten erfolgen können. Dadurch kann die Beleuchtung mit hohen Bildwiederholraten unabhängig von der geometrischen Komplexität der Szene erfolgen. Außerdem wurde eine parallele Darstellungstechnik für die interaktive Manipulation von Objekten in hochaufgelösten Volumendaten entwickelt. Dadurch lassen sich zum Beispiel virtuelle Instrumente in hochqualitativ dargestellten Computertomographieaufnahmen interaktiv positionieren. Aufgrund der inhärenten Asynchronität der beiden Darstellungsprozesse des Multi-Frame Rate Rendering Ansatzes können Artifakte während der Objektmigration zwischen den Graphikkarten auftreten. Eine intelligente Zustandsverwaltung in Kombination mit Prediktionstechniken kann diese Artifakte fast gänzlich verhindern, so dass Benutzer diese im allgemeinen nicht bemerken. Multi-Frame Rate Rendering beschleunigt die interaktive Manipulation von Objekten und Beleuchtungseffekten deutlich. Dadurch können deutlich umfangreichere virtuelle Szenarien bearbeitet werden als mit konventionellen Methoden. T2 - Multi-Frame Rate Rendering KW - Virtuelle Realität KW - Multi-Frame Rate Rendering KW - Multi-Frame Rate Composition KW - Interaction Fidelity KW - Visual Quality KW - Parallel Rendering Methods Y1 - 2008 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:gbv:wim2-20081127-14395 ER - TY - THES A1 - Riehmann, Patrick T1 - Advanced Visual Interfaces for Informed Decision-Making N2 - This thesis presents new interactive visualization techniques and systems intended to support users with real-world decisions such as selecting a product from a large variety of similar offerings, finding appropriate wording as a non-native speaker, and assessing an alleged case of plagiarism. The Product Explorer is a significantly improved interactive Parallel Coordinates display for facilitating the product selection process in cases where many attributes and numerous alternatives have to be considered. A novel visual representation for categorical and ordered data with only few occurring values, the so-called extended areas, in combination with cubic curves for connecting the parallel axes, are crucial for providing an effective overview of the entire dataset and to facilitate the tracing of individual products. The visual query interface supports users in quickly narrowing down the product search to a small subset or even a single product. The scalability of the approach towards a large number of attributes and products is enhanced by the possibility of setting some constraints on final attributes and, therefore, reducing the number of considered attributes and data items. Furthermore, an attribute repository allows users to focus on the most important attributes at first and to bring in additional criteria for product selection later in the decision process. A user study confirmed that the Product Explorer is indeed an excellent tool for its intended purpose for casual users. The Wordgraph is a layered graph visualization for the interactive exploration of search results for complex keywords-in-context queries. The system relies on the Netspeak web service and is designed to support non-native speakers in finding customary phrases. Uncertainties about the commonness of phrases are expressed with the help of wildcard-based queries. The visualization presents the alternatives for the wildcards in a multi-column layout: one column per wildcard with the other query fragments in between. The Wordgraph visualization displays the sorted results for all wildcards at once by appropriately arranging the words of each column. A user study confirmed that this is a significant advantage over simple textual result lists. Furthermore, visual interfaces to filter, navigate, and expand the graph allow interactive refinement and expansion of wildcard-containing queries. Furthermore, this thesis presents an advanced visual analysis tool for assessing and presenting alleged cases of plagiarism and provides a three-level approach for exploring the so-called finding spots in their context. The overview shows the relationship of the entire suspicious document to the set of source documents. An intermediate glyph-based view reveals the structural and textual differences and similarities of a set of finding spots and their corresponding source text fragments. Eventually, the actual fragments of the finding spot can be shown in a side-by-side view with a novel structured wrapping of both the source, as well as the suspicious text. The three different levels of detail are tied together by versatile navigation and selection operations. Reviews with plagiarism experts confirm that this tool can effectively support their workflow and provides a significant improvement over existing static visualizations for assessing and presenting plagiarism cases. The three main contributions of this research have a lot in common aside from being carefully designed and scientifically grounded solutions to real-world decision problems. The first two visualizations facilitate the decision for a single possibility out of many alternatives, whereas the latter ones deal with text at varying levels of detail. All visual representations are clearly structured based on horizontal and vertical layers contained in a single view and they all employ edges for depicting the most important relationships between attributes, words, or different levels of detail. A detailed analysis considering the context of the established decision-making literature reveals that important steps of common decision models are well-supported by the three visualization systems presented in this thesis. KW - Informatik KW - Visualisierung KW - Information Visualization KW - Preferential Choice KW - Text-based Visualization KW - Plagiarism Visualization KW - Product Search Y1 - 2015 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:gbv:wim2-20150907-24542 PB - Patrick Riehmann ER - TY - JOUR A1 - Kreskowski, Adrian A1 - Rendle, Gareth A1 - Fröhlich, Bernd T1 - Efficient Direct Isosurface Rasterization of Scalar Volumes JF - Computer Graphics Forum N2 - In this paper we propose a novel and efficient rasterization-based approach for direct rendering of isosurfaces. Our method exploits the capabilities of task and mesh shader pipelines to identify subvolumes containing potentially visible isosurface geometry, and to efficiently extract primitives which are consumed on the fly by the rasterizer. As a result, our approach requires little preprocessing and negligible additional memory. Direct isosurface rasterization is competitive in terms of rendering performance when compared with ray-marching-based approaches, and significantly outperforms them for increasing resolution in most situations. Since our approach is entirely rasterization based, it affords straightforward integration into existing rendering pipelines, while allowing the use of modern graphics hardware features, such as multi-view stereo for efficient rendering of stereoscopic image pairs for geometry-bound applications. Direct isosurface rasterization is suitable for applications where isosurface geometry is highly variable, such as interactive analysis scenarios for static and dynamic data sets that require frequent isovalue adjustment. KW - Rendering KW - Rastergrafik KW - Visualisierung KW - Maschinelles Sehen KW - isosurface KW - rendering KW - rasterization Y1 - 2023 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:gbv:wim2-20230525-63835 UR - https://onlinelibrary.wiley.com/doi/full/10.1111/cgf.14670 VL - 2022 IS - Volume 4, Issue 7 SP - 215 EP - 226 PB - Wiley Blackwell CY - Oxford ER - TY - THES A1 - Wetzstein, Gordon T1 - Radiometric Compensation of Global Illumination Effects with Projector-Camera Systems N2 - Projector-based displays have been evolving tremendously in the last decade. Reduced costs and increasing capabilities have let to a widespread use for home entertainment and scientific visualization. The rapid development is continuing - techniques that allow seamless projection onto complex everyday environments such as textured walls, window curtains or bookshelfs have recently been proposed. Although cameras enable a completely automatic calibration of the systems, all previously described techniques rely on a precise mapping between projector and camera pixels. Global illumination effects such as reflections, refractions, scattering, dispersion etc. are completely ignored since only direct illumination is taken into account. We propose a novel method that applies the light transport matrix for performing an image-based radiometric compensation which accounts for all possible lighting effects. For practical application the matrix is decomposed into clusters of mutually influencing projector and camera pixels. The compensation is modeled as a linear equation system that can be solved separately for each cluster. For interactive compensation rates this model is adapted to enable an efficient implementation on programmable graphics hardware. Applying the light transport matrix's pseudo-inverse allows to separate the compensation into a computational expensive preprocessing step (computing the pseudo-inverse) and an on-line matrix-vector multiplication. The generalized mathematical foundation for radiometric compensation with projector-camera systems is validated with several experiments. We show that it is possible to project corrected imagery onto complex surfaces such as an inter-reflecting statuette and glass. The overall sharpness of defocused projections is increased as well. Using the proposed optimization for GPUs, real-time framerates are achieved. KW - Association for Computing Machinery / Special Interest Group on Graphics KW - CGI KW - Maschinelles Sehen KW - Projektionssystem KW - radiometrische Kompensation KW - Licht Transport KW - Projector-Camera Systems KW - Radiometric Compensation KW - Inverse Light Transport Y1 - 2006 U6 - http://nbn-resolving.de/urn/resolver.pl?urn:nbn:de:gbv:wim2-20111215-8106 ER -