Refine
Document Type
- Conference Proceeding (132)
- Doctoral Thesis (15)
- Article (7)
- Master's Thesis (7)
- Report (4)
- Diploma Thesis (3)
- Preprint (3)
- Book (1)
- Part of a Book (1)
Institute
- In Zusammenarbeit mit der Bauhaus-Universität Weimar (93)
- Institut für Strukturmechanik (ISM) (16)
- Professur Baubetrieb und Bauverfahren (10)
- Professur Angewandte Mathematik (6)
- Institut für Europäische Urbanistik (5)
- Junior-Professur Augmented Reality (5)
- Professur Stahlbau (5)
- Institut für Konstruktiven Ingenieurbau (IKI) (4)
- Professur Informatik in der Architektur (4)
- Professur Computergestütztes kooperatives Arbeiten (3)
Keywords
- Architektur <Informatik> (125)
- CAD (125)
- Computerunterstütztes Verfahren (124)
- Baubetriebslehre (8)
- Nachtragsvereinbarung (8)
- Prozessmanagement (8)
- Weimar / Bauhaus-Universität / Professur Baubetrieb und Bauverfahren (8)
- Angewandte Mathematik (5)
- Strukturmechanik (5)
- CGI <Computergraphik> (4)
- Angewandte Informatik (3)
- Maschinelles Sehen (3)
- Association for Computing Machinery / Special Interest Group on Graphics (2)
- Bauablauf (2)
- Bildbasiertes Rendering (2)
- China (2)
- Erweiterte Realität <Informatik> (2)
- Finite-Elemente-Methode (2)
- MAG-Schweißen (2)
- Modellierung (2)
- Projektionsverfahren (2)
- Ubiquitous Computing (2)
- Virtuelle Realität (2)
- projector-camera systems (2)
- Abbinderegler (1)
- Abfälle (1)
- Alkalisulfate (1)
- Analyse (1)
- Architectural Theory (1)
- Architekturtheorie (1)
- Arthur (1)
- Augmented reality (1)
- Augmented studio (1)
- Aussagenlogik (1)
- Awareness (1)
- B-Spline (1)
- B-Spline Finite Elemente (1)
- B-spline (1)
- BIM (1)
- Bauphysik (1)
- Bausoll (1)
- Bauteil (1)
- Bearbeitung von Ingenieuraufgaben (1)
- Beteiligung (1)
- Betonverflüssiger (1)
- Bildkorrektur (1)
- Bodenaggressivität (1)
- Building Information Model (1)
- Bäderwesen (1)
- C-S-H Phasen (1)
- C-S-H phases (1)
- Cognition (1)
- Complex Systems (1)
- Computer-Supported Cooperative Work (1)
- Construction Product data (1)
- Content Management (1)
- Datenmodell (1)
- Design Science (1)
- Direkte numerische Simulation (1)
- Display (1)
- Drehrohr (1)
- Duplex-Stahl (1)
- Duplexstahl (1)
- Dynamische Analyse (1)
- Dörfer (1)
- Editor (1)
- Effizienzmessung (1)
- Eigenspannung (1)
- Elastomerlager (1)
- Epistemology (1)
- Erkenntnistheorie (1)
- Ersatzbrennstoffe (1)
- FVK (1)
- Fernsehproduktion (1)
- Festigkeit (1)
- Festkörpermechanik (1)
- Film (1)
- Fließmittel (1)
- Fließverhalten (1)
- Frankreich (1)
- Gebäudesimulation ESP-r (1)
- Gefügeumwandlung (1)
- Gegenständliches Modell (1)
- Gilles Deleuze (1)
- Globalisierung (1)
- Glocke (1)
- Glockenläuten (1)
- Gruppengewahrsein (1)
- Heuristik (1)
- Hochwasserschutz (1)
- Homogenisieren (1)
- Homogenisierung (1)
- Hydratation (1)
- Hydraulik (1)
- Hydrologische Messung (1)
- IFC (1)
- Immobilienmarkt (1)
- Informationsmodell (1)
- Instandhaltungsplanung (1)
- Inverse Light Transport (1)
- Jim Jarmusch (1)
- Kalkulation (1)
- Kognitive Komplexität (1)
- Komplexität (1)
- Konfiguration (1)
- Korrosionsgeschwindigkeit (1)
- Korrosionsprüfung (1)
- Kräfte und Bewegungen (1)
- Kunststoff (1)
- Kunststoff-Metall-Verbund (1)
- Kunststoffbauteil (1)
- Kunststoffbauten (1)
- Kunststoffherstellung (1)
- Kunststoffhäuser (1)
- Kunststoffindustrie (1)
- Kunststofftechnik (1)
- LEG (1)
- Landesentwicklungsgesellschaft (1)
- Latentwärmespeicher (1)
- Licht Transport (1)
- Lösungsverfahren (1)
- Maintenance (1)
- Maschine (1)
- Massivbrücke (1)
- Mathematisches Modell (1)
- Megastadt (1)
- Mehrgitterverfahren (1)
- Metallographie (1)
- Mikrostruktur (1)
- Mobilität (1)
- Modellbildung (1)
- Modelltechnik (1)
- Modelluntersuchung (1)
- Monitoring (1)
- Monitoring Bericht (1)
- Multi-Projektor Systeme (1)
- MySQL (1)
- Netzwerkwissenschaften (1)
- Numerische Mathematik (1)
- Nutzungsdauer (1)
- Nutzungsänderung (1)
- OPL (1)
- Opimization (1)
- Optimierung (1)
- PCM-Putz (1)
- PCM-plaster (1)
- PHP (1)
- PPP (1)
- Periodendauer (1)
- Pfeiler (1)
- Phasenübergangsmaterialien (1)
- Philosophie (1)
- Physikalisches Modell (1)
- Plausibilität (1)
- Privatsphaere (1)
- Produktinformation (1)
- Prognosemodell (1)
- Projector-Camera Systems (1)
- Projekor-Kamera System (1)
- Projektion (1)
- Projektion <Optik> (1)
- Projektionssystem (1)
- Projektor-Kamera Systeme (1)
- Public Private Partnership (1)
- Pyrolyse (1)
- RDA (1)
- RDF (1)
- Radiometric Compensation (1)
- Radiometrische Kompensation (1)
- Recycling (1)
- Rendering (1)
- Resttragfähigkeit (1)
- Revitalization (1)
- Reziprozitaet (1)
- Rohrleitung (1)
- Scale-Up (1)
- Schweißprozesssimulation (1)
- Schwimmbad (1)
- Schwingungsanregung (1)
- Schwingungsdämpfung (1)
- Schwingungstilger (1)
- Schwingungsverhalten (1)
- Schädigung (1)
- Sensor-Based Infrastructure (1)
- Sensor-basierte Infrastrukture (1)
- Service-Oriented Platform (1)
- Service-orientierte Plattform (1)
- Stadtentwicklung (1)
- Stadtforschung (1)
- Stadtplanung (1)
- Stahlbau (1)
- Stochastik (1)
- Strategic Development (1)
- Strategische Planung (1)
- Strategische Projekte (1)
- Strömungsmesstechnik (1)
- Studiotechnik (1)
- Stuttgart / Sonderforschungsbereich Rechnergestützte Modellierung und Simulation zur Analyse (1)
- Städtebau (1)
- Systemtheorie (1)
- Technik / Geschichte (1)
- Temperaturfeld (1)
- Tragkonstruktion (1)
- Tragwerk (1)
- Tragwerke (1)
- Transformation (1)
- Tsingtau (1)
- Umbau (1)
- Umnutzung (1)
- Urbanisierung (1)
- Urbanistik (1)
- Visuelle Wahrnehmung (1)
- Warschau (1)
- Wasserstandmessung (1)
- Wasserversorgung (1)
- Waste (1)
- Weimar / Sonderforschungsbereich Werkstoffe und Konstruktionen für die Revitalisierung von Bauwerken (1)
- Wiederholung (1)
- Winkelverzug (1)
- Wissenschaft / Geschichte (1)
- Wissenschaftsphilosophie (1)
- Wohnhäuser in Stahlbauweise (1)
- Wohnungsbau (1)
- Wohnungsmarkt (1)
- Wolke (1)
- Zement (1)
- Zivilgesellschaft (1)
- cement (1)
- cinema (1)
- civil society (1)
- configuration (1)
- cost estimation (1)
- digital light projection (1)
- duplex stainless steel (1)
- editor (1)
- effective properties (1)
- efficiency (1)
- engineer-technical integration (1)
- fachliche Integration (1)
- fiber reeinforced plastics (1)
- finite element (1)
- flood protection (1)
- fluidity (1)
- france (1)
- history of science (1)
- history of technology (1)
- hydration (1)
- hydraulic measurement (1)
- hydraulics (1)
- image correction (1)
- information sharing (1)
- interactive (1)
- interaktiv (1)
- job description (1)
- latent heat storage (1)
- leakages (1)
- load bearing structure (1)
- machine (1)
- megacity (1)
- mehrphasig (1)
- microstructure (1)
- modell analysis (1)
- modelling (1)
- multi-projector systems (1)
- multigrid (1)
- multiphase (1)
- numerical simulation (1)
- phase change materials (1)
- phase transformation (1)
- philosophy (1)
- plastic building (1)
- privacy (1)
- processing of engineering tasks (1)
- projection (1)
- radiometric compensation (1)
- radiometrische Kompensation (1)
- reciprocity (1)
- remaining load capacitiy (1)
- repetition (1)
- residential buildings (1)
- residual stress (1)
- sequence (1)
- set regulator (1)
- simulation (1)
- sommerlicher Wärmeschutz (1)
- spatial augmented reality (1)
- strength (1)
- structures (1)
- substitute fuel (1)
- superplasticizer (1)
- technische Nutzungsdauer (1)
- thermal building simulation (1)
- thermal protection (1)
- thermo-physical material properties (1)
- thermophysikalische Materialeigenschaften (1)
- traditional social system (1)
- traditionellen Sozialsystem (1)
- unterschiedliche Sicherheitskonzepte (1)
- urbanization (1)
- visual perception (1)
Year of publication
- 2006 (173) (remove)
A new approach to the non-linear analysis of cross-sections loaded by normal forces and bending moments is presented in the paper. The mechanical model is based on the LAGRANGE principle of minimum of total potential energy. Deformations, stresses and limit load parameters are obtained by solving a non-linear optimisation problem. The mathematical model is independent of the specifics of material. In addition to the stress strain relation and the specific strain energy W(ε) two further functions F(ε) and Φ(ε) are introduced to describe the material behaviour. Thus cracks in concrete, non-linearity of material etc. can be taken into account without basic modification of the numerical algorithm. For polygonal cross-sections the GAUSS' integral theorem is used. Numerical solutions of the non-linear optimisation problems can be found by application of standard software. Thus the analysis of reinforced concrete cross-sections or more general composite cross-sections with non-linear behaviour of material is as simple as in the case of linear elasticity. The application of the method is demonstrated for polygonal cross-sections. Pre-stresses or pre-strains can easily be included in the mathematical model.
The presented method for an physically non-linear analysis of stresses and deformations of composite cross-sections and members based on energy principles and their transformation to non-linear optimisation problems. From the LAGRANGE principle of minimum of total potential energy a kinematic formulation of the mechanical problem can be developed, which has the general advantage that pre-deformations excited by shrinkage, temperature, residual deformations after unloading et al., can be considered directly. Thus the non-linear analysis of composite cross-sections with layers of different mechanical properties and different preloading becomes possible and cracks in concrete, stiffness degradation and other specifics of the material behaviour can be taken into account without cardinal modification of the mathematical model. The impact of local defects on the bearing capacity of an entire element can also be analysed in this principle way. Standard computational systems for mathematical optimisation or general programs for spreadsheet analysis enable an uncomplicated implementation of the developed models and an effective non-linear analysis for composite cross-sections and elements.
The contribution focuses on the development of a basic computational scheme that provides a suitable calculation environment for the coupling of analytical near-field solutions with numerical standard procedures in the far-field of the singularity. The proposed calculation scheme uses classical methods of complex function theory, which can be generalized to 3-dimensional problems by using the framework of hypercomplex analysis. The adapted approach is mainly based on the factorization of the Laplace operator EMBED Equation.3 by the Cauchy-Riemann operator EMBED Equation.3 , where exact solutions of the respective differential equation are constructed by using an orthonormal basis of holomorphic and anti-holomorphic functions.
The extended finite element method (XFEM) offers an elegant tool to model material discontinuities and cracks within a regular mesh, so that the element edges do not necessarily coincide with the discontinuities. This allows the modeling of propagating cracks without the requirement to adapt the mesh incrementally. Using a regular mesh offers the advantage, that simple refinement strategies based on the quadtree data structure can be used to refine the mesh in regions, that require a high mesh density. An additional benefit of the XFEM is, that the transmission of cohesive forces through a crack can be modeled in a straightforward way without introducing additional interface elements. Finally different criteria for the determination of the crack propagation angle are investigated and applied to numerical tests of cracked concrete specimens, which are compared with experimental results.
Major problems of applying selective sensitivity to system identification are requirement of precise knowledge about the system parameters and realization of the required system of forces. This work presents a procedure which is able to deriving selectively sensitive excitation by iterative experiments. The first step is to determine the selectively sensitive displacement and selectively sensitive force patterns. These values are obtained by introducing the prior information of system parameters into an optimization which minimizes the sensitivities of the structure response with respect to the unselected parameters while keeping the sensitivities with respect to the selected parameters as a constant. In a second step the force pattern is used to derive dynamic loads on the tested structure and measurements are carried out. An automatic control ensures the required excitation forces. In a third step, measured outputs are employed to update the prior information. The strategy is to minimize the difference between a predicted displacement response, formulated as function of the unknown parameters and the measured displacements, and the selectively sensitive displacement calculated in the first step. With the updated values of the parameters a re-analysis of selective sensitivity is performed and the experiment is repeated until the displacement response of the model and the actual structure are conformed. As an illustration a simply supported beam made of steel, vibrated by harmonic excitation is investigated, thereby demonstrating that the adaptive excitation can be obtained efficiently.
The importance of modern simulation methods in the mechanical analysis of heterogeneous solids is presented in detail. Thereby the problem is noted that even for small bodies the required high-resolution analysis reaches the limits of today's computational power, in terms of memory demand as well as acceptable computational effort. A further problem is that frequently the accuracy of geometrical modelling of heterogeneous bodies is inadequate. The present work introduces a systematic combination and adaption of grid-based methods for achieving an essentially higher resolution in the numerical analysis of heterogeneous solids. Grid-based methods are as well primely suited for developing efficient and numerically stable algorithms for flexible geometrical modeling. A key aspect is the uniform data management for a grid, which can be utilized to reduce the effort and complexity of almost all concerned methods. A new finite element program, called Mulgrido, was just developed to realize this concept consistently and to test the proposed methods. Several disadvantages which generally result from grid discretizations are selectively corrected by modified methods. The present work is structured into a geometrical model, a mechanical model and a numerical model. The geometrical model includes digital image-based modeling and in particular several methods for the theory-based generation of inclusion-matrix models. Essential contributions refer to variable shape, size distribution, separation checks and placement procedures of inclusions. The mechanical model prepares the fundamentals of continuum mechanics, homogenization and damage modeling for the following numerical methods. The first topic of the numerical model introduces to a special version of B-spline finite elements. These finite elements are entirely variable in the order k of B-splines. For homogeneous bodies this means that the approximation quality can arbitrarily be scaled. In addition, the multiphase finite element concept in combination with transition zones along material interfaces yields a valuable solution for heterogeneous bodies. As the formulation is element-based, the storage of a global stiffness matrix is superseded such that the memory demand can essentially be reduced. This is possible in combination with iterative solver methods which represent the second topic of the numerical model. Here, the focus lies on multigrid methods where the number of required operations to solve a linear equation system only increases linearly with problem size. Moreover, for badly conditioned problems quite an essential improvement is achieved by preconditioning. The third part of the numerical model discusses certain aspects of damage simulation which are closely related to the proposed grid discretization. The strong efficiency of the linear analysis can be maintained for damage simulation. This is achieved by a damage-controlled sequentially linear iteration scheme. Finally a study on the effective material behavior of heterogeneous bodies is presented. Especially the influence of inclusion shapes is examined. By means of altogether more than one hundred thousand random geometrical arrangements, the effective material behavior is statistically analyzed and assessed.
In many applications such as parameter identification of oscillating systems in civil enginee-ring, speech processing, image processing and others we are interested in the frequency con-tent of a signal locally in time. As a start wavelet analysis provides a time-scale decomposition of signals, but this wavelet transform can be connected with an appropriate time-frequency decomposition. For instance in Matlab are defined pseudo-frequencies of wavelet scales as frequency centers of the corresponding bands. This frequency bands overlap more or less which depends on the choice of the biorthogonal wavelet system. Such a definition of frequency center is possible and useful, because different frequencies predominate at different dyadic scales of a wavelet decomposition or rather at different nodes of a wavelet packet decomposition tree. The goal of this work is to offer better algorithms for characterising frequency band behaviour and for calculating frequency centers of orthogonal and biorthogonal wavelet systems. This will be done with some product formulas in frequency domain. Now the connecting procedu-res are more analytical based, better connected with wavelet theory and more assessable. This procedures doesn’t need any time approximation of the wavelet and scaling functions. The method only works in the case of biorthogonal wavelet systems, where scaling functions and wavelets are defined over discrete filters. But this is the practically essential case, because it is connected with fast algorithms (FWT, Mallat Algorithm). At the end corresponding to the wavelet transform some closed formulas of pure oscillations are given. They can generally used to compare the application of different wavelets in the FWT regarding it’s frequency behaviour.
The design and application of high performance materials demands extensive knowledge of the materials damage behavior, which significantly depends on the meso- and microstructural complexity. Numerical simulations of crack growth on multiple length scales are promising tools to understand the damage phenomena in complex materials. In polycrystalline materials it has been observed that the grain boundary decohesion is one important mechanism that leads to micro crack initiation. Following this observation the paper presents a polycrystal mesoscale model consisting of grains with orthotropic material behavior and cohesive interfaces along grain boundaries, which is able to reproduce the crack initiation and propagation along grain boundaries in polycrystalline materials. With respect to the importance of modeling the geometry of the grain structure an advanced Voronoi algorithm is proposed to generate realistic polycrystalline material structures based on measured grain size distribution. The polycrystal model is applied to investigate the crack initiation and propagation in statically loaded representative volume elements of aluminum on the mesoscale without the necessity of initial damage definition. Future research work is planned to include the mesoscale model into a multiscale model for the damage analysis in polycrystalline materials.
The design of safety-critical structures, exposed to cyclic excitations demands for non-degrading or limited-degrading behavior during extreme events. Among others, the structural behavior is mainly determined by the amount of plastic cycles, completed during the excitation. Existing simplified methods often ignore this dependency, or assume/request sufficient cyclic capacity. The paper introduces a new performance based design method that considers explicitly a predefined number of re-plastifications. Hereby approaches from the shakedown theory and signal processing methods are utilized. The paper introduces the theoretical background, explains the steps of the design procedure and demonstrates the applicability with help of an example. This project was supported by German Science Foundation (Deutsche Forschungsgemeinschaft, DFG)
The present paper is part of a comprehensive approach of grid-based modelling. This approach includes geometrical modelling by pixel or voxel models, advanced multiphase B-spline finite elements of variable order and fast iterative solver methods based on the multigrid method. So far, we have only presented these grid-based methods in connection with linear elastic analysis of heterogeneous materials. Damage simulation demands further considerations. The direct stress solution of standard bilinear finite elements is severly defective, especially along material interfaces. Besides achieving objective constitutive modelling, various nonlocal formulations are applied to improve the stress solution. Such a corrective data processing can either refer to input data in terms of Young's modulus or to the attained finite element stress solution, as well as to a combination of both. A damage-controlled sequentially linear analysis is applied in connection with an isotropic damage law. Essentially by a high resolution of the heterogeneous solid, local isotropic damage on the material subscale allows to simulate complex damage topologies such as cracks. Therefore anisotropic degradation of a material sample can be simulated. Based on an effectively secantial global stiffness the analysis is numerically stable. The iteration step size is controlled for an adequate simulation of the damage path. This requires many steps, but in the iterative solution process each new step starts with the solution of the prior step. Therefore this method is quite effective. The present paper provides an introduction of the proposed concept for a stable simulation of damage in heterogeneous solids.
Advanced finite elements are proposed for the mechanical analysis of heterogeneous materials. The approximation quality of these finite elements can be controlled by a variable order of B-spline shape functions. An element-based formulation is developed such that the finite element problem can iteratively be solved without storing a global stiffness matrix. This memory saving allows for an essential increase of problem size. The heterogeneous material is modelled by projection onto a uniform, orthogonal grid of elements. Conventional, strictly grid-based finite element models show severe oscillating defects in the stress solutions at material interfaces. This problem is cured by the extension to multiphase finite elements. This concept enables to define a heterogeneous material distribution within the finite element. This is possible by a variable number of integration points to each of which individual material properties can be assigned. Based on an interpolation of material properties at nodes and further smooth interpolation within the finite elements, a continuous material function is established. With both, continuous B-spline shape function and continuous material function, also the stress solution will be continuous in the domain. The inaccuracy implied by the continuous material field is by far less defective than the prior oscillating behaviour of stresses. One- and two-dimensional numerical examples are presented.
In this paper an adaptive heterogeneous multiscale model, which couples two substructures with different length scales into one numerical model is introduced for the simulation of damage in concrete. In the presented approach the initiation, propagation and coalescence of microcracks is simulated using a mesoscale model, which explicitly represents the heterogeneous material structure of concrete. The mesoscale model is restricted to the damaged parts of the structure, whereas the undamaged regions are simulated on the macroscale. As a result an adaptive enlargement of the mesoscale model during the simulation is necessary. In the first part of the paper the generation of the heterogeneous mesoscopic structure of concrete, the finite element discretization of the mesoscale model, the applied isotropic damage model and the cohesive zone model are briefly introduced. Furthermore the mesoscale simulation of a uniaxial tension test of a concrete prism is presented and own obtained numerical results are compared to experimental results. The second part is focused on the adaptive heterogeneous multiscale approach. Indicators for the model adaptation and for the coupling between the different numerical models will be introduced. The transfer from the macroscale to the mesoscale and the adaptive enlargement of the mesoscale substructure will be presented in detail. A nonlinear simulation of a realistic structure using an adaptive heterogeneous multiscale model is presented at the end of the paper to show the applicability of the proposed approach to large-scale structures.
A fast solver method called the multigrid preconditioned conjugate gradient method is proposed for the mechanical analysis of heterogeneous materials on the mesoscale. Even small samples of a heterogeneous material such as concrete show a complex geometry of different phases. These materials can be modelled by projection onto a uniform, orthogonal grid of elements. As one major problem the possible resolution of the concrete specimen is generally restricted due to (a) computation times and even more critical (b) memory demand. Iterative solvers can be based on a local element-based formulation while orthogonal grids consist of geometrical identical elements. The element-based formulation is short and transparent, and therefore efficient in implementation. A variation of the material properties in elements or integration points is possible. The multigrid method is a fast iterative solver method, where ideally the computational effort only increases linear with problem size. This is an optimal property which is almost reached in the implementation presented here. In fact no other method is known which scales better than linear. Therefore the multigrid method gains in importance the larger the problem becomes. But for heterogeneous models with very large ratios of Young's moduli the multigrid method considerably slows down by a constant factor. Such large ratios occur in certain heterogeneous solids, as well as in the damage analysis of solids. As solution to this problem the multigrid preconditioned conjugate gradient method is proposed. A benchmark highlights the multigrid preconditioned conjugate gradient method as the method of choice for very large ratio's of Young's modulus. A proposed modified multigrid cycle shows good results, in the application as stand-alone solver or as preconditioner.
Projector-Based Augmentation
(2006)
Projector-based augmentation approaches hold the potential of combining the advantages of well-establishes spatial virtual reality and spatial augmented reality. Immersive, semi-immersive and augmented visualizations can be realized in everyday environments – without the need for special projection screens and dedicated display configurations. Limitations of mobile devices, such as low resolution and small field of view, focus constrains, and ergonomic issues can be overcome in many cases by the utilization of projection technology. Thus, applications that do not require mobility can benefit from efficient spatial augmentations. Examples range from edutainment in museums (such as storytelling projections onto natural stone walls in historical buildings) to architectural visualizations (such as augmentations of complex illumination simulations or modified surface materials in real building structures). This chapter describes projector-camera methods and multi-projector techniques that aim at correcting geometric aberrations, compensating local and global radiometric effects, and improving focus properties of images projected onto everyday surfaces.
Virtual studio technology plays an important role for modern television productions. Blue-screen matting is a common technique for integrating real actors or moderators into computer generated sceneries. Augmented reality offers the possibility to mix real and virtual in a more general context. This article proposes a new technological approach for combining real studio content with computergenerated information. Digital light projection allows a controlled spatial, temporal, chrominance and luminance modulation of illumination – opening new possibilities for TV studios.
Recent radiometric compensation techniques make it possible to project images onto colored and textured surfaces. This is realized with projector-camera systems by scanning the projection surface on a per-pixel basis. With the captured information, a compensation image is calculated that neutralizes geometric distortions and color blending caused by the underlying surface. As a result, the brightness and the contrast of the input image is reduced compared to a conventional projection onto a white canvas. If the input image is not manipulated in its intensities, the compensation image can contain values that are outside the dynamic range of the projector. They will lead to clipping errors and to visible artifacts on the surface. In this article, we present a novel algorithm that dynamically adjusts the content of the input images before radiometric compensation is carried out. This reduces the perceived visual artifacts while simultaneously preserving a maximum of luminance and contrast. The algorithm is implemented entirely on the GPU and is the first of its kind to run in real-time.
Summer overheating in buildings is a common problem, especially in office buildings with large glazed facades, high internal loads and low thermal mass. Phase change materials (PCM) that undergo a phase transition in the temperature range of thermal comfort can add thermal mass without increasing the structural load of the building. The investigated PCM were micro-encapsulated and mixed into gypsum plaster. The experiments showed a reduction of indoor-temperature of up to 4 K when using a 3 cm layer of PCM-plaster with micro-encapsulated paraffin. The measurement results could validate a numerical model that is based on a temperature dependent function for heat capacity. Thermal building simulation showed that a 3 cm layer of PCM-plaster can help to fulfil German regulations concerning heat protection of buildings in summer for most office rooms.
In this paper we study the structure of the solutions to higher dimensional Dirac type equations generalizing the known λ-hyperholomorphic functions, where λ is a complex parameter. The structure of the solutions to the system of partial differential equations (D- λ) f=0 show a close connection with Bessel functions of first kind with complex argument. The more general system of partial differential equations that is considered in this paper combines Dirac and Euler operators and emphasizes the role of the Bessel functions. However, contrary to the simplest case, one gets now Bessel functions of any arbitrary complex order.
The modeling of crack propagation in plain and reinforced concrete structures is still a field for many researchers. If a macroscopic description of the cohesive cracking process of concrete is applied, generally the Fictitious Crack Model is utilized, where a force transmission over micro cracks is assumed. In the most applications of this concept the cohesive model represents the relation between the normal crack opening and the normal stress, which is mostly defined as an exponential softening function, independently from the shear stresses in tangential direction. The cohesive forces are then calculated only from the normal stresses. By Carol et al. 1997 an improved model was developed using a coupled relation between the normal and shear damage based on an elasto-plastic constitutive formulation. This model is based on a hyperbolic yield surface depending on the normal and the shear stresses and on the tensile and shear strength. This model also represents the effect of shear traction induced crack opening. Due to the elasto-plastic formulation, where the inelastic crack opening is represented by plastic strains, this model is limited for applications with monotonic loading. In order to enable the application for cases with un- and reloading the existing model is extended in this study using a combined plastic-damage formulation, which enables the modeling of crack opening and crack closure. Furthermore the corresponding algorithmic implementation using a return mapping approach is presented and the model is verified by means of several numerical examples. Finally an investigation concerning the identification of the model parameters by means of neural networks is presented. In this analysis an inverse approximation of the model parameters is performed by using a given set of points of the load displacement curves as input values and the model parameters as output terms. It will be shown, that the elasto-plastic model parameters could be identified well with this approach, but require a huge number of simulations.
Die effektive Kooperation aller beteiligten Fachplaner im Bauplanungsprozess ist die Voraussetzung für wirtschaftliches und qualitativ hochwertiges Bauen. Bauprojektorganisationen bestehen in der Regel aus zahlreichen unabhängigen Planungspartnern, die örtlich verteilt spezifische Planungsaufgaben bearbeiten und die Ergebnisse in Teilproduktmodellen ablegen. Da Planungsprozesse im Bauwesen stark arbeitsteilig ablaufen, sind die Teilproduktmodelle der einzelnen Fachplanungen in hohem Maße voneinander abhängig. Ziel des hier vorgestellten Ansatzes ist die Integration der Teilproduktmodelle der Gebäudeplanung in einem netzwerkbasierten Modellverbund am Beispiel der Brandschutzplanung. Im Beitrag werden die Probleme der Verteiltheit und insbesondere der semantischen Heterogenität der involvierten Teilproduktmodelle betrachtet. Der verteilte Zugriff wird mithilfe mobiler Software-Agenten realisiert. Die Agenten können sich dabei frei im netzwerkbasierten Planungsverbund bewegen und agieren als Vertreter der Fachplaner. Das Problem der semantischen Heterogenität der Teilproduktmodelle wird auf der Basis von Ontologien gelöst. Dazu werden erstens Domänenontologien entwickelt, die Objekte der realen Welt einer abgeschlossenen Domäne, hier des Brandschutzes, abbilden. Zweitens werden Applikationsontologien entwickelt, die die einzelnen proprietären Datenhaltungen (im Sinne von Teilproduktmodellen) der jeweiligen Fachplanungen repräsentieren. Beide Ontologien werden mit einem regelbasierten Ansatz verknüpft. Im vorgestellten Anwendungsfall Brandschutz dient die Domänenontologie als einheitliche Schnittstelle für den Zugriff auf die verteilten Modelle und abstrahiert dabei von deren Datenbankspezifika und proprietären Schemata. Mithilfe von mobilen Agenten und semantischen Technologien kann so eine Plattform zur Verfügung gestellt werden, die erstens die dynamische Integration von Ressourcen in den Planungsverbund erlaubt und zweitens auf deren Basis unabhängig von der Verteiltheit und Heterogenität der eingebundenen Ressourcen ingenieurgerechte Verarbeitungsmethoden realisiert werden können.
In classical complex function theory the geometric mapping property of conformality is closely linked with complex differentiability. In contrast to the planar case, in higher dimensions the set of conformal mappings is only the set of Möbius transformations. Unfortunately, the theory of generalized holomorphic functions (by historical reasons they are called monogenic functions) developed on the basis of Clifford algebras does not cover the set of Möbius transformations in higher dimensions, since Möbius transformations are not monogenic. But on the other side, monogenic functions are hypercomplex differentiable functions and the question arises if from this point of view they can still play a special role for other types of 3D-mappings, for instance, for quasi-conformal ones. On the occasion of the 16th IKM 3D-mapping methods based on the application of Bergman's reproducing kernel approach (BKM) have been discussed. Almost all authors working before that with BKM in the Clifford setting were only concerned with the general algebraic and functional analytic background which allows the explicit determination of the kernel in special situations. The main goal of the abovementioned contribution was the numerical experiment by using a Maple software specially developed for that purpose. Since BKM is only one of a great variety of concrete numerical methods developed for mapping problems, our goal is to present a complete different from BKM approach to 3D-mappings. In fact, it is an extension of ideas of L. V. Kantorovich to the 3-dimensional case by using reduced quaternions and some suitable series of powers of a small parameter. Whereas until now in the Clifford case of BKM the recovering of the mapping function itself and its relation to the monogenic kernel function is still an open problem, this approach avoids such difficulties and leads to an approximation by monogenic polynomials depending on that small parameter.
The Element-free Galerkin Method has become a very popular tool for the simulation of mechanical problems with moving boundaries. The internally applied Moving Least Squares approximation uses in general Gaussian or cubic weighting functions and has compact support. Due to the approximative character of this method the obtained shape functions do not fulfill the interpolation condition, which causes additional numerical effort for the imposition of the essential boundary conditions. The application of a singular weighting function, which leads to singular coefficient matrices at the nodes, can solve this problem, but requires a very careful placement of the integration points. Special procedures for the handling of such singular matrices were proposed in literature, which require additional numerical effort. In this paper a non-singular weighting function is presented, which leads to an exact fulfillment of the interpolation condition. This weighting function leads to regular values of the weights and the coefficient matrices in the whole interpolation domain even at the nodes. Furthermore this function gives much more stable results for varying size of the influence radius and for strongly distorted nodal arrangements than classical weighting function types. Nevertheless, for practical applications the results are similar as these obtained with the regularized weighting type presented by the authors in previous publications. Finally a new concept will be presented, which enables an efficient analysis of systems with strongly varying node density. In this concept the nodal influence domains are adapted depending on the nodal configuration by interpolating the influence radius for each direction from the distances to the natural neighbor nodes. This approach requires a Voronoi diagram of the domain, which is available in this study since Delaunay triangles are used as integration background cells. In the numerical examples it will be shown, that this method leads to a more uniform and reduced number of influencing nodes for systems with varying node density than the classical circular influence domains, which means that the small additional numerical effort for interpolating the influence radius leads to remarkable reduction of the total numerical cost in a linear analysis while obtaining similar results. For nonlinear calculations this advantage would be even more significant.
In this paper we consider three different methods for generating monogenic functions. The first one is related to Fueter's well known approach to the generation of monogenic quaternion-valued functions by means of holomorphic functions, the second one is based on the solution of hypercomplex differential equations and finally the third one is a direct series approach, based on the use of special homogeneous polynomials. We illustrate the theory by generating three different exponential functions and discuss some of their properties. Formula que se usa em preprints e artigos da nossa UI&D (acho demasiado completo): Partially supported by the R\&D unit \emph{Matem\'atica a Aplica\c\~es} (UIMA) of the University of Aveiro, through the Portuguese Foundation for Science and Technology (FCT), co-financed by the European Community fund FEDER.
In engineering science the modeling and numerical analysis of complex systems and relations plays an important role. In order to realize such an investigation, for example a stochastic analysis, in a reasonable computational time, approximation procedure have been developed. A very famous approach is the response surface method, where the relation between input and output quantities is represented for example by global polynomials or local interpolation schemes as Moving Least Squares (MLS). In recent years artificial neural networks (ANN) have been applied as well for such purposes. Recently an adaptive response surface approach for reliability analyses was proposed, which is very efficient concerning the number of expensive limit state function evaluations. Due to the applied simplex interpolation the procedure is limited to small dimensions. In this paper this approach is extended for larger dimensions using combined ANN and MLS response surfaces for evaluating the adaptation criterion with only one set of joined limit state points. As adaptation criterion a combination by using the maximum difference in the conditional probabilities of failure and the maximum difference in the approximated radii is applied. Compared to response surfaces on directional samples or to plain directional sampling the failure probability can be estimated with a much smaller number of limit state points.
At the 16th IKM Bock, Falcão and Gürlebeck presented examples of the application of some specially developed Maple-Software in hypercomplex analysis. Other papers of those authors continued this work and showed the efficiency of such tools for concrete numerical calculations as well as for numerical experiments, supporting the detection of new relationships and even theorems in a highly technical theoretical work. The mentioned software has been developed mainly for the use on mapping problems in the Euclidean spaces of dimension 3 and 4 by means of Bergman kernel methods (BKM), which are related to monogenic functions as solutions of generalized Cauchy-Riemann equations with respect to the Euclidean metric (Riesz system). The developed procedures concerning generalized powers of totally regular variables and the corresponding homogeneous polynomials basically rely on results and conventions introduced in the paper "Power series representation for monogenic functions in Rm+1 based on a permutational product", Complex Variables, 15, No.3, 181-191 (1990) by H. Malonek. Since 1992 H. Leutwiler, S. L. Eriksson and others developed in a number of papers a modified Clifford analysis and, particularly, a modified quaternionic analysis. The modification mainly consists in considering generalized Cauchy-Riemann equations with respect to a hyperbolic metric in a half space. The aim of this contribution is to show how through a change of the basic combinatorial relations used in the modified quaternionic analysis the aforementioned Maple-software (that has been recently published on CD-Rom as integrated part of the text book "Funktionentheorie in der Ebene und im Raum" by K. Gürlebeck, K. Habetha, and W. Sprössig, in the series "Grundstudium Mathematik" of Birkhäuser Verlag, 2006) can directly be used for numerical calculations in the modified theory.
Am Beispiel eines 3-feldrigen Durchlaufträgers wird die Versagenswahrscheinlichkeit von wechselnd belasteten Stahlbetonbalken bezüglich des Grenzzustandes der Adaption (Einspielen, shakedown) untersucht. Die Adaptionsanalyse erfolgt unter Berücksichtigung der beanspruchungschabhängigen Degradation der Biegesteifigkeit infolge Rissbildung. Die damit verbundene mechanische Problemstellung kann auf die Adaptionsanalyse linear elastisch - ideal plastischer Balkentragwerke mit unbekannter aber begrenzter Biegesteifigkeit zurückgeführt werden. Die Versagenswahrscheinlichkeit wird unter Berücksichtigung stochastischer Tragwerks- und Belastungsgrößen berechnet. Tragwerkseigenschaften und ständige Lasten gelten als zeitunabhängige Zufallsgrößen. Zeitlich veränderliche Lasten werden als nutzungsdauerbezogene Extremwerte POISSONscher Rechteck-Pulsprozesse unter Berücksichtigung zeitlicher Überlagerungseffekte modelliert, so dass die Versagenswahrscheinlichkeit ebenfalls eine nutzungsdauerbezogene Größe ist. Die mechanischen Problemstellungen werden numerisch mit der mathematischen Optimierung gelöst. Die Versagenswahrscheinlichkeit wird auf statistischem Weg mit der Monte-Carlo-Methode geschätzt.
We propose a novel method that applies the light transport matrix for performing an image-based radiometric compensation which accounts for all possible types of light modulation. For practical application the matrix is decomposed into clusters of mutually influencing projector and camera pixels. The compensation is modeled as a linear system that can be solved with respect to the projector patterns. Precomputing the inverse light transport in combination with an efficient implementation on the GPU makes interactive compensation rates possible. Our generalized method unifies existing approaches that address individual problems. Based on examples, we show that it is possible to project corrected images onto complex surfaces such as an inter-reflecting statuette, glossy wallpaper, or through highly-refractive glass. Furthermore, we illustrate that a side-effect of our approach is an increase in the overall sharpness of defocused projections.
Moderne Büroarchitektur mit Räumen in Leichtbauweise und großen transparenten Fassa-denanteilen verschärft im Zusammenwirken mit hohen internen Lasten die Problematik der sommerlichen Überhitzung in Gebäuden. Phasenübergangsmaterialien (PCM: phase change materials) stellen eine interessante Möglichkeit dar, sommerliche Überhitzung in Gebäuden ohne aufwändige Anlagentechnik wie beispielsweise Klimaanlagen zu reduzieren. Der thermische Komfort in Räumen, die mit einem PCM-Putz ausgestattet sind, kann signifikant erhöht werden. Die Arbeit untersucht Anwendungsmöglichkeiten und Optimierungspotential eines PCM-Putzes auf experimentelle und numerische Weise. Zur Untersuchung des PCM-Putzes wurden materialtechnische und experimentelle sowie numerische und numerisch-analytische Methoden eingesetzt. Die Kenntnis der thermischen Parameter des PCM-Putzes ist unablässig für die Berechnung der möglichen Temperaturreduktionen. Zur Bestimmung der Latentwärme, des qualitativen Schmelz- und Erstarrungsprozesses sowie des Temperaturintervalls, in dem der Phasenübergang stattfindet, wurden Messungen mit einem Differential Scanning Calorimeter (DSC) durchgeführt. Für die experimentelle Untersuchung des PCM-Putzes wurden zwei identische Testräume in Leichtbauweise erstellt. Die Räume wurden im Verifikationsobjekt „Eiermannbau“ des Sonderforschungsbereiches SFB 524 der Bauhaus-Universität Weimar gemessen. Nach der Überprüfung, dass sich beide Räume thermisch gleich verhalten, wurde ein Raum mit dem PCM-Putz und der zweite Raum mit einem vergleichbaren Innenputz ohne PCM verputzt. Thermoelemente zur Temperaturmessung im Bauteil, an der Oberfläche und zur Raumlufttemperaturbestimmung wurden angebracht und mit einer Messwerterfassungsanlage verbunden. Der Verlauf der Außenlufttemperatur und die Globalstrahlung am Standort der Versuchsräume wurden aufgezeichnet, um einen Klimadatensatz zu erstellen. Für die Berechnung der Temperaturverteilung in einem PCM-Bauteil mit kontinuierlichem Phasenübergang existiert keine geschlossene analytische Lösung. Daher wurde ein numerischer Ansatz gewählt, bei dem der Phasenübergang im Temperaturbereich T1 bis T2 mit Hilfe einer temperaturabhängigen Wärmekapazität c(T) innerhalb der erweiterten Fou-rier’schen Wärmeleitungsgleichung dargestellt wird. Die Funktion c(T) wird auf Basis der DSC-Messungen bestimmt. Die Modellierung erfolgte mit einem Finite-Differenzen-Verfahren auf Grundlage der Fourier’schen Wärmeleitungsgleichung. Im Rahmen der Arbeit wurde ein PCM-Modul entwickelt, das in ein Gebäudesimulationsprogramm implementiert wurde. Mit dem neuen Modul lassen sich sowohl die Temperaturverläufe in einem PCM-Bauteil wie auch seine Wechselwirkung mit dem Raumklima darstellen. Eine Validierung des entwickelten PCM-Moduls anhand von zahlreichen experimentellen Daten der Versuchsräume wurde für das PCM-Modul erfolgreich durchgeführt. Sommerliche Überhitzungsstunden können durch PCM in Wand- und Deckenelementen deutlich reduziert werden. Der PCM-Putz eignet sich vor allem für Anwendungen in Leichtbauten wie z.B. moderne Büroräume. In Räumen, in denen bereits eine ausreichende thermische Masse vorhanden ist, ist die Temperaturreduktion durch PCM nur gering. Kann das PCM während der Nachtstunden nicht erstarren, erschöpft sich seine Fähigkeit zur Latentwärmespeicherung. Erhöhte Nachtlüftung führt bei entsprechend niedrigen Außentemperaturen zu höherem Wärmeübergang und kann damit zur besseren Entladung des PCM beitragen. Im Rahmen der Dissertation konnten Aussagen zur idealen Phasenübergangstemperatur in Abhängigkeit des verwendeten Materials und der Schichtdicke getroffen werden. Die Reduktion der Oberflächentemperaturen, die sich bei Einsatz eines PCM-Putzes unter geeigneten Randbedingungen ergibt, beträgt 2.0 - 3.5 K für eine Putzschicht von 1 cm und 3.0 - 5.0 K für eine Putzschicht von 3 cm. Diese Werte wurden sowohl numerisch als auch durch experimentelle Untersuchungen ermittelt. Die Reduktion der Lufttemperaturen aufgrund einer Konditionierung des Raumes mit PCM-Putz beträgt bei geeigneten thermischen Verhältnissen ca. 1.0 - 2.5 K für eine Putzschicht von 1 cm und 2.0 - 3.0 K für eine Putzschicht von 3 cm. Die operative Temperatur als wichtiger Komfortparameter kann durch den Einsatz des PCM-Putzes um bis zu 4 K gesenkt werden. Damit lässt sich mit Hilfe eines PCM-Putzes die thermische Behaglichkeit in einem Raum deutlich erhöhen.
ON THE NAVIER-STOKES EQUATION WITH FREE CONVECTION IN STRIP DOMAINS AND 3D TRIANGULAR CHANNELS
(2006)
The Navier-Stokes equations and related ones can be treated very elegantly with the quaternionic operator calculus developed in a series of works by K. Guerlebeck, W. Sproeossig and others. This study will be extended in this paper. In order to apply the quaternionic operator calculus to solve these types of boundary value problems fully explicitly, one basically needs to evaluate two types of integral operators: the Teodorescu operator and the quaternionic Bergman projector. While the integral kernel of the Teodorescu transform is universal for all domains, the kernel function of the Bergman projector, called the Bergman kernel, depends on the geometry of the domain. With special variants of quaternionic holomorphic multiperiodic functions we obtain explicit formulas for three dimensional parallel plate channels, rectangular block domains and regular triangular channels. The explicit knowledge of the integral kernels makes it then possible to evaluate the operator equations in order to determine the solutions of the boundary value problem explicitly.
Heutige Methoden zur Soll-Spezifikation von Bauleistungen (Kostenermittlung und zeitliche Ablaufplanung) gehen von einer abstrahierten und vereinfachten Betrachtung der Zusammenhänge bei Bauprojekten aus. Leistungsverzeichnisse, Kostenermittlungen und Bauzeitpläne orientieren sich nur indirekt an der Geometrie des Bauwerks und der Baustelle. Die dabei verwendeten Medien wie Papier, 2D-Dateien, digitale Leistungsbeschreibungen oder 3D-Darstellungen lassen die Suche nach Informationen auf der Baustelle zu einem zeitaufwändigen und in Anbetracht existierender Medientechnologien ineffizienten Prozess werden. Interaktive virtuelle Umgebungen erlauben die Auflösung starrer Zusammenhänge durch interaktive Eingriffe des Anwenders und visualisieren komplexe bauproduktionstechnische Vorgänge. Das Konzept der visuellen interaktiven Simulation der Bauproduktion sieht vor, die Soll-Spezifikation anhand eines interaktiven 3D-Modells zu entwickeln, um räumliche Veränderungen und parallele Prozesse auf der virtuellen Baustelle im Rahmen der Entscheidungsfindung zum Bauablauf besser berücksichtigen zu können. Verlangt man einen hohen Grad an Interaktivität mit dem 3D-Modell, dann bieten sich Computerspieltechnologien sehr gut zu Verifikationszwecken an. Die visuelle interaktive Simulation der Bauproduktion ist damit als eine 3D-modellbasierte Methode der Prozessmodellierung zu verstehen, die Entscheidungen als Input benötigt und die Kostenermittlung sowie die zeitliche Ablaufplanung als Output liefert.
Die Arbeit befasst sich mit der Anwendung faserverstärkter Kunststoffe für Tragwerke des Hochbaus. Es wird ein geschichtlicher Überblick über die Jahre 1950 bis 1980 gegeben und dabei herausgestellt, wie es 1. zur Einführung des bis 1950 unbekannten Werkstoffes im Bauwesen kommen konnte 2. welche Personen und Institute maßgeblich an der Einführung und Entwicklung des Bauens mit FVK beteiligt waren 3. welche Tragwerke verwendet wurden 4. wie die Pioniere diese Tragwerke bemaßen 5. welche konstruktiven Besonderheiten sich mit der Verwendung von FVK in der Tragstruktur ergaben Nach einer Einführung werden im Kapitel 2 die wichtigsten Faktoren der Entwicklung von Tragwerken aus GFK erörtert. Im Kapitel 3 wird die Technik der Fertigung von GFK-Teilen und deren Fügung beschrieben. Im Kapitel 4 werden die Tragwerke beschrieben und einzelne Tragwerkstypen eingehend erörtert. Im Kapitel 5 werden die Bemessungskonzepte und deren Entwicklung erörtert. In der Bilanz werden die Faktoren aufgezählt, die zum Abklingen des Bauens mit FVK in der Tragstruktur geführt haben. Die Arbeit wird ergänzt durch eine ca. 40-seitige Tabelle in der die gebauten Tragwerke in Abhängigkeit von den technischen Parametern Spannweiten und Lasten dargestellt werden. Im Anhang werden 10 exemplarische Bauten detailliert erörtert.
Eine zielführende Anwendung von Zusatzmitteln bei der Ausführung anspruchsvoller Betonbauten setzt einen hohen Kenntnisstand bezüglich der Wirkungsmechanismen und Interaktionen der einzelnen Betonkomponenten voraus. In der vorliegenden Arbeit wurden einige Aspekte der Zementhydratation in Abhängigkeit von der Fließmittelzugabe diskutiert. Im Ergebnis liefern die Teile eins und zwei der vorliegenden Arbeit einen Beitrag dazu, Veränderungen der Fließfähigkeit von Zementleim in Abhängigkeit der Zementhydratation und Fließmittelzugabe besser zu verstehen. Es konnte so z.B. gezeigt werden, dass Bildung langprismatischer Kristalle (z.B. Syngenit, Gips) die Fließfähigkeit von Zementleim und Beton vermindert. Infolge anhaltender Scherung von Zementleimen / Betonen mit langprismatischen Kristallen wird ein Zuwachs an Fließfähigkeit erzielt. Elektronenmikroskopische Untersuchungen zeigen, dass dies darauf zurückzuführen ist, dass die Kristalle in eine Vorzugsorientierung relativ zur Scherbewegung rotieren. Weiterhin wurde der Mechanismus einer so genannten Zement-Fließmittel-Inkompatibilität aufgezeigt. Durch diese Erweiterung des Kenntnisstandes zum Einfluss von Fließmitteln auf die Zementhydratation ist es möglich der Zement-Fließmittel-Inkompatibilität durch gezielte Auswahl des Zementes vorzubeugen. Dabei ist besonders darauf zu achten, dass der Zement ein ausgewogenes Verhältnis an zur Reaktion zur Verfügung stehendem C3A und Menge / Löslichkeit des Abbindereglers besitzt. Fließmittel verändern nicht nur die Verarbeitungseigenschaften sondern auch die Festigkeit und Dauerhaftigkeit von Zementstein und Beton. Im dritten Teil der vorliegenden Arbeit wird daher der Einfluss der Fließmittel und deren verflüssigender Wirkung auf die Festigkeitsentwicklung von Zementstein und C3S untersucht. Es konnte gezeigt werden, dass durch die dispergierende Wirkung der Fließmittel auch ohne Verminderung des Wasserzementwertes, eine Verdichtung des Zementsteingefüges erzielt werden kann. Es konnte weiterhin gezeigt werden, dass durch die Erhöhung der Partikelpackungsdichte am Anfang der Hydratation die Ausbildung der festigkeitsgebenden C-S-H Phasen verändert wird. Ein dichteres Verwachsen dieser nanostrukturierten C-S-H Phasen ermöglicht einen zusätzlichen Festigkeitszuwachs.
In this paper we evaluate 2D models for soil-water characteristic curve (SWCC), that incorporate the hysteretic nature of the relationship between volumetric water content Θ and suction Ψ. The models are based on nonlinear least squares estimation of the experimental data for sand. To estimate the dependent variable Θ the proposed models include two independent variables, suction and sensors reading position (depth d in the column test). The variable d represents not only the position where suction and water content are measured but also the initial suction distribution before each of the hydraulic loading test phases. Due to this the proposed 2D regression models acquire the advantage that they: (a) can be applied for prediction of Θ for any position along the column and (b) give the functional form for the scanning curves.
The quaternionic operator calculus can be applied very elegantly to solve many important boundary value problems arising in fluid dynamics and electrodynamics in an analytic way. In order to set up fully explicit solutions. In order to apply the quaternionic operator calculus to solve these types of boundary value problems fully explicitly, one has to evaluate two types of integral operators: the Teodorescu operator and the quaternionic Bergman projector. While the integral kernel of the Teodorescu transform is universal for all domains, the kernel function of the Bergman projector, called the Bergman kernel, depends on the geometry of the domain. Recently the theory of quaternionic holomorphic multiperiodic functions and automorphic forms provided new impulses to set up explicit representation formulas for large classes of hyperbolic polyhedron type domains. These include block shaped domains, wedge shaped domains (with or without additional rectangular restrictions) and circular symmetric finite and infinite cylinders as particular subcases. In this talk we want to give an overview over the recent developments in this direction.
... WITHOUT RIGHT ANGLE.
(2006)
Currently sculptural design is one of the most discussed themes in architecture. Due to their light weight, easy transportation and assembly, as well as an almost unlimited structural variety, parameterised spatial structures are excellently suited for constructive realisation of free formed claddings. They subdivide the continuous surface into a structure of small sized nodes, straight members and plane glass panels. Thus they provide an opportunity to realise arbitrary double-curved claddings with a high degree of transparency, using industrial semi-finished products (steel sections, flat glass). Digital design strategies and a huge number of similar looking but in detail unique structural members demand a continuous digital project handling. Within a research project, named MYLOMESH, a free-formed spatial structure was designed, constructed, fabricated and assembled. All required steps were carried out based on digital data. Different digital connections (scripts) between varying software tools, which are usually not used in the planning process of buildings, were created. They allow a completely digital workflow. The project, its design, meshing, constructive detailing and the above-mentioned scripts are described in this paper.
TOOL TO CHECK TOPOLOGY AND GEOMETRY FOR SPATIAL STRUCTURES ON BASIS OF THE EXTENDED MAXWELL'S RULE
(2006)
One of the simplest principle in the design of light-weight structures is to avoid bending. This can be achieved by dissolving girders into members acting purely in axial tension or compression. The employment of cables for the tensioned members leads to even lighter structures which are called cable-strut structures. They constitute a subclass of spatial structures. To give fast information about the general feasibility of an architectural concept employing cable-strut structures is a challenging task due to their sophisticated mechanical behavior. In this regard it is essential to control if the structure is stable and if pre-stress can be applied. This paper presents a tool using the spreadsheet software Microsoft (MS) Excel which can give such information. Therefore it is not necessary to purchase special software and the according time consuming training is much lower. The tool was developed on basis of the extended Maxwell's rule, which besides topology also considers the geometry of the structure. For this the rank of the node equilibrium matrix is crucial. Significance and determination of the rank and the implementation of the corresponding algorithms in MS Excel are described in the following. The presented tool is able to support the structural designer in an early stage of the project in finding a feasible architectural concept for cable-strut structures. As examples for the application of the software tool two special cable-strut structures, so called tensegrity structures, were examined for their mechanical behavior.
Wolken
(2006)
Projector-based displays have been evolving tremendously in the last decade. Reduced costs and increasing capabilities have let to a widespread use for home entertainment and scientific visualization. The rapid development is continuing - techniques that allow seamless projection onto complex everyday environments such as textured walls, window curtains or bookshelfs have recently been proposed. Although cameras enable a completely automatic calibration of the systems, all previously described techniques rely on a precise mapping between projector and camera pixels. Global illumination effects such as reflections, refractions, scattering, dispersion etc. are completely ignored since only direct illumination is taken into account. We propose a novel method that applies the light transport matrix for performing an image-based radiometric compensation which accounts for all possible lighting effects. For practical application the matrix is decomposed into clusters of mutually influencing projector and camera pixels. The compensation is modeled as a linear equation system that can be solved separately for each cluster. For interactive compensation rates this model is adapted to enable an efficient implementation on programmable graphics hardware. Applying the light transport matrix's pseudo-inverse allows to separate the compensation into a computational expensive preprocessing step (computing the pseudo-inverse) and an on-line matrix-vector multiplication. The generalized mathematical foundation for radiometric compensation with projector-camera systems is validated with several experiments. We show that it is possible to project corrected imagery onto complex surfaces such as an inter-reflecting statuette and glass. The overall sharpness of defocused projections is increased as well. Using the proposed optimization for GPUs, real-time framerates are achieved.
Für das städtische Leben hat die Existenz sich verändernder Personenströme eine grundlegende Bedeutung.
Ein Werkzeug, das solche kollektiven Bewegungsmuster sichtbar machen kann, wäre dabei ein bedeutendes Hilfsmittel für die Stadtplanung. Im Mittelpunkt der vorliegenden Arbeit steht die Auseinandersetzung mit der Space Syntax Methode.
Diese untersucht den Zusammenhang räumlicher Strukturen mit deren Nutzung. Eine wichtige Erkenntnis ist, dass das kollektive menschliche Verhalten im öffentlichen Raum berechenbar ist. Die Tatsache, dass Passanten bestimmte Wege bevorzugen und andere meiden, führt Space Syntax dabei auf stadträumliche Ursachen zurück. Der Begriff des Natural Movement beschreibt den Anteil dieser räumlich bedingten Nutzung.
Die vorliegende Arbeit gliedert sich in einen theoretischen und einen praktischen Teil. Zunächst werden die für das Verständnis der Space Syntax Methode wichtigen Begriffe und Maßgrößen beschrieben. Der methodische Teil wird ergänzt durch eine Gegenüberstellung nutzer-basierter wissenschaftlicher Ansätze mit der räumlich-orientierten Space Syntax Methode.
Im zweiten Teil der Arbeit wird die praktische Anwendung der Space Syntax Methode am Beispiel des Leipziger City-Tunnels vorgenommen. Das Fallbeispiel ist prädestiniert für die Untersuchung, wie die regionale Vernetzung des Schienenverkehrs aktuell betrieben wird. Von Interesse ist dabei sowohl die Konzeption des zukünftigen Liniennetzes als auch dessen stadträumliche Einbindung.