@inproceedings{EblingScheuermann, author = {Ebling, Julia and Scheuermann, G.}, title = {TEMPLATE MATCHING ON VECTOR FIELDS USING CLIFFORD ALGEBRA}, editor = {G{\"u}rlebeck, Klaus and K{\"o}nke, Carsten}, organization = {Bauhaus-Universit{\"a}t Weimar}, doi = {10.25643/bauhaus-universitaet.2946}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170327-29464}, pages = {25}, abstract = {Due to the amount of flow simulation and measurement data, automatic detection, classification and visualization of features is necessary for an inspection. Therefore, many automated feature detection methods have been developed in recent years. However, only one feature class is visualized afterwards in most cases, and many algorithms have problems in the presence of noise or superposition effects. In contrast, image processing and computer vision have robust methods for feature extraction and computation of derivatives of scalar fields. Furthermore, interpolation and other filter can be analyzed in detail. An application of these methods to vector fields would provide a solid theoretical basis for feature extraction. The authors suggest Clifford algebra as a mathematical framework for this task. Clifford algebra provides a unified notation for scalars and vectors as well as a multiplication of all basis elements. The Clifford product of two vectors provides the complete geometric information of the relative positions of these vectors. Integration of this product results in Clifford correlation and convolution which can be used for template matching of vector fields. For frequency analysis of vector fields and the behavior of vector-valued filters, a Clifford Fourier transform has been derived for 2D and 3D. Convolution and other theorems have been proved, and fast algorithms for the computation of the Clifford Fourier transform exist. Therefore the computation of Clifford convolution can be accelerated by computing it in Clifford Fourier domain. Clifford convolution and Fourier transform can be used for a thorough analysis and subsequent visualization of flow fields.}, subject = {Architektur }, language = {en} } @inproceedings{EbertLenzen, author = {Ebert, Carsten and Lenzen, Armin}, title = {OUTPUT-ONLY ANALYSIS FOR EXPERIMENTAL DAMAGE DETECTION OF A TIED-ARCH BRIDGE}, editor = {G{\"u}rlebeck, Klaus and K{\"o}nke, Carsten}, organization = {Bauhaus-Universit{\"a}t Weimar}, doi = {10.25643/bauhaus-universitaet.2945}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170327-29452}, pages = {13}, abstract = {In civil engineering it is very difficult and often expensive to excite constructions such as bridges and buildings with an impulse hammer or shaker. This problem can be avoided with the output-only method as special feature of stochastic system identification. The permanently existing ambient noise (e.g. wind, traffic, waves) is sufficient to excite the structures in their operational conditions. The output-only method is able to estimate the observable part of a state-space-model which contains the dynamic characteristics of the measured mechanical system. Because of the assumption that the ambient excitation is white there is no requirement to measure the input. Another advantage of the output-only method is the possibility to get high detailed models by a special method, called polyreference setup. To pretend the availability of a much larger set of sensors the data from varying sensor locations will be collected. Several successive data sets are recorded with sensors at different locations (moving sensors) and fixed locations (reference sensors). The covariance functions of the reference sensors are bases to normalize the moving sensors. The result of the following subspace-based system identification is a high detailed black-box-model that contains the weighting function including the well-known dynamic parameters eigenfrequencies and mode shapes of the mechanical system. Emphasis of this lecture is the presentation of an extensive damage detection experiment. A 53-year old prestressed concrete tied-arch-bridge in H{\"u}nxe (Germany) was deconstructed in 2005. Preliminary numerous vibration measurements were accomplished. The first experiment for system modification was an additional support near the bridge bearing of one main girder. During a further experiment one hanger from one tied arch was cut through as an induced damage. Some first outcomes of the described experiments will be presented.}, subject = {Architektur }, language = {en} } @inproceedings{SampaioHenriquesStuderetal., author = {Sampaio, Alcinia Zita and Henriques, Pedro and Studer, P. and Luizi, Rui}, title = {VIRTUAL REALITY TECHNOLOGY TO REPRESENT CONSTRUCTION ACTIVITIES}, editor = {G{\"u}rlebeck, Klaus and K{\"o}nke, Carsten}, organization = {Bauhaus-Universit{\"a}t Weimar}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170327-30090}, pages = {9}, abstract = {The use of virtual reality techniques in the development of educational applications brings new perspectives to the teaching of subjects related to the field of civil construction in Civil Engineering domain. In order to obtain models, which would be able to visually simulate the construction process of two types of construction work, the research turned to the techniques of geometric modelling and virtual reality. The applications developed for this purpose are concerned with the construction of a cavity wall and a bridge. These models make it possible to view the physical evolution of the work, to follow the planned construction sequence and to visualize details of the form of every component of the works. They also support the study of the type and method of operation of the equipment necessary for these construction procedures. These models have been used to distinct advantage as educational aids in first-degree courses in Civil Engineering. Normally, three-dimensional geometric models, which are used to present architectural and engineering works, show only their final form, not allowing the observation of their physical evolution. The visual simulation of the construction process needs to be able to produce changes to the geometry of the project dynamically. In the present study, two engineering construction work models were created, from which it was possible to obtain three-dimensional models corresponding to different states of their form, simulating distinct stages in their construction. Virtual reality technology was applied to the 3D models. Virtual reality capacities allow the interactive real-time viewing of 3D building models and facilitate the process of visualizing, evaluating and communicating.}, subject = {Architektur }, language = {en} } @inproceedings{BrackxDeSchepperSommen, author = {Brackx, Fred and De Schepper, Nele and Sommen, Frank}, title = {Clifford-Hermite and Two-Dimensional Clifford-Gabor Filters For Early Vision}, editor = {G{\"u}rlebeck, Klaus and K{\"o}nke, Carsten}, organization = {Bauhaus-Universit{\"a}t Weimar}, doi = {10.25643/bauhaus-universitaet.2930}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170327-29303}, pages = {22}, abstract = {Image processing has been much inspired by the human vision, in particular with regard to early vision. The latter refers to the earliest stage of visual processing responsible for the measurement of local structures such as points, lines, edges and textures in order to facilitate subsequent interpretation of these structures in higher stages (known as high level vision) of the human visual system. This low level visual computation is carried out by cells of the primary visual cortex. The receptive field profiles of these cells can be interpreted as the impulse responses of the cells, which are then considered as filters. According to the Gaussian derivative theory, the receptive field profiles of the human visual system can be approximated quite well by derivatives of Gaussians. Two mathematical models suggested for these receptive field profiles are on the one hand the Gabor model and on the other hand the Hermite model which is based on analysis filters of the Hermite transform. The Hermite filters are derivatives of Gaussians, while Gabor filters, which are defined as harmonic modulations of Gaussians, provide a good approximation to these derivatives. It is important to note that, even if the Gabor model is more widely used than the Hermite model, the latter offers some advantages like being an orthogonal basis and having better match to experimental physiological data. In our earlier research both filter models, Gabor and Hermite, have been developed in the framework of Clifford analysis. Clifford analysis offers a direct, elegant and powerful generalization to higher dimension of the theory of holomorphic functions in the complex plane. In this paper we expose the construction of the Hermite and Gabor filters, both in the classical and in the Clifford analysis framework. We also generalize the concept of complex Gaussian derivative filters to the Clifford analysis setting. Moreover, we present further properties of the Clifford-Gabor filters, such as their relationship with other types of Gabor filters and their localization in the spatial and in the frequency domain formalized by the uncertainty principle.}, subject = {Architektur }, language = {en} } @inproceedings{BrackxDeKnockDeSchepper, author = {Brackx, Fred and De Knock, B. and De Schepper, Hennie}, title = {A MULTI--DIMENSIONAL HILBERT TRANSFORM IN ANISOTROPIC CLIFFORD ANALYSIS}, editor = {G{\"u}rlebeck, Klaus and K{\"o}nke, Carsten}, organization = {Bauhaus-Universit{\"a}t Weimar}, doi = {10.25643/bauhaus-universitaet.2929}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170327-29297}, pages = {15}, abstract = {In earlier research, generalized multidimensional Hilbert transforms have been constructed in m-dimensional Euclidean space, in the framework of Clifford analysis. Clifford analysis, centred around the notion of monogenic functions, may be regarded as a direct and elegant generalization to higher dimension of the theory of the holomorphic functions in the complex plane. The considered Hilbert transforms, usually obtained as a part of the boundary value of an associated Cauchy transform in m+1 dimensions, might be characterized as isotropic, since the metric in the underlying space is the standard Euclidean one. In this paper we adopt the idea of a so-called anisotropic Clifford setting, which leads to the introduction of a metric dependent m-dimensional Hilbert transform, showing, at least formally, the same properties as the isotropic one. The Hilbert transform being an important tool in signal analysis, this metric dependent setting has the advantage of allowing the adjustment of the co-ordinate system to possible preferential directions in the signals to be analyzed. A striking result to be mentioned is that the associated anisotropic (m+1)-dimensional Cauchy transform is no longer uniquely determined, but may stem from a diversity of (m+1)-dimensional "mother" metrics.}, subject = {Architektur }, language = {en} } @inproceedings{BrackxDeSchepperDeSchepperetal., author = {Brackx, Fred and De Schepper, Hennie and De Schepper, Nele and Sommen, Frank}, title = {HERMITIAN CLIFFORD-HERMITE WAVELETS}, editor = {G{\"u}rlebeck, Klaus and K{\"o}nke, Carsten}, organization = {Bauhaus-Universit{\"a}t Weimar}, doi = {10.25643/bauhaus-universitaet.2931}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170327-29313}, pages = {13}, abstract = {The one-dimensional continuous wavelet transform is a successful tool for signal and image analysis, with applications in physics and engineering. Clifford analysis offers an appropriate framework for taking wavelets to higher dimension. In the usual orthogonal case Clifford analysis focusses on monogenic functions, i.e. null solutions of the rotation invariant vector valued Dirac operator ∂, defined in terms of an orthogonal basis for the quadratic space Rm underlying the construction of the Clifford algebra R0,m. An intrinsic feature of this function theory is that it encompasses all dimensions at once, as opposed to a tensorial approach with products of one-dimensional phenomena. This has allowed for a very specific construction of higher dimensional wavelets and the development of the corresponding theory, based on generalizations of classical orthogonal polynomials on the real line, such as the radial Clifford-Hermite polynomials introduced by Sommen. In this paper, we pass to the Hermitian Clifford setting, i.e. we let the same set of generators produce the complex Clifford algebra C2n (with even dimension), which we equip with a Hermitian conjugation and a Hermitian inner product. Hermitian Clifford analysis then focusses on the null solutions of two mutually conjugate Hermitian Dirac operators which are invariant under the action of the unitary group. In this setting we construct new Clifford-Hermite polynomials, starting in a natural way from a Rodrigues formula which now involves both Dirac operators mentioned. Due to the specific features of the Hermitian setting, four different types of polynomials are obtained, two types of even degree and two types of odd degree. These polynomials are used to introduce a new continuous wavelet transform, after thorough investigation of all necessary properties of the involved polynomials, the mother wavelet and the associated family of wavelet kernels.}, subject = {Architektur }, language = {en} } @inproceedings{WittwerBecker, author = {Wittwer, Christof and Becker, Rainer}, title = {MODELLBASIERTE ERTRAGSKONTROLLE F{\"U}R PV ANLAGEN IN VERNETZTEN GEB{\"A}UDEN}, editor = {G{\"u}rlebeck, Klaus and K{\"o}nke, Carsten}, organization = {Bauhaus-Universit{\"a}t Weimar}, doi = {10.25643/bauhaus-universitaet.3036}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170327-30363}, pages = {5}, abstract = {Subject of the paper is the realisation of a model based efficiency control system for PV generators using a simulation model. A standard 2-diodes model of PV generator is base of the ColSim model, which is implemented in ANSI C code for flexible code exporting. The algorithm is based on discretisized U-I characteristics, which allows the calculation of string topologies witch parallel and serial PV cells and modules. Shadowing effects can be modelled down to cell configuration using polar horizon definitions. The simulation model was ported to a real time environment, to calculate the efficiency of a PV system. Embedded System technology allows the networked operation and the integration of standard I/O devices. Futher work focus on the adaption of shadowing routine, which will be adapted to get the environment conditions from the real operation.}, subject = {Architektur }, language = {de} } @inproceedings{Geyer, author = {Geyer, Philipp}, title = {MODELS FOR MULTIDISCIPLINARY DESIGN OPTIMIZATION: AN EXEMPLARY OFFICE BUILDING}, editor = {G{\"u}rlebeck, Klaus and K{\"o}nke, Carsten}, organization = {Bauhaus-Universit{\"a}t Weimar}, doi = {10.25643/bauhaus-universitaet.2957}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170327-29574}, pages = {10}, abstract = {The mathematical and technical foundations of optimization have been developed to a large extent. In the design of buildings, however, optimization is rarely applied because of insufficient adaptation of this method to the needs of building design. The use of design optimization requires the consideration of all relevant objectives in an interactive and multidisciplinary process. Disciplines such as structural, light, and thermal engineering, architecture, and economics impose various objectives on the design. A good solution calls for a compromise between these often contradictory objectives. This presentation outlines a method for the application of Multidisciplinary Design Optimization (MDO) as a tool for the designing of buildings. An optimization model is established considering the fact that in building design the non-numerical aspects are of major importance than in other engineering disciplines. A component-based decomposition enables the designer to manage the non-numerical aspects in an interactive design optimization process. A fa{\c{c}}ade example demonstrates a way how the different disciplines interact and how the components integrate the disciplines in one optimization model. In this grid-based fa{\c{c}}ade example, the materials switch between a discrete number of materials and construction types. For light and thermal engineering, architecture, and economics, analysis functions calculate the performance; utility functions serve as an important means for the evaluation since not every increase or decrease of a physical value improves the design. For experimental purposes, a genetic algorithm applied to the exemplary model demonstrates the use of optimization in this design case. A component-based representation first serves to manage non-numerical characteristics such as aesthetics. Furthermore, it complies with usual fabrication methods in building design and with object-oriented data handling in CAD. Therefore, components provide an important basis for an interactive MDO process in building design.}, subject = {Architektur }, language = {en} } @inproceedings{ChangChang, author = {Chang, Wei-Tsang and Chang, Teng-Wen}, title = {TIME-BASED FORM TRANSFORMATION WITH FOLDING SPACE}, editor = {G{\"u}rlebeck, Klaus and K{\"o}nke, Carsten}, organization = {Bauhaus-Universit{\"a}t Weimar}, doi = {10.25643/bauhaus-universitaet.2937}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170327-29371}, pages = {10}, abstract = {Design activity could be treated as state transition computationally. In stepwise processing, in-between form-states are not easily observed. However, in this research time-based concept is introduced and applied in order to bridge the gap. In architecture, folding is one method of form manipulation and architects also want to search for alternatives by this operation. Besides, folding operation has to be defined and parameterized before time factor is involved as a variable of folding. As a result, time-based transformation provides sequential form states and redirects design activity.}, subject = {Architektur }, language = {en} } @inproceedings{Kisil, author = {Kisil, Vladimir}, title = {FILLMORE-SPRINGER-CNOPS CONSTRUCTION IMPLEMENTED IN GINAC}, editor = {G{\"u}rlebeck, Klaus and K{\"o}nke, Carsten}, organization = {Bauhaus-Universit{\"a}t Weimar}, doi = {10.25643/bauhaus-universitaet.2974}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170327-29744}, pages = {103}, abstract = {This is an implementation of the Fillmore-Springer-Cnops construction (FSCc) based on the Clifford algebra capacities of the GiNaC computer algebra system. FSCc linearises the linear-fraction action of the Mobius group. This turns to be very useful in several theoretical and applied fields including engineering. The core of this realisation of FSCc is done for an arbitrary dimension, while a subclass for two dimensional cycles add some 2D-specific routines including a visualisation to PostScript files through the MetaPost or Asymptote software. This library is a backbone of many result published in, which serve as illustrations of its usage. It can be ported (with various level of required changes) to other CAS with Clifford algebras capabilities.}, subject = {Architektur }, language = {en} } @inproceedings{Krasnov, author = {Krasnov, Yakov}, title = {ANALYTIC FUNCTIONS IN OPERATOR VARIABLES AS SOLUTION TO PDES}, editor = {G{\"u}rlebeck, Klaus and K{\"o}nke, Carsten}, organization = {Bauhaus-Universit{\"a}t Weimar}, doi = {10.25643/bauhaus-universitaet.2982}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170327-29822}, pages = {16}, abstract = {Procedures of a construction of general solutions for some classes of partial differential equations (PDEs) are proposed and a symmetry operators approach to the raising the orders of the polynomial solutions to linear PDEs are develops. We touch upon an ''operator analytic function theory'' as the solution of a frequent classes of the equations of mathematical physics, when its symmetry operators forms vast enough space. The MAPLE© package programs for the building the operator variables is elaborated also.}, subject = {Architektur }, language = {en} } @inproceedings{LehnerHartmann, author = {Lehner, Karlheinz and Hartmann, Dietrich}, title = {USING INTERVAL ANALYSIS FOR STRUCTURAL ENGINEERING PROBLEMS}, editor = {G{\"u}rlebeck, Klaus and K{\"o}nke, Carsten}, organization = {Bauhaus-Universit{\"a}t Weimar}, doi = {10.25643/bauhaus-universitaet.2984}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170327-29844}, pages = {10}, abstract = {Interval analysis extends the concept of computing with real numbers to computing with real intervals. As a consequence, some interesting properties appear, such as the delivery of guaranteed results or confirmed global values. The former property is given in the sense that unknown numerical values are in known to lie in a computed interval. The latter property states that the global minimum value, for example, of a given function is also known to be contained in a interval (or a finite set of intervals). Depending upon the amount computation effort invested in the calculation, we can often find tight bounds on these enclosing intervals. The downside of interval analysis, however, is the mathematically correct, but often very pessimistic size of the interval result. This is in particularly due to the so-called dependency effect, where a single variable is used multiple times in one calculation. Applying interval analysis to structural analysis problems, the dependency has a great influence on the quality of numerical results. In this paper, a brief background of interval analysis is presented and shown how it can be applied to the solution of structural analysis problems. A discussion of possible improvements as well as an outlook to parallel computing is also given.}, subject = {Architektur }, language = {en} } @inproceedings{LourensvanRooyen, author = {Lourens, Eliz-Mari and van Rooyen, G.C.}, title = {Automating Preliminary Column Force Calculations In Multy-Storey Buildings}, editor = {G{\"u}rlebeck, Klaus and K{\"o}nke, Carsten}, organization = {Bauhaus-Universit{\"a}t Weimar}, doi = {10.25643/bauhaus-universitaet.2986}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170327-29864}, pages = {10}, abstract = {In civil engineering practice, values of column forces are often required before any detailed analysis of the structure has been performed. One of the reasons for this arises from the fast-tracked nature of the majority of construction projects: foundations are laid and base columns constructed whilst analysis and design are still in progress. A need for quick results when feasibility studies are performed or when evaluating the effect of design changes on supporting columns form other situations in which column forces are required, but where a detailed analysis to get these forces seems superfluous. Thus it was concluded that the development of an efficient tool for column force calculations, in which the extensive input required in a finite element analysis is to be avoided, would be highly beneficial. The automation of the process is achieved by making use of a Voronoi diagram. The Voronoi diagram is used a) for subdividing the floor into influence areas and b) as a basis for automatic load assignment. The implemented procedure is integrated into a CAD system in which the relevant geometric information of the floor, i.e. its shape and column layout, can be defined or uploaded. A brief description of the implementation is included. Some comparative results and considerations regarding the continuation of the study are given.}, subject = {Architektur }, language = {en} } @inproceedings{Knyziak, author = {Knyziak, Piotr}, title = {ANALYSIS THE TECHNICAL STATE FOR LARGE-PANEL RESIDENTIAL BUILDINGS BEHIND ASSISTANCE OF ARTIFICIAL NEURAL NETWORKS}, editor = {G{\"u}rlebeck, Klaus and K{\"o}nke, Carsten}, organization = {Bauhaus-Universit{\"a}t Weimar}, doi = {10.25643/bauhaus-universitaet.2979}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170327-29792}, pages = {9}, abstract = {This paper presents two new methods for analysis of a technical state of large-panel residential buildings. The first method is based on elements extracted from the classical methods and on data about repairs and modernization collected from building documentations. The technical state of a building is calculated as a sum of several groups of elements defining the technical state. The deterioration in this method depends on: - time, which has passed since last repair of element or time which has passed since construction, - estimate of the state of element groups which can be determined on basis of yearly controls. This is a new unique method. it is easy to use, does not need expertise. The required data could be extracted easily from building documentations. For better accuracy the data from building inspections should be applied (in Poland inspections are made every year). The second method is based on the extracted data processing by means of the artificial neural networks. The aim is to learn the artificial neural network configurations for a set of data containing values of the technical state and information about building repairs for last years (or other information and building parameters) and next to analyse new buildings by the instructed neural network. The second profit from using artificial neural networks is the reduction of number of parameters. Instead of more then 40 parameters describing building, about 6-12 are usually sufficient for satisfactory accuracy. This method could have lower accuracy but it is less prone to data errors.}, subject = {Architektur }, language = {en} } @inproceedings{Kravchenko, author = {Kravchenko, Vladislav}, title = {NEW APPLICATIONS OF PSEUDOANALYTIC FUNCTION THEORY}, editor = {G{\"u}rlebeck, Klaus and K{\"o}nke, Carsten}, organization = {Bauhaus-Universit{\"a}t Weimar}, doi = {10.25643/bauhaus-universitaet.2983}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170327-29835}, pages = {3}, abstract = {We show a close relation between the Schr{\"o}dinger equation and the conductivity equation to a Vekua equation of a special form. Under quite general conditions we propose an algorithm for explicit construction of pseudoanalytic positive formal powers for the Vekua equation that as a consequence gives us a complete system of solutions for the Schr{\"o}dinger and the conductivity equations. Besides the construction of complete systems of exact solutions for the above mentioned second order equations and the Dirac equation, we discuss some other applications of pseudoanalytic function theory.}, subject = {Architektur }, language = {en} } @inproceedings{EygelaarvanRooyen, author = {Eygelaar, Anton and van Rooyen, G.C.}, title = {ENGINEERING PROCESS MODEL SPECIFICATION AND RESOURCE LEVELING}, editor = {G{\"u}rlebeck, Klaus and K{\"o}nke, Carsten}, organization = {Bauhaus-Universit{\"a}t Weimar}, doi = {10.25643/bauhaus-universitaet.2952}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170327-29529}, pages = {18}, abstract = {The use of process models in the analysis, optimization and simulation of processes has proven to be extremely beneficial in the instances where they could be applied appropriately. However, the Architecture/Engineering/Construction (AEC) industries present unique challenges that complicate the modeling of their processes. A simple Engineering process model, based on the specification of Tasks, Datasets, Persons and Tools, and certain relations between them, have been developed, and its advantages over conventional techniques have been illustrated. Graph theory is used as the mathematical foundation mapping Tasks, Datasets, Persons and Tools to vertices and the relations between them to edges forming a directed graph. The acceptance of process modeling in AEC industries not only depends on the results it can provide, but the ease at which these results can be attained. Specifying a complex AEC process model is a dynamic exercise that is characterized by many modifications over the process model's lifespan. This article looks at reducing specification complexity, reducing the probability for erroneous input and allowing consistent model modification. Furthermore, the problem of resource leveling is discussed. Engineering projects are often executed with limited resources and determining the impact of such restrictions on the sequence of Tasks is important. Resource Leveling concerns itself with these restrictions caused by limited resources. This article looks at using Task shifting strategies to find a near-optimal sequence of Tasks that guarantees consistent Dataset evolution while resolving resource restrictions.}, subject = {Architektur }, language = {en} } @inproceedings{Klauer, author = {Klauer, Thomas}, title = {MOBILE FACILITY MANAGEMENT ZUR INSPEKTION UND INSTANDHALTUNG VON INGENIEURBAUWERKEN}, editor = {G{\"u}rlebeck, Klaus and K{\"o}nke, Carsten}, organization = {Bauhaus-Universit{\"a}t Weimar}, doi = {10.25643/bauhaus-universitaet.2975}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170327-29759}, pages = {10}, abstract = {In diesem Beitrag wird eine mobile Software-Komponente zur Vor-Ort-Unterst{\"u}tzung von Bauwerkspr{\"u}fungen gem{\"a}ß DIN 1076 „Ingenieurbauwerke im Zuge von Strassen und Wegen, {\"U}berwachung und Pr{\"u}fung" vorgestellt, welche sich im praktischen Einsatz bei der Hochbahn AG Hamburg befindet. Mit Hilfe dieses Werkzeugs kann die Aktivit{\"a}t am Bauwerk in den gesamten softwaregest{\"u}tzten Gesch{\"a}ftsprozess der auwerksinstandhaltung integriert und somit die Bearbeitungszeit einer Bauwerkspr{\"u}fung von der Vorbereitung bis zur Pr{\"u}fbericht-Erstellung reduziert werden. Die Technologie des Mobile Computing wird unter Ber{\"u}cksichtigung spezieller fachlicher Randbedingungen, wie z.B. dem Einsatzort unter freiem Himmel, erl{\"a}utert und es werden Methoden zur effizienten Datenerfassung mit Stift und Sprache vorgestellt und bewertet. Ferner wird die Einschr{\"a}nkung der Hardware durch die geringere Gr{\"o}ße der Endger{\"a}te, die sich durch die Bedingung der Mobilit{\"a}t ergibt, untersucht.}, subject = {Architektur }, language = {de} } @inproceedings{BultheelJansenMaesetal., author = {Bultheel, Adhemar and Jansen, M. and Maes, J. and Van Aerschot, W. and Vanraes, E.}, title = {SUBDIVIDE AND CONQUER RESOLUTION}, editor = {G{\"u}rlebeck, Klaus and K{\"o}nke, Carsten}, organization = {Bauhaus-Universit{\"a}t Weimar}, doi = {10.25643/bauhaus-universitaet.2909}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170327-29091}, pages = {47}, abstract = {This contribution will be freewheeling in the domain of signal, image and surface processing and touch briefly upon some topics that have been close to the heart of people in our research group. A lot of the research of the last 20 years in this domain that has been carried out world wide is dealing with multiresolution. Multiresolution allows to represent a function (in the broadest sense) at different levels of detail. This was not only applied in signals and images but also when solving all kinds of complex numerical problems. Since wavelets came into play in the 1980's, this idea was applied and generalized by many researchers. Therefore we use this as the central idea throughout this text. Wavelets, subdivision and hierarchical bases are the appropriate tools to obtain these multiresolution effects. We shall introduce some of the concepts in a rather informal way and show that the same concepts will work in one, two and three dimensions. The applications in the three cases are however quite different, and thus one wants to achieve very different goals when dealing with signals, images or surfaces. Because completeness in our treatment is impossible, we have chosen to describe two case studies after introducing some concepts in signal processing. These case studies are still the subject of current research. The first one attempts to solve a problem in image processing: how to approximate an edge in an image efficiently by subdivision. The method is based on normal offsets. The second case is the use of Powell-Sabin splines to give a smooth multiresolution representation of a surface. In this context we also illustrate the general method of construction of a spline wavelet basis using a lifting scheme.}, subject = {Architektur }, language = {en} } @inproceedings{KlawitterOstrowski, author = {Klawitter, Arne and Ostrowski, M.}, title = {INTEGRATED RAINFALL RUNOFF MODELLING IN SMALL URBANIZED CATCHMENTS}, editor = {G{\"u}rlebeck, Klaus and K{\"o}nke, Carsten}, organization = {Bauhaus-Universit{\"a}t Weimar}, doi = {10.25643/bauhaus-universitaet.2976}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170327-29767}, pages = {12}, abstract = {A concept for integrated modeling of urban and rural hydrology is introduced. The concept allows for simulations on the catchment scale as well as on the local scale. It is based on a 2-layer-approach which facilitates the parallel coupling of a catchment hydrology model with an urban hydrology model, considering the interactions between the two systems. The concept has been implemented in a computer model combining a grid based distributed hydrological catchment model and a hydrological urban stormwater model based on elementary units. The combined model provides a flexible solution for time and spatial scale integration and offers to calculate separate water balances for urban and rural hydrology. Furthermore, it is GIS-based which allows for easy and accurate geo-referencing of urban overflow structures, which are considered as points of interactions between the two hydrologic systems. Due to the two-layer-approach, programs of measures can be incorporated in each system separately. The capabilities of the combined model have been tested on a hypothetical test case and a real world application. It could be shown that the model is capable of accurately quantifying the effects of urbanization in a catchment. The affects of urbanization can be analyzed at the catchment outlet, but can also be traced back to its origins, due to the geo-referencing of urban overflow structures. This is a mayor advantage over conventional hydrological catchment models for the analysis of land use changes.}, subject = {Architektur }, language = {en} } @inproceedings{BeranDlask, author = {Beran, V{\´a}clav and Dlask, Petr}, title = {CONSTRUCTION SPEED AND CASH FLOW OPTIMISATION}, editor = {G{\"u}rlebeck, Klaus and K{\"o}nke, Carsten}, organization = {Bauhaus-Universit{\"a}t Weimar}, doi = {10.25643/bauhaus-universitaet.2926}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170327-29269}, pages = {10}, abstract = {Practical examples show that the improvement in cost flow and total amount of money spend in construction and further use may be cut significantly. The calculation is based on spreadsheets calculation, very easy to develop on most PC´s now a days. Construction works, are a field where the evaluation of Cash Flow can be and should be applied. Decisions about cash flow in construction are decisions with long-term impact and long-term memory. Mistakes from the distant past have a massive impact on situations in the present and into the far economic future of economic activities. Two approaches exist. The Just-in-Time (JIT) approach and life cycle costs (LCC) approach. The calculation example shows the dynamic results for the production speed in opposition to stable flow of production in duration of activities. More sophisticated rescheduling in optimal solution might bring in return extra profit. In the technologies and organizational processes for industrial buildings, railways and road reconstruction, public utilities and housing developments there are assembly procedures that are very appropriate for the given purpose, complicated research-, development-, innovation-projects are all very good aspects of these kinds of applications. The investors of large investments and all public invested money may be spent more efficiently if an optimisation speed-strategy can be calculated.}, subject = {Architektur }, language = {en} }