### Refine

#### Institute

- In Zusammenarbeit mit der Bauhaus-Universität Weimar (92) (remove)

#### Keywords

- Computerunterstütztes Verfahren (92) (remove)

#### Year of publication

- 2006 (92) (remove)

Practical examples show that the improvement in cost flow and total amount of money spend in construction and further use may be cut significantly. The calculation is based on spreadsheets calculation, very easy to develop on most PC´s now a days. Construction works, are a field where the evaluation of Cash Flow can be and should be applied. Decisions about cash flow in construction are decisions with long-term impact and long-term memory. Mistakes from the distant past have a massive impact on situations in the present and into the far economic future of economic activities. Two approaches exist. The Just-in-Time (JIT) approach and life cycle costs (LCC) approach. The calculation example shows the dynamic results for the production speed in opposition to stable flow of production in duration of activities. More sophisticated rescheduling in optimal solution might bring in return extra profit. In the technologies and organizational processes for industrial buildings, railways and road reconstruction, public utilities and housing developments there are assembly procedures that are very appropriate for the given purpose, complicated research-, development-, innovation-projects are all very good aspects of these kinds of applications. The investors of large investments and all public invested money may be spent more efficiently if an optimisation speed-strategy can be calculated.

Objects for civil engineering applications can be identified with their reference in memory, their alpha-numeric name or their geometric location. Particularly in graphic user interfaces, it is common to identify objects geometrically by selection with the mouse. As the number of geometric objects in a graphic user interface grows, it becomes increasingly more important to treat the basic operations add, search and remove for geometric objects with great efficiency. Guttmann has proposed the Region-Tree (R-tree) for geometric identification in an environment which uses pages on disc as data structure. Minimal bounding rectangles are used to structure the data in such a way that neighborhood relations can be described effectively. The literature shows that the parameters which influence the efficiency of the R-trees have been studied extensively, but without conclusive results. The goal of the research which is reported in this paper is to determine reliably the parameters which significantly influence the efficiency of R-trees for geometric identification in technical drawings. In order to make this investigation conclusive, it must be performed with the best available software technology. Therefore an object-oriented software for the method is developed. This implementation is tested with technical drawings containing many thousands of geometric objects. These drawings are created automatically by a stochastic generator which is incorporated into a test bed consisting of an editor and a visualisor. This test bed is used to obtain statistics for the main factors which affect the efficiency of R-trees. The investigation shows that the following main factors which affect the efficiency can be identified reliably : number of geometric objects on the drawing the minimum und maximum number of children of a node of the tree the maximum width and height of the minimal bounding rectangles of the geometric objects relative to the size of the drawing.

In earlier research, generalized multidimensional Hilbert transforms have been constructed in m-dimensional Euclidean space, in the framework of Clifford analysis. Clifford analysis, centred around the notion of monogenic functions, may be regarded as a direct and elegant generalization to higher dimension of the theory of the holomorphic functions in the complex plane. The considered Hilbert transforms, usually obtained as a part of the boundary value of an associated Cauchy transform in m+1 dimensions, might be characterized as isotropic, since the metric in the underlying space is the standard Euclidean one. In this paper we adopt the idea of a so-called anisotropic Clifford setting, which leads to the introduction of a metric dependent m-dimensional Hilbert transform, showing, at least formally, the same properties as the isotropic one. The Hilbert transform being an important tool in signal analysis, this metric dependent setting has the advantage of allowing the adjustment of the co-ordinate system to possible preferential directions in the signals to be analyzed. A striking result to be mentioned is that the associated anisotropic (m+1)-dimensional Cauchy transform is no longer uniquely determined, but may stem from a diversity of (m+1)-dimensional "mother" metrics.

Image processing has been much inspired by the human vision, in particular with regard to early vision. The latter refers to the earliest stage of visual processing responsible for the measurement of local structures such as points, lines, edges and textures in order to facilitate subsequent interpretation of these structures in higher stages (known as high level vision) of the human visual system. This low level visual computation is carried out by cells of the primary visual cortex. The receptive field profiles of these cells can be interpreted as the impulse responses of the cells, which are then considered as filters. According to the Gaussian derivative theory, the receptive field profiles of the human visual system can be approximated quite well by derivatives of Gaussians. Two mathematical models suggested for these receptive field profiles are on the one hand the Gabor model and on the other hand the Hermite model which is based on analysis filters of the Hermite transform. The Hermite filters are derivatives of Gaussians, while Gabor filters, which are defined as harmonic modulations of Gaussians, provide a good approximation to these derivatives. It is important to note that, even if the Gabor model is more widely used than the Hermite model, the latter offers some advantages like being an orthogonal basis and having better match to experimental physiological data. In our earlier research both filter models, Gabor and Hermite, have been developed in the framework of Clifford analysis. Clifford analysis offers a direct, elegant and powerful generalization to higher dimension of the theory of holomorphic functions in the complex plane. In this paper we expose the construction of the Hermite and Gabor filters, both in the classical and in the Clifford analysis framework. We also generalize the concept of complex Gaussian derivative filters to the Clifford analysis setting. Moreover, we present further properties of the Clifford-Gabor filters, such as their relationship with other types of Gabor filters and their localization in the spatial and in the frequency domain formalized by the uncertainty principle.

The one-dimensional continuous wavelet transform is a successful tool for signal and image analysis, with applications in physics and engineering. Clifford analysis offers an appropriate framework for taking wavelets to higher dimension. In the usual orthogonal case Clifford analysis focusses on monogenic functions, i.e. null solutions of the rotation invariant vector valued Dirac operator ∂, defined in terms of an orthogonal basis for the quadratic space Rm underlying the construction of the Clifford algebra R0,m. An intrinsic feature of this function theory is that it encompasses all dimensions at once, as opposed to a tensorial approach with products of one-dimensional phenomena. This has allowed for a very specific construction of higher dimensional wavelets and the development of the corresponding theory, based on generalizations of classical orthogonal polynomials on the real line, such as the radial Clifford-Hermite polynomials introduced by Sommen. In this paper, we pass to the Hermitian Clifford setting, i.e. we let the same set of generators produce the complex Clifford algebra C2n (with even dimension), which we equip with a Hermitian conjugation and a Hermitian inner product. Hermitian Clifford analysis then focusses on the null solutions of two mutually conjugate Hermitian Dirac operators which are invariant under the action of the unitary group. In this setting we construct new Clifford-Hermite polynomials, starting in a natural way from a Rodrigues formula which now involves both Dirac operators mentioned. Due to the specific features of the Hermitian setting, four different types of polynomials are obtained, two types of even degree and two types of odd degree. These polynomials are used to introduce a new continuous wavelet transform, after thorough investigation of all necessary properties of the involved polynomials, the mother wavelet and the associated family of wavelet kernels.

In this paper we study the structure of the solutions to higher dimensional Dirac type equations generalizing the known λ-hyperholomorphic functions, where λ is a complex parameter. The structure of the solutions to the system of partial differential equations (D- λ) f=0 show a close connection with Bessel functions of first kind with complex argument. The more general system of partial differential equations that is considered in this paper combines Dirac and Euler operators and emphasizes the role of the Bessel functions. However, contrary to the simplest case, one gets now Bessel functions of any arbitrary complex order.

Design activity could be treated as state transition computationally. In stepwise processing, in-between form-states are not easily observed. However, in this research time-based concept is introduced and applied in order to bridge the gap. In architecture, folding is one method of form manipulation and architects also want to search for alternatives by this operation. Besides, folding operation has to be defined and parameterized before time factor is involved as a variable of folding. As a result, time-based transformation provides sequential form states and redirects design activity.

ON THE NAVIER-STOKES EQUATION WITH FREE CONVECTION IN STRIP DOMAINS AND 3D TRIANGULAR CHANNELS
(2006)

The Navier-Stokes equations and related ones can be treated very elegantly with the quaternionic operator calculus developed in a series of works by K. Guerlebeck, W. Sproeossig and others. This study will be extended in this paper. In order to apply the quaternionic operator calculus to solve these types of boundary value problems fully explicitly, one basically needs to evaluate two types of integral operators: the Teodorescu operator and the quaternionic Bergman projector. While the integral kernel of the Teodorescu transform is universal for all domains, the kernel function of the Bergman projector, called the Bergman kernel, depends on the geometry of the domain. With special variants of quaternionic holomorphic multiperiodic functions we obtain explicit formulas for three dimensional parallel plate channels, rectangular block domains and regular triangular channels. The explicit knowledge of the integral kernels makes it then possible to evaluate the operator equations in order to determine the solutions of the boundary value problem explicitly.

In this paper we consider three different methods for generating monogenic functions. The first one is related to Fueter's well known approach to the generation of monogenic quaternion-valued functions by means of holomorphic functions, the second one is based on the solution of hypercomplex differential equations and finally the third one is a direct series approach, based on the use of special homogeneous polynomials. We illustrate the theory by generating three different exponential functions and discuss some of their properties. Formula que se usa em preprints e artigos da nossa UI&D (acho demasiado completo): Partially supported by the R\&D unit \emph{Matem\'atica a Aplica\c\~es} (UIMA) of the University of Aveiro, through the Portuguese Foundation for Science and Technology (FCT), co-financed by the European Community fund FEDER.

In classical complex function theory the geometric mapping property of conformality is closely linked with complex differentiability. In contrast to the planar case, in higher dimensions the set of conformal mappings is only the set of Möbius transformations. Unfortunately, the theory of generalized holomorphic functions (by historical reasons they are called monogenic functions) developed on the basis of Clifford algebras does not cover the set of Möbius transformations in higher dimensions, since Möbius transformations are not monogenic. But on the other side, monogenic functions are hypercomplex differentiable functions and the question arises if from this point of view they can still play a special role for other types of 3D-mappings, for instance, for quasi-conformal ones. On the occasion of the 16th IKM 3D-mapping methods based on the application of Bergman's reproducing kernel approach (BKM) have been discussed. Almost all authors working before that with BKM in the Clifford setting were only concerned with the general algebraic and functional analytic background which allows the explicit determination of the kernel in special situations. The main goal of the abovementioned contribution was the numerical experiment by using a Maple software specially developed for that purpose. Since BKM is only one of a great variety of concrete numerical methods developed for mapping problems, our goal is to present a complete different from BKM approach to 3D-mappings. In fact, it is an extension of ideas of L. V. Kantorovich to the 3-dimensional case by using reduced quaternions and some suitable series of powers of a small parameter. Whereas until now in the Clifford case of BKM the recovering of the mapping function itself and its relation to the monogenic kernel function is still an open problem, this approach avoids such difficulties and leads to an approximation by monogenic polynomials depending on that small parameter.

Designing a structure follows a pattern of creating a structural design concept, executing a finite element analysis and developing a design model. A project was undertaken to create computer support for executing these tasks within a collaborative environment. This study focuses on developing a software architecture that integrates the various structural design aspects into a seamless functional collaboratory that satisfies engineering practice requirements. The collaboratory is to support both homogeneous collaboration i.e. between users operating on the same model and heterogeneous collaboration i.e. between users operating on different model types. Collaboration can take place synchronously or asynchronously, and the information exchange is done either at the granularity of objects or at the granularity of models. The objective is to determine from practicing engineers which configurations they regard as best and what features are essential for working in a collaborative environment. Based on the suggestions of these engineers a specification of a collaboration configuration that satisfies engineering practice requirements will be developed.

Reasonably accurate cost estimation of the structural system is quite desirable at the early stages of the design process of a construction project. However, the numerous interactions among the many cost-variables make the prediction difficult. Artificial neural networks (ANN) and case-based reasoning (CBR) are reported to overcome this difficulty. This paper presents a comparison of CBR and ANN augmented by genetic algorithms (GA) conducted by using spreadsheet simulations. GA was used to determine the optimum weights for the ANN and CBR models. The cost data of twenty-nine actual cases of residential building projects were used as an example application. Two different sets of cases were randomly selected from the data set for training and testing purposes. Prediction rates of 84% in the GA/CBR study and 89% in the GA/ANN study were obtained. The advantages and disadvantages of the two approaches are discussed in the light of the experiments and the findings. It appears that GA/ANN is a more suitable model for this example of cost estimation where the prediction of numerical values is required and only a limited number of cases exist. The integration of GA into CBR and ANN in a spreadsheet format is likely to improve the prediction rates.

Der Begriff der Zuverlässigkeit spielt eine zentrale Rolle bei der Bewertung von Verkehrsnetzen. Aus der Sicht der Nutzer des öffentlichen Personennahverkehrs (ÖPNV) ist eines der wichtigsten Kriterien zur Beurteilung der Qualität des Liniennetzes, ob es möglich ist, mit einer großen Sicherheit das Reiseziel in einer vorgegebenen Zeit zu erreichen. Im Vortrag soll dieser Zuverlässigkeitsbegriff mathematisch gefasst werden. Dabei wird zunächst auf den üblichen Begriff der Zuverlässigkeit eines Netzes im Sinne paarweiser Zusammenhangswahrscheinlichkeiten eingegangen. Dieser Begriff wird erweitert durch die Betrachtung der Zuverlässigkeit unter Einbeziehung einer maximal zulässigen Reisezeit. In vergangenen Arbeiten hat sich die Ring-Radius-Struktur als bewährtes Modell für die theoretische Beschreibung von Verkehrsnetzen erwiesen. Diese Überlegungen sollen nun durch Einbeziehung realer Verkehrsnetzstrukturen erweitert werden. Als konkretes Beispiel dient das Straßenbahnnetz von Krakau. Hier soll insbesondere untersucht werden, welche Auswirkungen ein geplanter Ausbau des Netzes auf die Zuverlässigkeit haben wird. This paper is involved with CIVITAS-CARAVEL project: "Clean and better transport in cites". The project has received research funding from the Community's Sixth Framework Programme. The paper reflects only the author's views and the Community is not liable for any use that may be made of the information contained therein.

Wir betrachten im ÖPNV (Öffentlichen Personennahverkehr) diejenige Situation, daß zwei Bus- oder Straßenbahnlinien gemeinsame Haltestellen haben. Ziel unserer Untersuchungen ist es, für beide Linien einen solchen Fahrplan zu finden, der für die Fahrgäste möglichst viel Bequemlichkeit bietet. Die Bedarfsstruktur - die Anzahl von Personen, die die beiden Linien benutzen - setzt dabei gewisse Beschränkungen für die Taktzeiten der beiden Linien. Die verbleibenden Entscheidungsfreiheiten sollen im Sinne der Zielstellung ausgenutzt werden. Im Vortrag wird folgenden Fragen nachgegangen: - nach welchen Kriterien kann man die "Bequemlichkeit" oder die "Synchonisationsgüte" messen? - wie kann man die einzelnen "Synchronisationsmaße" berechnen ? - wie kann man die verbleibenden Entscheidungsfreiheiten nutzen, um eine möglichst gute Synchronisation zu erreichen ? Die Ergebnisse werden dann auf einige Beispiele angewandt und mit den bereitgestellten Methoden Lösungsvorschläge unterbreitet.

In civil engineering it is very difficult and often expensive to excite constructions such as bridges and buildings with an impulse hammer or shaker. This problem can be avoided with the output-only method as special feature of stochastic system identification. The permanently existing ambient noise (e.g. wind, traffic, waves) is sufficient to excite the structures in their operational conditions. The output-only method is able to estimate the observable part of a state-space-model which contains the dynamic characteristics of the measured mechanical system. Because of the assumption that the ambient excitation is white there is no requirement to measure the input. Another advantage of the output-only method is the possibility to get high detailed models by a special method, called polyreference setup. To pretend the availability of a much larger set of sensors the data from varying sensor locations will be collected. Several successive data sets are recorded with sensors at different locations (moving sensors) and fixed locations (reference sensors). The covariance functions of the reference sensors are bases to normalize the moving sensors. The result of the following subspace-based system identification is a high detailed black-box-model that contains the weighting function including the well-known dynamic parameters eigenfrequencies and mode shapes of the mechanical system. Emphasis of this lecture is the presentation of an extensive damage detection experiment. A 53-year old prestressed concrete tied-arch-bridge in Hünxe (Germany) was deconstructed in 2005. Preliminary numerous vibration measurements were accomplished. The first experiment for system modification was an additional support near the bridge bearing of one main girder. During a further experiment one hanger from one tied arch was cut through as an induced damage. Some first outcomes of the described experiments will be presented.

Due to the amount of flow simulation and measurement data, automatic detection, classification and visualization of features is necessary for an inspection. Therefore, many automated feature detection methods have been developed in recent years. However, only one feature class is visualized afterwards in most cases, and many algorithms have problems in the presence of noise or superposition effects. In contrast, image processing and computer vision have robust methods for feature extraction and computation of derivatives of scalar fields. Furthermore, interpolation and other filter can be analyzed in detail. An application of these methods to vector fields would provide a solid theoretical basis for feature extraction. The authors suggest Clifford algebra as a mathematical framework for this task. Clifford algebra provides a unified notation for scalars and vectors as well as a multiplication of all basis elements. The Clifford product of two vectors provides the complete geometric information of the relative positions of these vectors. Integration of this product results in Clifford correlation and convolution which can be used for template matching of vector fields. For frequency analysis of vector fields and the behavior of vector-valued filters, a Clifford Fourier transform has been derived for 2D and 3D. Convolution and other theorems have been proved, and fast algorithms for the computation of the Clifford Fourier transform exist. Therefore the computation of Clifford convolution can be accelerated by computing it in Clifford Fourier domain. Clifford convolution and Fourier transform can be used for a thorough analysis and subsequent visualization of flow fields.

Solid behavior as well as liquid behavior characterizes the flow of granular material in silos. The presented model is based on an appropriate interaction of a displacement field and a velocity field. The constitutive equations and the applied algorithm are developed from the exact solution for a standard case. The standard case evolves from a very tall vertical plane strain silo containing material that flows at a constant speed. No horizontal displacements and velocities take place. No changes regarding the field values arise in the vertical direction and in time. Tension is not allowed at any point. Coulomb friction represents the effects of the vertical walls. The interaction between the flowing material and the walls is covered by a forced boundary condition resulting in an additional matrix for the solid component as well as for the liquid component. The resulting integral equations are designed to be solved directly. Three coefficients describe the properties of the granular material. They govern elastic solid behavior in combination with viscous liquid behavior.

We propose a new approach to the numerical solution of quasi-static elastic-plastic problems based on the Moreau-Yosida theorem. After the time discretization, the problem is expressed as an energy minimization problem for unknown displacement and plastic strain fields. The dependency of the minimization functional on the displacement is smooth whereas the dependency on the plastic strain is non-smooth. Besides, there exists an explicit formula, how to calculate the plastic strain from a given displacement field. This allows us to reformulate the original problem as a minimization problem in the displacement only. Using the Moreau-Yosida theorem from the convex analysis, the minimization functional in the displacements turns out to be Frechet-differentiable, although the hidden dependency on the plastic strain is non-differentiable. The seconds derivative exists everywhere apart from the elastic-plastic interface dividing elastic and plastic zones of the continuum. This motivates to implement a Newton-like method, which converges super-linearly as can be observed in our numerical experiments.

RESEARCH OF DEFORMATION OF MULTILAYERED PLATES ON UNDEFORMABLE BASIS BY UNFLEXURAL SPECIFIED MODEL
(2006)

Stress-strain state (SSS) of multilayered plates on undeformable foundation is investigated. The settlement circuit of transverse loaded plate is formed by symmetrical attaching of a plate concerning a surface of contact to the foundation. The plate of the double thickness becomes bilateral symmetrically loaded concerning its median surface. It allows to model only unflexural deformation that reduces amount of unknown and the general order of differentiation of resolving system of the equations. The developed refined continual model takes into account deformations of transverse shear and transverse compression in high iterative approximation. Rigid contact between the foundation and a plate, and also shear without friction on a surface of contact of a plate with the foundation is considered. Calculations confirm efficiency of such approach, allowing to receive decisions which is qualitative and quantitatively close to three-dimensional solutions.

Adopting the European laws concerning environmental protection will require sustained efforts of the authorities and communities from Romania; implementing modern solutions will become a fast and effective option for the improvement of the functioning systems, in order to prevent disasters. As a part of the urban infrastructure, the drainage networks of pluvial and residual waters are included in the plan of promoting the systems which protect the environmental quality, with the purpose of integrated and adaptive management. The paper presents a distributed control system for sewer network of Iasi town. Unsatisfactory technical state of the actual sewer system is exposed, focusing on objectives related to implementation of the control system. The proposed distributed control system of Iasi drainage network is based on the implementation of the hierarchic control theory for diagnose, sewer planning and management. There are proposed two control levels: coordinating and local execution. Configuration of the distributed control system, including data acquisition and conversion equipment, interface characteristics, local data bus, data communication network, station configuration are widely described. The project wish to be an useful instrument for the local authorities in the preventing and reducing the impact of future natural disasters over the urban areas by means of modern technologies.

The concrete is modeled as a material with damage and plasticity, whereat the viscoplastic and the viscoelastic behaviour depends on the rate of the total strains. Due to the damage behaviour the compliance tensor develops different properties in tension and compression. There have been tested various yield surfaces and flow rules, damage rules respectively to their usability in a concrete model. One three-dimensional yield surface was developed from a failure surface based on the Willam--Warnke five-parameter model by the author. Only one general uni-axial stress-strain-relation is used for the numeric control of the yield surface. From that curve all necessary parameters for different strengths of concrete and different strain rates can be derived by affine transformations. For the flow rule in the compression zone a non associated inelastic potential is used, in the tension zone a Rankine potential. Conditional on the time-dependent formulation, the symmetry of the system equations is maintained in spite of the usage of non-associated potentials for the derivation of the inelastic strains. In case of quasi statical computations a simple viscoplastic law is used that is rested on an approach to Perzyna. The principle of equality of dissipation power in the uni-axial and the three-axial state of stress is used. It is modified by a factor that depends on the actual stress ratio and in comparison with the Kupfer experiments it implicates strains that are more realistic. The implementation of the concrete model is conducted in a mixed hybrid finite element. Examples in the structural level are introduced for verification of the concrete model.

Die Liquiditätsplanung von Bauunternehmen XE "Liquiditätsplanung" gilt als ein wesentliches Steuerungs-, Kontroll- sowie Informationsinstrument für interne und externe Adressaten und übt eine Entscheidungsunterstützungsfunktion aus. Da die einzelnen Bauprojekte einen wesentlichen Anteil an den Gesamtkosten des Unternehmens ausmachen, besitzen diese auch einen erheblichen Einfluß auf die Liquidität und die Zahlungsfähigkeit der Bauunternehmung. Dem folgend ist es in der Baupraxis eine übliche Verfahrensweise, die Liquiditätsplanung zuerst projektbezogen zu erstellen und anschließend auf Unternehmensebene zu verdichten. Ziel der Ausführungen ist es, die Zusammenhänge von Arbeitskalkulation XE "Arbeitskalkulation" , Ergebnisrechnung XE "Ergebnisrechnung" und Finanzrechnung XE "Finanzrechnung" in Form eines deterministischen XE "Erklärungsmodells" Planungsmodells auf Projektebene darzustellen. Hierbei soll das Verständnis und die Bedeutung der Verknüpfungen zwischen dem technisch-orientierten Bauablauf und dessen Darstellung im Rechnungs- und Finanzwesen herausgestellt werden. Die Vorgänge aus der Bauabwicklung, das heißt die Abarbeitung der Bauleistungsverzeichnispositionen und deren zeitliche Darstellung in einem Bauzeitenplan sind periodisiert in Größen der Betriebsbuchhaltung (Leistung, Kosten) zu transformieren und anschließend in der Finanzrechnung (Einzahlungen., Auszahlungen) nach Kreditoren und Debitoren aufzuschlüsseln.

We consider efficient numerical methods for the solution of partial differential equations with stochastic coefficients or right hand side. The discretization is performed by the stochastic finite element method (SFEM). Separation of spatial and stochastic variables in the random input data is achieved via a Karhunen-Loève expansion or Wiener's polynomial chaos expansion. We discuss solution strategies for the Galerkin system that take advantage of the special structure of the system matrix. For stochastic coefficients linear in a set of independent random variables we employ Krylov subspace recycling techniques after having decoupled the large SFEM stiffness matrix.

HYPERMONOGENIC POLYNOMIALS
(2006)

It is well know that the power function is not monogenic. There are basically two ways to include the power function into the set of solutions: The hypermonogenic functions or holomorphic Cliffordian functions. L. Pernas has found out the dimension of the space of homogenous holomorphic Cliffordian polynomials of degree m, but his approach did not include a basis. It is known that the hypermonogenic functions are included in the space of holomorphic Cliffordian functions. As our main result we show that we can construct a basis for the right module of homogeneous holomorphic Cliffordian polynomials of degree m using hypermonogenic polynomials and their derivatives. To that end we first recall the function spaces of monogenic, hypermonogenic and holomorphic Cliffordian functions and give the results needed in the proof of our main theorem. We list some basic polynomials and their properties for the various function spaces. In particular, we consider recursive formulas, rules of differentiation and properties of linear independency for the polynomials.

Traffic simulation is a valuable tool for the design and evaluation of road networks. Over the years, the level of detail to which urban and freeway traffic can be simulated has increased steadily, shifting from a merely qualitative macroscopic perspective to a very detailed microscopic view, where the behavior of individual vehicles is emulated realistically. With the improvement of behavioral models, however, the computational complexity has also steadily increased, as more and more aspects of real-life traffic have to be considered by the simulation environment. Despite the constant increase in computing power of modern personal computers, microscopic simulation stays computationally expensive, limiting the maximum network size than can be simulated on a single-processor computer in reasonable time. Parallelization can distribute the computing load from a single computer system to a cluster of several computing nodes. To this end, the exisiting simulation framework had to be adapted to allow for a distributed approach. As the simulation is ultimately targeted to be executed in real-time, incorporating real traffic data, only a spatial partition of the simulation was considered, meaning the road network has to be partitioned into subnets of comparable complexity, to ensure a homogenous load balancing. The partition process must also ensure, that the division between subnets does only occur in regions, where no strong interaction between the separated road segments occurs (i.e. not in the direct vicinity of junctions). In this paper, we describe a new microscopic reasoning voting strategy, and discuss in how far the increasing computational costs of these more complex behaviors lend themselves to a parallelized approach. We show the parallel architecture employed, the communication between computing units using MPIJava, and the benefits and pitfalls of adapting a single computer application to be used on a multi-node computing cluster.

The use of process models in the analysis, optimization and simulation of processes has proven to be extremely beneficial in the instances where they could be applied appropriately. However, the Architecture/Engineering/Construction (AEC) industries present unique challenges that complicate the modeling of their processes. A simple Engineering process model, based on the specification of Tasks, Datasets, Persons and Tools, and certain relations between them, have been developed, and its advantages over conventional techniques have been illustrated. Graph theory is used as the mathematical foundation mapping Tasks, Datasets, Persons and Tools to vertices and the relations between them to edges forming a directed graph. The acceptance of process modeling in AEC industries not only depends on the results it can provide, but the ease at which these results can be attained. Specifying a complex AEC process model is a dynamic exercise that is characterized by many modifications over the process model's lifespan. This article looks at reducing specification complexity, reducing the probability for erroneous input and allowing consistent model modification. Furthermore, the problem of resource leveling is discussed. Engineering projects are often executed with limited resources and determining the impact of such restrictions on the sequence of Tasks is important. Resource Leveling concerns itself with these restrictions caused by limited resources. This article looks at using Task shifting strategies to find a near-optimal sequence of Tasks that guarantees consistent Dataset evolution while resolving resource restrictions.

At the 16th IKM Bock, Falcão and Gürlebeck presented examples of the application of some specially developed Maple-Software in hypercomplex analysis. Other papers of those authors continued this work and showed the efficiency of such tools for concrete numerical calculations as well as for numerical experiments, supporting the detection of new relationships and even theorems in a highly technical theoretical work. The mentioned software has been developed mainly for the use on mapping problems in the Euclidean spaces of dimension 3 and 4 by means of Bergman kernel methods (BKM), which are related to monogenic functions as solutions of generalized Cauchy-Riemann equations with respect to the Euclidean metric (Riesz system). The developed procedures concerning generalized powers of totally regular variables and the corresponding homogeneous polynomials basically rely on results and conventions introduced in the paper "Power series representation for monogenic functions in Rm+1 based on a permutational product", Complex Variables, 15, No.3, 181-191 (1990) by H. Malonek. Since 1992 H. Leutwiler, S. L. Eriksson and others developed in a number of papers a modified Clifford analysis and, particularly, a modified quaternionic analysis. The modification mainly consists in considering generalized Cauchy-Riemann equations with respect to a hyperbolic metric in a half space. The aim of this contribution is to show how through a change of the basic combinatorial relations used in the modified quaternionic analysis the aforementioned Maple-software (that has been recently published on CD-Rom as integrated part of the text book "Funktionentheorie in der Ebene und im Raum" by K. Gürlebeck, K. Habetha, and W. Sprössig, in the series "Grundstudium Mathematik" of Birkhäuser Verlag, 2006) can directly be used for numerical calculations in the modified theory.

We establish the basis of a discrete function theory starting with a Fischer decomposition for difference Dirac operators. Discrete versions of homogeneous polynomials, Euler and Gamma operators are obtained. As a consequence we obtain a Fischer decomposition for the discrete Laplacian. For the sake of simplicity we consider in the first part only Dirac operators which contain only forward or backward finite differences. Of course, these Dirac operators do not factorize the classic discrete Laplacian. Therefore, we will consider a different definition of a difference Dirac operator in the quaternionic case which do factorizes the discrete Laplacian.

LIFETIME-ORIENTED OPTIMIZATION OF BRIDGE TIE RODS EXPOSED TO VORTEX-INDUCED ACROSS-WIND VIBRATIONS
(2006)

In recent years, damages in welded connections plates of vertical tie rods of several arched steel bridges have been reported. These damages are due to fatigue caused by wind-induced vibrations. In the present study, such phenomena are examined, and the corresponding lifetime of a reference bridge in Münster-Hiltrup, Germany, is estimated, based on the actual shape of the connection plate. Also, the results obtained are compared to the expected lifetime of a connection plate, whose geometry has been optimized separately. The structural optimization, focussing on the shape of the cut at the hanger ends, has been carried out using evolution strategies. The oscillation amplitudes have been computed by means of the Newmark-Wilson time-step method, using an appropriate load model, which has been validated by on-site experiments on the selected reference bridge. Corresponding stress-amplitudes are evaluated by multiplying the oscillation amplitudes with a stress concentration factor. This factor has been computed on the basis of a finite element model of the system "hanger-welding-connection plate", applying solid elements, according to the notch stress approach. The damage estimation takes into account the stochastics of the exciting wind process, as well as the stochastics of the material parameters (fatigue strength) given in terms of Woehler-curves. The shape optimization results in a substantial increase of the estimated hanger lifetime. The comparison of the lifetimes of the bulk plate and of the welding revealed that, in the optimized structure, the welding, being the most sensitive part in the original structure, shows much more resistance against potential damages than the bulk material.

The mathematical and technical foundations of optimization have been developed to a large extent. In the design of buildings, however, optimization is rarely applied because of insufficient adaptation of this method to the needs of building design. The use of design optimization requires the consideration of all relevant objectives in an interactive and multidisciplinary process. Disciplines such as structural, light, and thermal engineering, architecture, and economics impose various objectives on the design. A good solution calls for a compromise between these often contradictory objectives. This presentation outlines a method for the application of Multidisciplinary Design Optimization (MDO) as a tool for the designing of buildings. An optimization model is established considering the fact that in building design the non-numerical aspects are of major importance than in other engineering disciplines. A component-based decomposition enables the designer to manage the non-numerical aspects in an interactive design optimization process. A façade example demonstrates a way how the different disciplines interact and how the components integrate the disciplines in one optimization model. In this grid-based façade example, the materials switch between a discrete number of materials and construction types. For light and thermal engineering, architecture, and economics, analysis functions calculate the performance; utility functions serve as an important means for the evaluation since not every increase or decrease of a physical value improves the design. For experimental purposes, a genetic algorithm applied to the exemplary model demonstrates the use of optimization in this design case. A component-based representation first serves to manage non-numerical characteristics such as aesthetics. Furthermore, it complies with usual fabrication methods in building design and with object-oriented data handling in CAD. Therefore, components provide an important basis for an interactive MDO process in building design.

A new application of software technology is the application area of smart living or sustainable living. Within this area application platforms are designed and realized with the goal to support value added services. In this context value added services integrates microelectronics, home automation and services to enhance the attractiveness of flats, homes and buildings. Especially real estate companies or service providers dealing with home services are interested in an effective design and management of their services. Service Engineering is the approved approach for designing customer oriented service processes. Service engineering consists of several phases; from situation analysis to service creation and service design to service management. This article will describe how the method service blueprint can be used to design service processes. Smart living includes all actions to enlarge a flat to a smart home for living. One special requirement of this application domain is the use of local components (actuators, sensors) within service processes. This article will show how this extended method supports service providers to improve the quality of customer oriented service processes and the derivation of needed interfaces of involved actors. For the civil engineering process it will be possible to derive needed information from a built in home automation system. The aim is to show, how to get needed smart local components to fullfill later offered it-supported value added services. Value added services focused on inhabitants are grouped to consulting and information, care and supervision, leisure time activities, repairs, mobility and delivery, safety and security, supply and disposal.

Requires for reliability and durability of structures and their elements with simultaneous material economy have stimulated improvement of constitutive equations for description of elasto-plastic deformation processes. This has led to the development of phenomenological modelling of complex phenomena of irreversible deformation including history-dependent and rate-dependent effects. During the last several decades many works have been devoted to the development of elasto-plastic models, in order to better predict the material behavior under combined variable thermo-mechanical loading. The increase of accuracy of stress analysis and safety factors for complex structures with the help of modern finite-element packages (ABAQUS, ANSYS, COSMOS, LS-DYNA, MSC.MARC, MSC.NASTRAN, PERMAS and other) can be provided only by use of complex and special variants of plasticity theories, which are adequate for the considered loading conditions and based on authentic information about properties of materials. The areas of application of the various theories (models) are as a rule unknown to the users of finite-element packages at the existing variety loading condition sin machine-building designs. At the moment a universal theory of inelasticity is absent and even the most accomplished theories can not guarantee adequate description of deformation processes for arbitrary structure under wide range of loading programs. The classifier of materials, loading conditions, effects (phenomena) and list of basic experiments are developed by the authors. Use of these classifiers for an establishment of hierarchy of models is a first step for introduction of the multimodel analysis into computational practice. The set of the classic and modern inelasticity theories is considered, so that they are applicable for stress analysis of structures under complex loading programs. Among them there are plastic flow theories with linear and nonlinear isotropic and kinematic hardening, multisurface theories, endochronic theory, holonomic theory, rheologic models, theory of elasto-plastic processes, slip theory, physical theories (single crystal and polycrystalline models) and others. The classification of materials provides rearranging by a degree of homogeneous, chemical composition, level of strength and plasticity, behavior under cyclic loading, anisotropy of properties at initial condition, anisotropy of properties during deformation process, structural stability. The classification of loading conditions takes into consideration proportional and non-proportional loading, temperature range, combination of cyclic and monotonous loading, one-axial, two-axial and complex stress state, curvature of strain path, presence of stress concentrators and level of strain gradient. A unified general form of constitutive equations is presented for all used material models based upon the concept of internal state variables. The wide range of mentioned above inelastic material models has been implemented into finite element program PANTOCRATOR developed by authors (see for details www.pantocrator.narod.ru). Application possibility of different material models is considered both for material element and for complex structures subjected to complex non-proportional loading.

This paper deals with the development of a new multi-objective evolution strategy in combination with an integrated pollution-load and water-quality model. The optimization algorithm combines the advantages of the Non-Dominated Sorting Genetic Algorithm and Self-Adaptive Evolution Strategies. The identification of a good spread of solutions on the pareto-optimum front and the optimization of a large number of decision variables equally demands numerous simulation runs. In addition, statements with regard to the frequency of critical concentrations and peak discharges require continuous long-term simulations. Therefore, a fast operating integrated simulation model is needed providing the required precision of the results. For this purpose, a hydrological deterministic pollution-load model has been coupled with a river water-quality and a rainfall-runoff model. Wastewater treatment plants are simulated in a simplified way. The functionality of the optimization and simulation tool has been validated by analyzing a real catchment area including sewer system, WWTP, water body and natural river basin. For the optimization/rehabilitation of the urban drainage system, both innovative and approved measures have been examined and used as decision variables. As objective functions, investment costs and river water quality criteria have been used.

The paper is a proposal of calculation of internal forces and dislocations in the reinforced concrete beams before and after cracking. For the ideally elastic bars transfer matrix proposed by Rakowski was applied. The effects associated with cracking were introduced by means of the Borcz's theory in the spectrally way. Numerical example was shown. The presented attitude also enables to calculate dynamic problems and those connected with the stability of the compressed and bending cracked beams and columns.

Water resources development and management is a complex problem. It includes the design and operation of single system components, often as part of larger interrelated systems and usually on the basis of river basins. While several decades ago the dominant objective was the maximization of economic benefit, other objectives have evolved as part of the sustainable development envisaged. Today, planning and operation of larger water resources systems is practically impossible without adequate computer tools, normally being one or several models, increasingly combined with data bank management systems and multi criteria assessment procedures in decision support systems. The use of models in civil engineering already has a long history when structural engineering is considered. These design support models, however, must rather be seen as expert systems made to support the engineer with his daily work. They often have no direct link to stakeholders and the decision makers community. The scale of investigation is often much larger in water resources engineering than in structural engineering which is related to different stakeholders and decision making procedures. Still, several similarities are obvious which can be summarized as the search for a compromise solution on a complex, i.e. multiobjective and interdisciplinary decision problem. While in structural engineering e.g. aestetics, stability and energy consumption might be important evaluation criteria in addition to construction and maintenance cost other or additional criteria have to be considered in water resources planning such as political, environmental and social criteria. In this respect civil engineers tend to overemphasize technical criteria. For the future the existing expert systems should be embedded into an improved decision support shell, keeping in mind that decision makers are hardly interested in numerical modelling results. The paper will introduce into the problem and demonstrate the state of the art by means of an example.

The paper is dedicated to decidability exploration of market segmentation problem with the help of linear convolution algorithms. Mathematical formulation of this problem represents interval task of bipartite graph cover by stars. Vertices of the first partition correspond to types of commodities, vertices of the second – to customers groups. Appropriate method is offered for interval problem reduction to two-criterion task that has one implemented linear convolution algorithm. Unsolvability with the help of linear convolution algorithm of multicriterion, and consequently interval, market segmentation problem is proved.

Procedures of a construction of general solutions for some classes of partial differential equations (PDEs) are proposed and a symmetry operators approach to the raising the orders of the polynomial solutions to linear PDEs are develops. We touch upon an ''operator analytic function theory'' as the solution of a frequent classes of the equations of mathematical physics, when its symmetry operators forms vast enough space. The MAPLE© package programs for the building the operator variables is elaborated also.

We show a close relation between the Schrödinger equation and the conductivity equation to a Vekua equation of a special form. Under quite general conditions we propose an algorithm for explicit construction of pseudoanalytic positive formal powers for the Vekua equation that as a consequence gives us a complete system of solutions for the Schrödinger and the conductivity equations. Besides the construction of complete systems of exact solutions for the above mentioned second order equations and the Dirac equation, we discuss some other applications of pseudoanalytic function theory.

Interval analysis extends the concept of computing with real numbers to computing with real intervals. As a consequence, some interesting properties appear, such as the delivery of guaranteed results or confirmed global values. The former property is given in the sense that unknown numerical values are in known to lie in a computed interval. The latter property states that the global minimum value, for example, of a given function is also known to be contained in a interval (or a finite set of intervals). Depending upon the amount computation effort invested in the calculation, we can often find tight bounds on these enclosing intervals. The downside of interval analysis, however, is the mathematically correct, but often very pessimistic size of the interval result. This is in particularly due to the so-called dependency effect, where a single variable is used multiple times in one calculation. Applying interval analysis to structural analysis problems, the dependency has a great influence on the quality of numerical results. In this paper, a brief background of interval analysis is presented and shown how it can be applied to the solution of structural analysis problems. A discussion of possible improvements as well as an outlook to parallel computing is also given.

In civil engineering practice, values of column forces are often required before any detailed analysis of the structure has been performed. One of the reasons for this arises from the fast-tracked nature of the majority of construction projects: foundations are laid and base columns constructed whilst analysis and design are still in progress. A need for quick results when feasibility studies are performed or when evaluating the effect of design changes on supporting columns form other situations in which column forces are required, but where a detailed analysis to get these forces seems superfluous. Thus it was concluded that the development of an efficient tool for column force calculations, in which the extensive input required in a finite element analysis is to be avoided, would be highly beneficial. The automation of the process is achieved by making use of a Voronoi diagram. The Voronoi diagram is used a) for subdividing the floor into influence areas and b) as a basis for automatic load assignment. The implemented procedure is integrated into a CAD system in which the relevant geometric information of the floor, i.e. its shape and column layout, can be defined or uploaded. A brief description of the implementation is included. Some comparative results and considerations regarding the continuation of the study are given.

This research focuses on an approach to describe principles in architectural layout planning within the domain of revitalization. With the aid of mathematical rules, which are executed by a computer, solutions to design problems are generated. Provided that "design" is in principle a combinatorial problem, i.e. a constraint-based search for an overall optimal solution of a problem, an exemplary method will be described to solve such problems in architectural layout planning. To avoid conflicts relating to theoretical subtleness, a customary approach adopted from Operations Research has been chosen in this work. In this approach, design is a synonym for planning, which could be described as a systematic and methodical course of action for the analysis and solution of current or future problems. The planning task is defined as an analysis of a problem with the aim to prepare optimal decisions by the use of mathematical methods. The decision problem of a planning task is represented by an optimization model and the application of an efficient algorithm in order to aid finding one or more solutions to the problem. The basic principle underlying the approach presented herein is the understanding of design in terms of searching for solutions that fulfill specific criteria. This search is executed by the use of a constraint programming language.

This contribution will be freewheeling in the domain of signal, image and surface processing and touch briefly upon some topics that have been close to the heart of people in our research group. A lot of the research of the last 20 years in this domain that has been carried out world wide is dealing with multiresolution. Multiresolution allows to represent a function (in the broadest sense) at different levels of detail. This was not only applied in signals and images but also when solving all kinds of complex numerical problems. Since wavelets came into play in the 1980's, this idea was applied and generalized by many researchers. Therefore we use this as the central idea throughout this text. Wavelets, subdivision and hierarchical bases are the appropriate tools to obtain these multiresolution effects. We shall introduce some of the concepts in a rather informal way and show that the same concepts will work in one, two and three dimensions. The applications in the three cases are however quite different, and thus one wants to achieve very different goals when dealing with signals, images or surfaces. Because completeness in our treatment is impossible, we have chosen to describe two case studies after introducing some concepts in signal processing. These case studies are still the subject of current research. The first one attempts to solve a problem in image processing: how to approximate an edge in an image efficiently by subdivision. The method is based on normal offsets. The second case is the use of Powell-Sabin splines to give a smooth multiresolution representation of a surface. In this context we also illustrate the general method of construction of a spline wavelet basis using a lifting scheme.

DECENTRALIZED APPROACHES TO ADAPTIVE TRAFFIC CONTROL AND AN EXTENDED LEVEL OF SERVICE CONCEPT
(2006)

Traffic systems are highly complex multi-component systems suffering from instabilities and non-linear dynamics, including chaos. This is caused by the non-linearity of interactions, delays, and fluctuations, which can trigger phenomena such as stop-and-go waves, noise-induced breakdowns, or slower-is-faster effects. The recently upcoming information and communication technologies (ICT) promise new solutions leading from the classical, centralized control to decentralized approaches in the sense of collective (swarm) intelligence and ad hoc networks. An interesting application field is adaptive, self-organized traffic control in urban road networks. We present control principles that allow one to reach a self-organized synchronization of traffic lights. Furthermore, vehicles will become automatic traffic state detection, data management, and communication centers when forming ad hoc networks through inter-vehicle communication (IVC). We discuss the mechanisms and the efficiency of message propagation on freeways by short-range communication. Our main focus is on future adaptive cruise control systems (ACC), which will not only increase the comfort and safety of car passengers, but also enhance the stability of traffic flows and the capacity of the road (“traffic assistance”). We present an automated driving strategy that adapts the operation mode of an ACC system to the autonomously detected, local traffic situation. The impact on the traffic dynamics is investigated by means of a multi-lane microscopic traffic simulation. The simulation scenarios illustrate the efficiency of the proposed driving strategy. Already an ACC equipment level of 10% improves the traffic flow quality and reduces the travel times for the drivers drastically due to delaying or preventing a breakdown of the traffic flow. For the evaluation of the resulting traffic quality, we have recently developed an extended level of service concept (ELOS). We demonstrate our concept on the basis of travel times as the most important variable for a user-oriented quality of service.

Digital models of buildings are widely used in civil engineering. In these models, geometric information is used as leading information. Engineers are used to have geometric information, and, for instance, it is state of the art to specify a point by its three coordinates. However, the traditional approaches have disadvantages. Geometric information is over-determined. Thus, more geometric information is specified and stored than needed. In addition, engineers already deal with topological information. A denotation of objects in buildings is of topological nature. It has to be answered whether approaches where topological information becomes a leading role would be more efficient in civil engineering. This paper presents such an approach. Topological information is modelled independently of geometric information. It is used for denoting the objects of a building. Geometric information is associated to topological information so that geometric information “weights” a topology.
The concept presented in this paper has already been used in surveying existing buildings. Experiences in the use of this concept showed that the number of geometric information that is required for a complete specification of a building could be reduced by a factor up to 100. Further research will show how this concept can be used in planning processes.

Mikroelektronik und Mikrosystemtechnik in Kombination mit Informations- und Kommunikations-technik erlauben es mittlerweile, Rechenleistung und Kommunikationsfähigkeit in kleinsten Formaten, mit geringsten Energien und zu günstigen Preisen nutzbringend in unser privates und berufliches Umfeld einzubringen. Beispiele sind Notebook-PC, PDA, Handy und das Navigationßystem im Auto. Aber auch eingebettete Elektronik in Komponenten, Geräten und Systemen ist nunmehr zur Selbstverständlichkeit geworden. Bekannte Beispiele aus der Haustechnik sind Mikroprozeßoren in Heizungs- und Alarmanlagen und aber auch in Komponenten wie Brand- und Bewegungsmelder. Wir nähern uns dem vor einigen Jahren noch als Vision bezeichneten Zustand der überall vorhandenen elektronischen Rechenleistung (engl. ubiquitous computing) bzw. des von Informationsverarbeitung durchdrungenen täglichen Umfelds (engl. pervasive computing). Werden die TGA-Komponenten genau wie die größeren Computerkomponenten (z.B. PCs, Server) über Datenschnittstellen zu räumlich verteilten Netzwerken verknüpft (z.B. Internet, Intranet) und mit einer systemübergreifenden und adäquaten Intelligenz (Software) programmiert, so können neuartige Funktionalitäten im jeweiligen Anwendungsumfeld (engl. ambient intelligence, kurz AmI, [1]) entstehen. Hier liegt bei Gebäuden und Räumen speziell eine große Chance, die bislang einer ganzheitlichen Systemkonzeption unter Einschluß von Architektur, Gebäudephysik, technischer Gebäudeausrüstung (TGA) und Gebäudeautomation (GA) im Wege stehende Gewerketrennung zu überwinden. Es entstehen für div. Anwendungszwecke systemisch integrierte >smart areas< (nach Prof. Becker, FH Biberach). Im vorliegenden Beitrag erläuterte Beispiele für AmI-Lösungen im Immobilienbereich sind Raumsysteme zur automatischen und sicheren Erkennung von Notfällen, z.B. in Pflegeheimen; sich automatisch an die Nutzung und den Nutzer bzgl. Klima und Beleuchtung adaptierende Raumsysteme im Büro- oder Hotelbereich und die elektronische Aßistenz des Bau- und Betriebsprozeßes von Gebäuden. Im Duisburger inHaus-Innovationszentrum für Intelligente Raum- und Gebäudesysteme der Fraunhofer-Gesellschaft wurden in den letzten Jahren erste Lösungen mit diesem neuartigen Ansatz konzipiert, entwickelt und erprobt. Der Beitrag beschreibt nach einer kurzen Skizzierung des Ambient-Intelligence-Ansatzes an Beispielen Möglichkeiten für den Transfer dieser neuen Technologie in den Raum- und Gebäudebereich. Es folgt eine abschließende Zusammenfaßung und eine Einschätzung der Zukunftspotenziale der Ambient Intelligence in Raum und Bau.

Mit diesen Ausführungen wird ein Beitrag zum weiteren Erhalt der historischen Bausubstanz in Mecklenburg aus der Sicht der Tragwerksanalyse geleistet. Dabei bestätigt es sich immer mehr, dass mit dem Modell der Geometrie, der Belastung und des Materials gleichberechtigte Modelle für eine wirklichkeitsnahe Einschätzung des Tragverhaltens eines Tragwerks vorliegen müssen. Es zeigt sich, dass dabei die besten Berechnungsprogramme nur die Ergebnisse liefern können, die mit den Eingabedaten zu erzielen sind. So hat sich der Forschungsschwerpunkt im Lehrgebiet Tragwerkslehre des FB Architektur an der Hochschule Wismar in den letzten Jahren auf die realistische Abbildung der Wechselwirkung zwischen der Bauaufnahme und der geometrischen Modellierung konzentriert. In diesem Bereich zeigen sich als Schwerpunkte die Wechselwirkung zwischen Schäden und Tragwerksanalyse und die Wechselwirkung zwischen der aufgenommenen Geometrie und dem geometrischen Modell für die Tragwerksanalyse. Die Fülle der aufgenommenen Daten sind dabei in der Regel mehr hinderlich als ein Segen für die Tragwerksanalyse. Hier wurde gezeigt, welche und wie viele geometrische Daten für das geometrische Modell für die Tragwerksanalyse sinnvoll sind. Da die eigene Datenaufnahme relativ viel Zeit beansprucht, wurde eine "geistige" Bauaufnahme durchgeführt. Dazu wird der historische Planungsprozess in den einzelnen Formfindungsschritten nachvollzogen und in die virtuelle Realität überführt. Mit dieser Methode ergeben sich unterschiedliche Bauzustände und es lassen sich auch mögliche Bauphasen abbilden. Die Tragwerksanalyse dieser virtuellen Realität zeigt dann mögliche Schwächen der Tragwerke und/oder die Notwendigkeit konstruktiver Veränderungen. Ein Vergleich der Ergebnisse der Tragwerksanalyse mit der Realität anhand des vorliegenden Datenbestands liefert die Grundlage für den aktuellen Handlungsbedarf. Da der Bauzustand eines Bauwerkes unter einer zeitlichen Veränderung steht, werden Methoden überprüft, die es ermöglichen, einen einmal vorgelegten Datenbestand aufzubereiten und weiter zu verwalten.

The design of challenging space structures frequently relies on the theory of folded plates. The models are composed of plane facets of which the bending and membrane stiffness are coupled along the folds. In conventional finite element analysis of faceted structures the continuity of the displacement field is enforced exclusively at the nodes. Since approximate solutions for transverse and for in-plane displacements are not members of the same function space, separation occurs in between the common nodes of adjacent elements. It is shown that the kinematic assumptions of Bernoulli are accounted for this incompatibility along the edges in facet models. A general answer to this problem involves substantial modification of plate and membrane theory, but a straight forward formulation can be derived for simply folded plates, structures, whose folds do not intersect. A broad class of faceted structures, including models of various curved shells, belong to this category and can be calculated consistently. The additional requirements to assure continuity concern the mapping of displacement derivatives on the edges. An appropriate finite facet element provides node and edge-oriented degrees of freedom, whose transformation to system degrees of freedom, depends on the geometric configuration at each node. The concept is implemented using conform triangular elements. To evaluate the new approach, the energy norm of representative structures for refined meshes is calculated. The focus is placed on the mathematical convergence towards reliable solutions obtained from finite volume models.

The present study was designed to investigate the underlying factors determining the visual impressions of design-patterns that have complex textures. Design-patterns produced by "the dynamical system defined by iterations of discrete Laplacians on the plane lattice" were adopted as stimuli because they were not only complex, but also defined mathematically. In the experiment, 21 graduate and undergraduate students sorted 102 design-patterns into several groups by visual impressions. Those 102 patterns were classified into 12 categories by the cluster analysis. The results showed that the regularity of pattern was a most efficient factor for determining visual impressions of design-pattern, and there were some correspondence between visual impressions and mathematical variables of design-pattern. Especially, the visual impressions were influenced greatly by the neighborhood, and less influenced by steps of iterations.

This is an implementation of the Fillmore–Springer–Cnops construction (FSCc) based on the Clifford algebra capacities of the GiNaC computer algebra system. FSCc linearises the linear-fraction action of the Mobius group. This turns to be very useful in several theoretical and applied fields including engineering. The core of this realisation of FSCc is done for an arbitrary dimension, while a subclass for two dimensional cycles add some 2D-specific routines including a visualisation to PostScript files through the MetaPost or Asymptote software. This library is a backbone of many result published in, which serve as illustrations of its usage. It can be ported (with various level of required changes) to other CAS with Clifford algebras capabilities.

In diesem Beitrag wird eine mobile Software-Komponente zur Vor-Ort-Unterstützung von Bauwerksprüfungen gemäß DIN 1076 „Ingenieurbauwerke im Zuge von Strassen und Wegen, Überwachung und Prüfung“ vorgestellt, welche sich im praktischen Einsatz bei der Hochbahn AG Hamburg befindet. Mit Hilfe dieses Werkzeugs kann die Aktivität am Bauwerk in den gesamten softwaregestützten Geschäftsprozess der auwerksinstandhaltung integriert und somit die Bearbeitungszeit einer Bauwerksprüfung von der Vorbereitung bis zur Prüfbericht-Erstellung reduziert werden. Die Technologie des Mobile Computing wird unter Berücksichtigung spezieller fachlicher Randbedingungen, wie z.B. dem Einsatzort unter freiem Himmel, erläutert und es werden Methoden zur effizienten Datenerfassung mit Stift und Sprache vorgestellt und bewertet. Ferner wird die Einschränkung der Hardware durch die geringere Größe der Endgeräte, die sich durch die Bedingung der Mobilität ergibt, untersucht.

A concept for integrated modeling of urban and rural hydrology is introduced. The concept allows for simulations on the catchment scale as well as on the local scale. It is based on a 2-layer-approach which facilitates the parallel coupling of a catchment hydrology model with an urban hydrology model, considering the interactions between the two systems. The concept has been implemented in a computer model combining a grid based distributed hydrological catchment model and a hydrological urban stormwater model based on elementary units. The combined model provides a flexible solution for time and spatial scale integration and offers to calculate separate water balances for urban and rural hydrology. Furthermore, it is GIS-based which allows for easy and accurate geo-referencing of urban overflow structures, which are considered as points of interactions between the two hydrologic systems. Due to the two-layer-approach, programs of measures can be incorporated in each system separately. The capabilities of the combined model have been tested on a hypothetical test case and a real world application. It could be shown that the model is capable of accurately quantifying the effects of urbanization in a catchment. The affects of urbanization can be analyzed at the catchment outlet, but can also be traced back to its origins, due to the geo-referencing of urban overflow structures. This is a mayor advantage over conventional hydrological catchment models for the analysis of land use changes.

For assessment of old buildings, thermal graphic analysis aided with infra-red camera have been employed in a wide range nowadays. Image processing and evaluation can be economically practicable only if the image evaluation can also be automated to the largest extend. For that reason methods of computer vision are presented in this paper to evaluate thermal images. To detect typical thermal image elements, such as thermal bridges and lintels in thermal images respectively gray value images, methods of digital image processing have been applied, of which numerical procedures are available to transform, modify and encode images. At the same time, image processing can be regarded as a multi-stage process. In order to be able to accomplish the process of image analysis from image formation through perfecting and segmentation to categorization, appropriate functions must be implemented. For this purpose, different measuring procedures and methods for automated detection and evaluation have been tested.

The Lucas-Kanade tracker has proven to be an efficient and accurate method for calculation of the optical flow. However, this algorithm can reliably track only suitable image features like corners and edges. Therefore, the optical flow can only be calculated for a few points in each image, resulting in sparse optical flow fields. Accumulation of these vectors over time is a suitable method to retrieve a dense motion vector field. However, the accumulation process limits application of the proposed method to fixed camera setups. Here, a histogram based approach is favored to allow more than a single typical flow vector per pixel. The resulting vector field can be used to detect roads and prescribed driving directions which constrain object movements. The motion structure can be modeled as a graph. The nodes represent entry and exit points for road users as well as crossings, while the edges represent typical paths.

This paper presents two new methods for analysis of a technical state of large-panel residential buildings. The first method is based on elements extracted from the classical methods and on data about repairs and modernization collected from building documentations. The technical state of a building is calculated as a sum of several groups of elements defining the technical state. The deterioration in this method depends on: - time, which has passed since last repair of element or time which has passed since construction, - estimate of the state of element groups which can be determined on basis of yearly controls. This is a new unique method. it is easy to use, does not need expertise. The required data could be extracted easily from building documentations. For better accuracy the data from building inspections should be applied (in Poland inspections are made every year). The second method is based on the extracted data processing by means of the artificial neural networks. The aim is to learn the artificial neural network configurations for a set of data containing values of the technical state and information about building repairs for last years (or other information and building parameters) and next to analyse new buildings by the instructed neural network. The second profit from using artificial neural networks is the reduction of number of parameters. Instead of more then 40 parameters describing building, about 6-12 are usually sufficient for satisfactory accuracy. This method could have lower accuracy but it is less prone to data errors.

Optimum technological solutions must take into account the entire life cycle of structures including design procedures as well as quality assurance, inspection, maintenance, and repair strategies. Unfortunately, current design standards do not provide a satisfactory basis to ensure expected structural lifetimes. The latter may vary from only a few years for temporary structures to over a century for bridges, water dams or nuclear repositories. Consistent scientific concepts are urgently required to cover this wide spectrum of lifetimes in structural design and maintenance. This was a motivation for a group of scientists at the Ruhr-University Bochum (RUB) to start a special research program supported by the German Research Foundation (DFG) within the Cooperative Research Center SFB 398 since 1996. Institutes of the University Wuppertal and of the University Essen-Duisburg joined the research group. The goal of the Center is to study sources of damage and deterioration in materials and structures, to develop consistent models and simulation methods, to predict structural lifetimes and finally to integrate this predictions into new lifetime-oriented design strategies.
Research activities in our center are organised in three Project Groups as follows:
- Modelling of lifetime effects
- Methods for lifetime-oriented structural analyses
- Future lifespan-oriented design strategies.

For the dynamic behavior of lightweight structures like thin shells and membranes exposed to fluid flow the interaction between the two fields is often essential. Computational fluid-structure interaction provides a tool to predict this interaction and complement or eventually replace expensive experiments. Partitioned analyses techniques enjoy great popularity for the numerical simulation of these interactions. This is due to their computational superiority over simultaneous, i.e. fully coupled monolithic approaches, as they allow the independent use of suitable discretization methods and modular analysis software. We use, for the fluid, GLS stabilized finite elements on a moving domain based on the incompressible instationary Navier-Stokes equations, where the formulation guarantees geometric conservation on the deforming domain. The structure is discretized by nonlinear, three-dimensional shell elements.
Commonly used sequential staggered coupling schemes may exhibit instabilities due to the so-called artificial added mass effect. As best remedy to this problem subiterations should be invoked to guarantee kinematic and dynamic continuity across the fluid-structure interface. Since iterative coupling algorithms are computationally very costly, their convergence rate is very decisive for their usability. To ensure and accelerate the convergence of this iteration the updates of the interface position are relaxed. The time dependent, 'optimal' relaxation parameter is determined automatically without any user-input via exploiting a gradient method or applying an Aitken iteration scheme.

In this study we introduce a concept of discrete Laplacian on the plane lattice and consider its iteration dynamical system. At first we discuss some basic properties on the dynamical system to be proved. Next making their computer simulations, we show that we can realize the following phenomena quite well:(1) The crystal of waters (2) The designs of carpets, embroideries (3) The time change of the numbers of families of extinct animals, and (4) The echo systems of life things. Hence we may expect that we can understand the evolutions and self organizations by use of the dynamical systems. Here we want to make a stress on the following fact: Although several well known chaotic dynamical systems can describe chaotic phenomena, they have difficulties in the descriptions of the evolutions and self organizations.

Interactive visualization based on 3D computer graphics nowadays is an indispensable part of any simulation software used in engineering. Nevertheless, the implementation of such visualization software components is often avoided in research projects because it is a challenging and potentially time consuming task. In this contribution, a novel Java framework for the interactive visualization of engineering models is introduced. It supports the task of implementing engineering visualization software by providing adequate program logic as well as high level classes for the visual representation of entities typical for engineering models. The presented framework is built on top of the open source visualization toolkit VTK. In VTK, a visualization model is established by connecting several filter objects in a so called visualization pipeline. Although designing and implementing a good pipeline layout is demanding, VTK does not support the reuse of pipeline layouts directly. Our framework tailors VTK to engineering applications on two levels. On the first level it adds new – engineering model specific – filter classes to VTK. On the second level, ready made pipeline layouts for certain aspects of engineering models are provided. For instance there is a pipeline class for one-dimensional elements like trusses and beams that is capable of showing the elements along with deformations and member forces. In order to facilitate the implementation of a graphical user interface (GUI) for each pipeline class, there exists a reusable Java Swing GUI component that allows the user to configure the appearance of the visualization model. Because of the flexible structure, the framework can be easily adapted and extended to new problem domains. Currently it is used in (i) an object-oriented p-version finite element code for design optimization, (ii) an agent based monitoring system for dam structures and (iii) the simulation of destruction processes by controlled explosives based on multibody dynamics. Application examples from all three domains illustrates that the approach presented is powerful as well as versatile.

Due to economical, technical or political reasons all over the world about 100 nuclear power plants have been disconnected until today. All these power stations are still waiting for their complete dismantling which, considering one reactor, causes cost of up to one Bil. Euros and lasts up to 15 years. In our contribution we present a resource-constrained project scheduling approach minimizing the total discounted cost of dismantling a nuclear power plant. A project of dismantling a nuclear power plant can be subdivided into a number of disassembling activities. The execution of these activities requires time and scarce resources like manpower, special equipment or storage facilities for the contaminated material arising from the dismantling. Moreover, we have to regard several minimum and maximum time lags (temporal constraints) between the start times of the different activities. Finally, each disassembling activity can be processed in two alternative execution modes, which lead to different disbursements and determine the resource requirements of the considered activity. The optimization problem is to determine a start time and an execution mode for each activity, such that the discounted cost of the project is minimum, and neither the temporal constraints are violated nor the activities' resource requirements exceed the availability of any scarce resource at any point in time. In our contribution we introduce an appropriate multi-mode project scheduling model with minimum and maximum time lags as well as renewable and cumulative resources for the described optimization problem. Furthermore, we show that the considered optimization problem is NP-hard in the strong sense. For small problem instances, optimal solutions can be gained from a relaxation based enumeration approach which is incorporated into a branch and bound algorithm. In order to be able to solve large problem instances, we also propose a truncated version of the devised branch and bound algorithm.

MODEL OF TRAM LINE OPERATION
(2006)

From passenger's perspective punctuality is one of the most important features of trams operations. Unfortunately in most cases this feature is only insufficiently fulfilled. In this paper we present a simulation model for trams operation with special focus on punctuality. The aim is to get a helpful tool for designing time-tables and for analyzing the effects by changing priorities for trams in traffic lights respectively the kind of track separation. A realization of trams operations is assumed to be a sequence of running times between successive stops and times spent by tram at the stops. In this paper the running time is modeled by the sum of its mean value and a zero-mean random variable. With the help of multiple regression we find out that the average running time is a function depending on the length of the sections and the number of intersections. The random component is modeled by a sum of two independent zero-mean random variables. One of these variables describes the disturbance caused by the process of waiting at an intersection and the other the disturbance caused by the process of driving. The time spent at a stop is assumed to be a random variable, too. Its distribution is estimated from given measurements of these stop times for different tram lines in Kraków. Finally a special case of the introduced model is considered and numerical results are presented. This paper is involved with CIVITAS-CARAVEL project: "Clean and better transport in cites". The project has received research funding from the Community's Sixth Framework Programme. The paper reflects only the author's views and the Community is not liable for any use that may be made of the information contained therein.

The ride of the tram along the line, defined by a time-table, consists of the travel time between the subsequent sections and the time spent by tram on the stops. In the paper, statistical data collected in the city of Krakow is presented and evaluated. In polish conditions, for trams the time spent on stops makes up the remarkable amount of 30 % of the total time of tram line operation. Moreover, this time is characterized by large variability. The time spent by tram on a stop consists of alighting and boarding time and time lost by tram on stop after alighting and boarding time ending, but before departure. Alighting and boarding time itself usually depends on the random number of alighting and boarding passengers and also on the number of passengers which are inside the vehicle. However, the time spent by tram on stop after alighting and boarding time ending is an effect of certain random events, mainly because of impossibility of departure from stop, caused by lack of priorities for public transport vehicles. The main focus of the talk lies on the description and the modelling of these effects. This paper is involved with CIVITAS-CARAVEL project: "Clean and better transport in cites". The project has received research funding from the Community's Sixth Framework Programme. The paper reflects only the author's views and the Community is not liable for any use that may be made of the information contained therein.

The contribution presents a model that is able to simulate construction duration and cost for a building project. This model predicts set of expected project costs and duration schedule depending on input parameters such as production speed, scope of work, time schedule, bonding conditions and maximum and minimum deviations from scope of work and production speed. The simulation model is able to calculate, on the basis of input level of probability, the adequate construction cost and time duration of a project. The reciprocal view attends to finding out the adequate level of probability for construction cost and activity durations. Among interpretive outputs of the application software belongs the compilation of a presumed dynamic progress chart. This progress chart represents the expected scenario of development of a building project with the mapping of potential time dislocations for particular activities. The calculation of a presumed dynamic progress chart is based on an algorithm, which calculates mean values as a partial result of the simulated building project. Construction cost and time models are, in many ways, useful tools in project management. Clients are able to make proper decisions about the time and cost schedules of their investments. Consequently, building contractors are able to schedule predicted project cost and duration before any decision is finalized.

The paper presents a linear static analysis on continuous orthotropic thin-walled shell structures simply supported at the transverse ends with a random deformable contour of the cross section. The external loads can be random as well. The class of this structures involves most of the bridges, scaffold bridges, some roof structures etc. A numerical example of steel continuous structures on five spans with an open contour of the cross-section has been solved. The examination of the structure has used the following two computation models: a prismatic structure consisting of isotropic strips, a plates and ribs, with considering their real interaction, and a smooth orthotropic plate equivalent to the structure in the first model. The displacements and forces of the structure characterizing its stressed and deformed condition have been determined. The results obtained from the two solutions have been analyzed. The study on the structure is made with the force method in combination with the analytical finite strip method (AFSM) in displacements. The basic system is obtained by separating the superstructure from the understructure at the places of intermediate supports and consists of two parts. The first part is a single span thin-walled prismatic shell structure; the second part presents supports (columns, space frames etc.). The connection between the superstructure and intermediate supports is made under random supporting conditions. The forces at the supporting points in the direction of the connections removed are assumed to be the basic unknowns of the force method. The solution of the superstructure has been accomplished by the AFSM in displacements. The structure is divided in only one (transverse) direction into a finite number of plain strips connected to each other in longitudinal linear nodes. The three displacements of the points on the node lines and the rotation around those lines have been assumed to be the basic unknown in each node. The boundary conditions of each strip of the basic system correspond to the simply support along the transverse ends and the restraint along the longitudinal ones. The particular strip of the basic system has been solved by the method of the single trigonometric series. The method is reduced to solving a discrete structure in displacements and restoring its continuity at the places of the sections made in respect to both the displacements and forces. The two parts of the basic system have been solved in sequence under the action of single values of each of the basic unknowns and with the external load. The solution of the support part is accomplished using software for analyzing structures by the FEM. The basic unknown forces have been determined from system of canonic equations, the conditions of the deformations continuity on the places of the removed connections under superstructure and intermediate supports. The final displacements and forces at a random point of a continuous superstructure have been determined using the principle of superposition. The computations have been carried by software developed with Visual Fortran version 5.0 for PC.

MODELLING THE PLASTIC HINGE IN THE STATICALLY INDETERMINABLE REINFORCED CONCRETE BAR ELEMENTS
(2006)

The paper presents the example numerical model to calculate the reinforced concrete bar structures. Usually applied methods of structure dimensioning do not include the case of plastic hinges occurrence under the limit load of construction. The model represented by A. Borcz is based on the differential equation of deflection line of the beam and it includes the effects of rearrangement of the internal forces and reological effects. The experimental parameters obtained in earlier investigations describe effects resulting from the rise of plastic hinges in the proposed equation.

Durch die Betrachtung des Produktions-Prozesses als zentrales Transformationselement wird die Struktur der Bauproduktion realitätsnah gefasst. Die Integration der prozessorientierten Kostendefinition setzt relevante Kostenparameter und Produktionsfaktoren so in Beziehung, dass sie im Einklang mit der realen Kostenstruktur und Kostendynamik einer Baustelle stehen. Die Beziehung zwischen Bauzeit und Kosten wird direkt erfasst und ausgewertet. Der hohen Dynamik der Bauproduktion zwischen kapazitätsbeschränkten Einsatzmitteln und Produktionsprozessen wurde durch das Poolmodell und der Simulation als Berechnungsmethode Rechnung getragen. Eine einfache Modellierung von sich zyklusartig wiederholenden Arbeitsvorgängen (Taktplanung) ist möglich. Die Taktbildung vollzieht sich bei der Simulation durch Kapazitätsbeschränkungen ohne Zutun des Benutzers. Durch eine Optimierungsmethode kann automatisiert nach der kostengünstigsten oder zeitlich schnellsten Produktionsvariante gesucht werden

Die effektive Kooperation aller beteiligten Fachplaner im Bauplanungsprozess ist die Voraussetzung für wirtschaftliches und qualitativ hochwertiges Bauen. Bauprojektorganisationen bestehen in der Regel aus zahlreichen unabhängigen Planungspartnern, die örtlich verteilt spezifische Planungsaufgaben bearbeiten und die Ergebnisse in Teilproduktmodellen ablegen. Da Planungsprozesse im Bauwesen stark arbeitsteilig ablaufen, sind die Teilproduktmodelle der einzelnen Fachplanungen in hohem Maße voneinander abhängig. Ziel des hier vorgestellten Ansatzes ist die Integration der Teilproduktmodelle der Gebäudeplanung in einem netzwerkbasierten Modellverbund am Beispiel der Brandschutzplanung. Im Beitrag werden die Probleme der Verteiltheit und insbesondere der semantischen Heterogenität der involvierten Teilproduktmodelle betrachtet. Der verteilte Zugriff wird mithilfe mobiler Software-Agenten realisiert. Die Agenten können sich dabei frei im netzwerkbasierten Planungsverbund bewegen und agieren als Vertreter der Fachplaner. Das Problem der semantischen Heterogenität der Teilproduktmodelle wird auf der Basis von Ontologien gelöst. Dazu werden erstens Domänenontologien entwickelt, die Objekte der realen Welt einer abgeschlossenen Domäne, hier des Brandschutzes, abbilden. Zweitens werden Applikationsontologien entwickelt, die die einzelnen proprietären Datenhaltungen (im Sinne von Teilproduktmodellen) der jeweiligen Fachplanungen repräsentieren. Beide Ontologien werden mit einem regelbasierten Ansatz verknüpft. Im vorgestellten Anwendungsfall Brandschutz dient die Domänenontologie als einheitliche Schnittstelle für den Zugriff auf die verteilten Modelle und abstrahiert dabei von deren Datenbankspezifika und proprietären Schemata. Mithilfe von mobilen Agenten und semantischen Technologien kann so eine Plattform zur Verfügung gestellt werden, die erstens die dynamische Integration von Ressourcen in den Planungsverbund erlaubt und zweitens auf deren Basis unabhängig von der Verteiltheit und Heterogenität der eingebundenen Ressourcen ingenieurgerechte Verarbeitungsmethoden realisiert werden können.

The use of virtual reality techniques in the development of educational applications brings new perspectives to the teaching of subjects related to the field of civil construction in Civil Engineering domain. In order to obtain models, which would be able to visually simulate the construction process of two types of construction work, the research turned to the techniques of geometric modelling and virtual reality. The applications developed for this purpose are concerned with the construction of a cavity wall and a bridge. These models make it possible to view the physical evolution of the work, to follow the planned construction sequence and to visualize details of the form of every component of the works. They also support the study of the type and method of operation of the equipment necessary for these construction procedures. These models have been used to distinct advantage as educational aids in first-degree courses in Civil Engineering. Normally, three-dimensional geometric models, which are used to present architectural and engineering works, show only their final form, not allowing the observation of their physical evolution. The visual simulation of the construction process needs to be able to produce changes to the geometry of the project dynamically. In the present study, two engineering construction work models were created, from which it was possible to obtain three-dimensional models corresponding to different states of their form, simulating distinct stages in their construction. Virtual reality technology was applied to the 3D models. Virtual reality capacities allow the interactive real-time viewing of 3D building models and facilitate the process of visualizing, evaluating and communicating.

The reduction of oscillation amplitudes of structural elements is necessary not only for maintenance of their durability and longevity but also for elimination of a harmful effect of oscillations on people and technology operations. The dampers are widely applied for this purpose. One of the most widespread models of structural friction forces having piecewise linear relation to displacement was analysed. T The author suggests the application of phase trajectories mapping in plane "acceleration – displacement". Unlike the trajectories mapping in a plane "velocity – displacement", they don't require large number of geometrical constructions for identification of the characteristics of dynamic systems. It promotes improving the accuracy. The analytical assumptions had been verified by numerical modeling. The results show good enough coincide between numerical and analytical estimation of dissipative characteristic.

Die meisten Insolvenzen in Deutschland kommen aus der Bauindustrie. Die Gründe hierfür sind vielschichtig, jedoch kann mittels eines modern ausgerichteten M-I-S und Baustellen-Controllings frühzeitig erkannt werden, wie sich die Baustellenergebnisse entwickeln. Hierzu ist es notwendig, dass die Arbeitskalkulation ständig auf dem Laufenden gehalten wird. Nur wenn dies geschieht, sind monatliche Soll-/ Ist-Vergleiche und eine Betrachtung der cost-to-complete möglich und sinnvoll. Eine monatlich rollierende Prognose des Baustellenergebnisses zum Bauende ermöglicht, dass gravierende Veränderungen des Ergebnisses umgehend aufgedeckt werden. Nur in Kenntnis dieser Entwicklungen kann das Management frühzeitig (im Sinne eines Frühwarnsystems) agieren und Steuerungsmaßnahmen ergreifen. Die Ergebnisprognose zum Bauende ist allein als Steuerungsinstrument nicht ausreichend. Die Finanzsituation der Baustelle muß auch regelmäßig geprüft werden, d.h. der Leistungsstand mit der Rechnungsstellung an den Bauherren abgeglichen sowie die unbezahlten Rechnungen des Bauherren überprüft werden. Das beste Prognoseergebnis ist wertlos, wenn der Bauherr seine bezogenen Leistungen nicht vergütet. Die wirtschaftlichen Daten stehen den Verantwortlichen online im Baustellen-Informations-System (B-I-S) zur Verfügung. Ein Ampelsystem verdeutlicht die wirtschaftliche Lage der Baustelle.

Subject of the paper is the realisation of a model based efficiency control system for PV generators using a simulation model. A standard 2-diodes model of PV generator is base of the ColSim model, which is implemented in ANSI C code for flexible code exporting. The algorithm is based on discretisized U-I characteristics, which allows the calculation of string topologies witch parallel and serial PV cells and modules. Shadowing effects can be modelled down to cell configuration using polar horizon definitions. The simulation model was ported to a real time environment, to calculate the efficiency of a PV system. Embedded System technology allows the networked operation and the integration of standard I/O devices. Futher work focus on the adaption of shadowing routine, which will be adapted to get the environment conditions from the real operation.

Analysis of the reinforced concrete chimney geometry changes and their influence on the stresses in the chimney mantle was made. All the changes were introduced to a model chimney and compared. Relations between the stresses in the mantle of the chimney and the deformations determined by the change of the chimney's vertical axis geometry were investigated. The vertical axis of chimney was described by linear function (corresponding to the real rotation of the chimney together with the foundation), and by parabolic function (corresponding to the real dislocation of the chimney under the influence of the horizontal forces - wind). The positive stress pattern in the concrete as well as the negative stress pattern in the reinforcing steel have been presented. The two cases were compared. Analysis of the stress changes in the chimney mantle depending on the modification in the thickness of the mantle (the thickness of the chimney mantle was altered in the linear or the abrupt way) was carried out. The relation between the stresses and the chimney's diameter change from the bottom to the top of the chimney was investigated. All the analyses were conducted by means of a specially developed computer program created in Mathematica environment. The program makes it also possible to control calculations and to visualize the results of the calculations at every stage of the calculation process.

We present an algebraically extended 2D image representation in this paper. In order to obtain more degrees of freedom, a 2D image is embedded into a certain geometric algebra. Combining methods of differential geometry, tensor algebra, monogenic signal and quadrature filter, the novel 2D image representation can be derived as the monogenic extension of a curvature tensor. The 2D spherical harmonics are employed as basis functions to construct the algebraically extended 2D image representation. From this representation, the monogenic signal and the monogenic curvature signal for modeling intrinsically one and two dimensional (i1D/i2D) structures are obtained as special cases. Local features of amplitude, phase and orientation can be extracted at the same time in this unique framework. Compared with the related work, our approach has the advantage of simultaneous estimation of local phase and orientation. The main contribution is the rotationally invariant phase estimation, which enables phase-based processing in many computer vision tasks.

The execution of project activities generally requires the use of (renewable) resources like machines, equipment or manpower. The resource allocation problem consists in assigning time intervals to the execution of the project activities while taking into account temporal constraints between activities emanating from technological or organizational requirements and costs incurred by the resource allocation. If the total procurement cost of the different renewable resources has to be minimized we speak of a resource investment problem. If the cost depends on the smoothness of the resource utilization over time the underlying problem is called a resource levelling problem. In this paper we consider a new tree-based enumeration method for solving resource investment and resource levelling problems exploiting some fundamental properties of spanning trees. The enumeration scheme is embedded in a branch-and-bound procedure using a workload-based lower bound and a depth first search. Preliminary computational results show that the proposed procedure is promising for instances with up to 30 activities.

The aim of this paper is to present so-called discrete-continual boundary element method (DCBEM) of structural analysis. Its field of application comprises buildings constructions, structures and also parts and components for the residential, commercial and un-inhabitant structures with invariability of physical and geometrical parameters in some dimensions. We should mention here in particular such objects as beams, thin-walled bars, strip foundations, plates, shells, deep beams, high-rise buildings, extensional buildings, pipelines, rails, dams and others. DCBEM comes under group of semianalytical methods. Semianalytical formulations are contemporary mathematical models which currently becoming available for realization due to substantial speed-up of computer productivity. DCBEM is based on the theory of the pseudodifferential boundary equations. Corresponding pseudodifferential operators are discretely approximated using Fourier analysis or wavelet analysis. The main DCBEM advantages against the other methods of the numerical analysis is a double reduction in dimension of the problem (discrete numerical division applied not to the full region of the interest but only to the boundary of the region cross section, as a matter of fact one is solving an one-dimensional problem with the finite step on the boundary area of the region), one has opportunities to carrying out very detailed analysis of the specific chosen zones, simplified initial data preparation, simplistic and adaptive algorithms. There are two methods to define and conduct DCBEM analysis developed – indirect (IDCBEM) and direct (DDCBEM), thus indirect like in boundary element method (BEM) applied and used little bit more than direct.

In distributed project organisations and collaboration there is a need for integrating unstructured self-contained text information with structured project data. We consider this a process of text integration in which various text technologies can be used to externalise text content and consolidate it into structured information or flexibly interlink it with corresponding information bases. However, the effectiveness of text technologies and the potentials of text integration greatly vary with the type of documents, the project setup and the available background knowledge. The goal of our research is to establish text technologies within collaboration environments to allow for (a) flexibly combining appropriate text and data management technologies, (b) utilising available context information and (c) the sharing of text information in accordance to the most critical integration tasks. A particular focus is on Semantic Service Environments that leverage on Web service and Semantic Web technologies and adequately support the required systems integration and parallel processing of semi-structured and structured information. The paper presents an architecture for text integration that extends Semantic Service Environments with two types of integration services. Backbone to the Information Resource Sharing and Integration Service is a shared environment ontology that consolidates information on the project context and the available model, text and general linguistic resources. It also allows for the configuration of Semantic Text Analysis and Annotation Services to analyse the text documents as well as for capturing the discovered text information and sharing it through semantic notification and retrieval engines. A particular focus of the paper is the definition of the overall integration process configuring a complementary set of analyses and information sharing components.

Räume und Gebäude sind heute wegen der enormen Funktionalität der technischen Gebäudeausrüstung (TGA) in Kombination mit der sonstigen Ausstattung und den diversen Anwendungsprozessen und Nutzergruppen ohne innovative Konzepte der integrierten Bedienung kaum noch beherrschbar bzw. optimal nutzbar. Dies gilt sowohl für Wohn- als auch für Zweckimmobilien. Die Gebäudeleittechnik (GLT) und die Gebäudeautomation (GA) können hier unter sinnvoller Integration der Möglichkeiten der Mikroelektronik, Multimedia-, Kommunikations- und Informationstechnik erheblich zu nutzbringenden Innovationen beitragen. Die Automobilindustrie hat in den letzten Jahren gezeigt, wie durch einen integralen Systemansatz und durch Einsatz von Elektronik, Kommunikations- und Informationstechnik eine sinnvolle technische Assistenz der Anwender machbar ist. Genannt sei hier das Konzept des Cockpits mit integrierter Funktionsbündelung und der Informationskonzentration am Armaturenbrett. Im Gegensatz zum Automobil ist der Bereich der technischen Gebäudeausstattung in Wohn- und Nutzimmobilien gekennzeichnet durch eine starke Fragmentierung in unterschiedlichste Gewerke unter Beteiligung vieler oft schlecht koordinierter Akteure. Durch das Duisburger inHaus-Innovationszentrum für Intelligente Raum- und Gebäudesysteme der Fraunhofer-Gesellschaft wurden in den letzten Jahren neuartige Konzepte der Systemintegration heterogener Technik auf der Basis von Middleware-Plattformen und Multimedia-Technologien und -Geräten entwickelt, getestet und in die Anwendung getragen. Einer der ersten Systemanwendungen dieses offenen Infrastrukturkonzepts ist die integrierte Systembedienung mit zum Teil völlig neuen Bedienkonzepten und einer starken Bedienungsvereinfachung auch komplexester Technikausrüstungen in Immobilien. Der Beitrag beschreibt nach einer Analyse der Ausgangslage die technologischen Grundzüge der integrierten Systembedienung. Es folgen einige Anwendungsbeispiele und eine zusammenfassende Bewertung mit einem Ausblick auf weiterführende Aktivitäten.

Unconstrained models are very often found in the broad spectrum of different theories of traffic demand models. In these models there are none or only one-sided restrictions influencing the choice of the individual. However in the traffic demand different deciding dependencies of the traffic volume with regard to the specific conditions of the territory structure potentials exist. Kichhoff and Lohse introduced bi- and tri-linearly constrained models to show these dependencies. In principle, the dependencies are described as hard, elastic and open boundary sum criteria. In this article a model is formulated which gets away from these predefined boundary sum criteria and allows a free determination of minimal and maximal boundary sum criteria. The iterative solution algorithm is shown according to a FURNESS procedure at the same time. With the approach of freely selectable minimal and maximal boundary sum criteria the modeling transport planner gets the possibility to show the traffic event even better. Furthermore all common boundary sum criteria can be calculated with this model. Therewith the often necessary and sensible standard and special cases can also be modeled.

Im Bereich der Altbausanierung und der Bestandserfassung im Bauwesen ist es häufig notwendig, bestehende Pläne hinsichtlich des Bauwerkszustandes zu aktualisieren oder, wenn diese Pläne nicht (mehr) zugänglich sind, gänzlich neue Planunterlagen des Ist-Zustandes zu erstellen. Ein komfortabler Weg, diese Bauwerksdaten zu erheben, eröffnet die Technologie der Laservermessung. Der vorliegende Artikel stellt in diesem Zusammenhang Ansätze zur Teilautomatisierung der Generierung eines dreidimensionalen Computermodells eines Bauwerkes vor. Als Ergebnis wird ein Volumenmodell bereitgestellt, in dem zunächst die geometrischen und topologischen Informationen über Flächen, Kanten und Punkte im Sinne eines B-rep Modells beschrieben sind. Die Objekte dieses Volumenmodells werden mit Verfahren aus dem Bereich der künstlichen Intelligenz analysiert und in Bauteilklassen systematisch kategorisiert. Die Kenntnis der Bauteilsemantik erlaubt es somit, aus den Daten ein Bauwerks-Produktmodell abzuleiten und dieses einzelnen Fachplanern – etwa zur Erstellung eines Energiepasses – zugänglich zu machen. Der Aufsatz zeigt den erfolgreichen Einsatz virtueller neuronaler Netze im Bereich der Bestandserfassung anhand eines komplexen Beispiels.

Die Kommunale Wohnungsgesellschaft mbH Erfurt(KoWo) ist mit ihren rund 20.000 Wohnungen in der Landeshauptstadt das größte Wohnungsunternehmen in Thüringen. Der Immobilienbestand ist heterogen in seinem technischen Zustand und im Bezug auf die unterschiedlichen Lagen der Objekte. Bedingt durch Leerstände und unterschiedliche Modernisierungsmaßnahmen und -stände unterscheidet sich die Wirtschaftlichkeit verschiedener Objekte deutlich. Ohne eine einheitliche Einwertung des Immobilienbestandes im Bezug auf die Objektattraktivität, die Standortqualität und die Objektwirtschaftlichkeit fällt eine langfristige strategische Entwicklung des Immobilienportfolios schwer. Über die Schritte der technischen Bestandserfassung, die Einwertung über ein Scorintmodell, die Abbildung in einem Portfoliomodell mit zugehöriger Normstrategie bis hin zur Weiterverarbeitung der Daten in der 20-jährigen Instandsetzungsplanung wird praxisnah aufgezeigt, wie die Vorgehensweise bei der Einwertung des Immobilienportfolios ist.

In many branches companies often lose the visibility of their human and technical resources of their field service. On the one hand the people in the fieldservice are often free like kings on the other hand they do not take part of the daily communication in the central office and suffer under the lacking involvement in the decisions inside the central office. The result is inefficiency. Reproaches in both directions follow. With the radio systems and then mobile phones the ditch began to dry up. But the solutions are far from being productive.

Car following models are used to describe the behavior of a number of cars on the road dependent on the distance to the car in front. We introduce a system of ordinary differential equations and perform a theoretical and numerical analysis in order to find solutions that reflect various traffic situations. We present three different variations of the model motivated by reality.

VARIATION OF ROTATIONAL RESTRAINT IN GRID DECK CONNECTION DUE TO CORROSION DAMAGE AND STRENGTHENING
(2006)

The approach to assessment of rotational restraint of stringer-to-crossbeam connection in a deck of 100-year old steel truss bridge is presented. Sensitivity of rotational restraint coefficient of the connection to corrosion damage and strengthening is analyzed. Two criteria of the assessment of the rotational restraint coefficient are applied: static and kinematic one. The former is based on bending moment distribution in the considered member, the latter one – on the member rotation at the given joint. 2D-element model of finite element method is described: webs and flanges are modeled with shell elements, while rivets in the connection – with system of beam and spring elements. The method of rivet modeling is verified by T-stub connection test results published in literature. FEM analyses proved that recorded extent of corrosion damage does not alter the initial rotational restraint of stringer-to-crossbeam connection. Strengthening of stringer midspan influences midspan bending moment and stringer end rotation in a different way. Usually restoring member load bearing capacity means strengthening its critical regions (where the highest stress levels occur). This alters flexural stiffness distribution over member length and influences rotational restraint at its connection to other members. The impact depends on criterion chosen for rotational restraint coefficient assessment.

The concept is presented of the sensitivity analysis of the limit state of the structure with respect to selected basic variables. The sensitivity is presented in the form of the probability distribution of the limit state of the structure. The analysis is performed by the problem-oriented Monte Carlo simulation procedure. The procedure is based on the problem's definition of the elementary event, as a structural limit state. Thus the sample space consists of limit states of the structure. Defined on the sample space the one-dimensional random multiplier is introduced. This multiplier refers to the dominant basic variable (group of variables) of the problem. Numerical procedure results in the set of random numbers. Normalized relative histogram of this set is an estimator of the PDF of the limit state of the structure. Estimators of reliability, or the probability of failure are statistical characteristics of this histogram. The procedure is illustrated by the example of sensitivity analysis of the serviceability limit state of monumental structure. It is the colonnade of Licheń Basilica, situated in central Poland. Limit state of the structure is examined with reference to the upper deck horizontal deflection. Wind actions are taken as dominant variables. An assumption is made that the wind load intensities acting on the lower and on the upper storey of the colonnade, respectively, are identically distributed, but correlated random variables. Three correlation variants of these variables are considered. Relevant limit state histograms are analysed thereafter. The paper ends with the conclusions referring to the method and some general remarks on the fully probabilistic design.

This paper presents results of applying Fuzzy Inference System for estimation of the number of potential Park and Ride users. Usually it is difficult to evaluate the number of users because it depends on human factor and data in the considered system are uncertain. In such situation the traditional mathematical approaches can not take into consideration rough data. Therefore a fuzzy approach can be applied in this case. A fuzzy methodology is treated as a proper way to describe choice of mode of transport, and especially that uncertainty accompanied of choosing process has rather fuzzy character. The proposed approach is based on the Mamdani Fuzzy Inference System and for calculation there is used Matlab software with Fuzzy Logic Toolbox. Mamdani model requires, as an input data, knowledge of the shape of membership function. These functions can be calibrated taking into consideration results of questionnaires conducted among users of Park and Ride system. Due to lack of representative sample of users, one has decided to use results of experts' questionnaires as a input data for calibration the shape of membership functions. Describing factor will be generalized cost of the trip for different modes of transport. Proposed approach consists of two main stages: modeling of share of public/private transport trips and Multimodal model estimating number of Park and Ride users. Verification of presented methodology is treated as an indirect proof. Proposed approach can be applied for estimation of bi-modal split. Then the results are compared with traditional approaches based on logit functions. Comparable results of proposed fuzzy approach with traditional logit models can be treated as a confirmation of chosen methodology.

In this paper proposed the application of two-parameters damage model, based on non-linear finite element approach, to the analysis of masonry panels. Masonry is treated as a homogenized material, for which the material characteristics can be defined by using homogenization technique. The masonry panels subjected to shear loading are studied by using the proposed procedure within the framework of three-dimensional analyses. The nonlinear behaviour of masonry can be modelled using concepts of damage theory. In this case an adequate damage function is defined for taking into account different response of masonry under tension and compression states. Cracking can, therefore, be interpreted as a local damage effect, defined by the evolution of known material parameters and by one or several functions which control the onset and evolution of damage. The model takes into account all the important aspects which should be considered in the nonlinear analysis of masonry structures such as the effect of stiffness degradation due to mechanical effects and the problem of objectivity of the results with respect to the finite element mesh. Finally the proposed damage model is validated with a comparison with experimental results available in the literature.

The paper proposes a new method for general 3D measurement and 3D point reconstruction. Looking at its features, the method explicitly aims at practical applications. These features especially cover low technical expenses and minimal user interaction, a clear problem separation into steps that are solved by simple mathematical methods (direct, stable and optimal with respect to least error squares), and scalability. The method expects the internal and radial distortion parameters of the used camera(s) as inputs, and a plane quadrangle with known geometry within the scene. At first, for each single picture the 3D position of the reference quadrangle (with respect to each camera coordinate frame) is calculated. These 3D reconstructions of the reference quadrangle are then used to yield the relative external parameters of each camera regarding the first one. With known external parameters, triangulation is finally possible. The differences from other known procedures are outlined, paying attention to the stable mathematical methods (no usage of nonlinear optimization) and the low user interaction with good results at the same time.

The paper contains a description of dynamic effects in the silo wall during the outflow of a stored material. The work allows for determining the danger of construction damage due to resonant vibrations and is of practical importance by determining the influence of cyclic pressures and vibro–creeping during prolonged use of a silo. The paper was devised as a result of tests on silo walls in semi-technical scale. The model is generally applicable and allows for identification of parameters in real- size silos as well.

The paper is devoted to the investigation of dynamical behavior of a cable under influence of various types of excitations. Such element has a low rigidity and is sensitive to dynamic effect. The structural scheme is a cable which ends are located at different level. The analysis of dynamical behavior of the cable under effect of kinematical excitation which is represented by the oscillations of the upper part of tower is given. The scheme of cable is accepted such, that lower end of an inclined cable is motionless. The motion of the upper end is assumed only in horizontal direction. The fourth-order Runge-Kutta method was realized in software. The fast Fourier transform was used for spectral analysis. Standard graphical software was adopted for presenting results of investigations. The mathematical model of oscillations of a cable was developed by the account of the viscous damping. The analysis of dynamical characteristics of a cable for various parameters of damping and kinematical excitation was carried out. The time series, spectral characteristics and amplitude-frequencies characteristics was obtained. The resonance amplitude for different oscillating regimes was estimated. It is noted that increasing of the coefficient of the viscous damping and decreasing of the amplitude of tower's oscillations reduces the value of the critical frequency and the resonant amplitudes.