### Refine

#### Has Fulltext

- yes (289) (remove)

#### Document Type

- Conference Proceeding (286)
- Article (3)

#### Institute

- In Zusammenarbeit mit der Bauhaus-Universität Weimar (174)
- Graduiertenkolleg 1462 (31)
- Professur Informatik im Bauwesen (24)
- Institut für Strukturmechanik (22)
- Professur Angewandte Mathematik (18)
- Institut für Konstruktiven Ingenieurbau (8)
- Professur Stahlbau (4)
- Institut für Mathematik-Bauphysik (3)
- Juniorprofessur Stochastik und Optimierung (3)
- Professur Informatik in der Architektur (3)

#### Keywords

- Computerunterstütztes Verfahren (289) (remove)

The aim of this paper is to present so-called discrete-continual boundary element method (DCBEM) of structural analysis. Its field of application comprises buildings constructions, structures and also parts and components for the residential, commercial and un-inhabitant structures with invariability of physical and geometrical parameters in some dimensions. We should mention here in particular such objects as beams, thin-walled bars, strip foundations, plates, shells, deep beams, high-rise buildings, extensional buildings, pipelines, rails, dams and others. DCBEM comes under group of semianalytical methods. Semianalytical formulations are contemporary mathematical models which currently becoming available for realization due to substantial speed-up of computer productivity. DCBEM is based on the theory of the pseudodifferential boundary equations. Corresponding pseudodifferential operators are discretely approximated using Fourier analysis or wavelet analysis. The main DCBEM advantages against the other methods of the numerical analysis is a double reduction in dimension of the problem (discrete numerical division applied not to the full region of the interest but only to the boundary of the region cross section, as a matter of fact one is solving an one-dimensional problem with the finite step on the boundary area of the region), one has opportunities to carrying out very detailed analysis of the specific chosen zones, simplified initial data preparation, simplistic and adaptive algorithms. There are two methods to define and conduct DCBEM analysis developed – indirect (IDCBEM) and direct (DDCBEM), thus indirect like in boundary element method (BEM) applied and used little bit more than direct.

The execution of project activities generally requires the use of (renewable) resources like machines, equipment or manpower. The resource allocation problem consists in assigning time intervals to the execution of the project activities while taking into account temporal constraints between activities emanating from technological or organizational requirements and costs incurred by the resource allocation. If the total procurement cost of the different renewable resources has to be minimized we speak of a resource investment problem. If the cost depends on the smoothness of the resource utilization over time the underlying problem is called a resource levelling problem. In this paper we consider a new tree-based enumeration method for solving resource investment and resource levelling problems exploiting some fundamental properties of spanning trees. The enumeration scheme is embedded in a branch-and-bound procedure using a workload-based lower bound and a depth first search. Preliminary computational results show that the proposed procedure is promising for instances with up to 30 activities.

We present recent developments of adaptive wavelet solvers for elliptic eigenvalue problems. We describe the underlying abstract iteration scheme of the preconditioned perturbed iteration. We apply the iteration to a simple model problem in order to identify the main ideas which a numerical realization of the abstract scheme is based upon. This indicates how these concepts carry over to wavelet discretizations. Finally we present numerical results for the Poisson eigenvalue problem on an L-shaped domain.

We present an algebraically extended 2D image representation in this paper. In order to obtain more degrees of freedom, a 2D image is embedded into a certain geometric algebra. Combining methods of differential geometry, tensor algebra, monogenic signal and quadrature filter, the novel 2D image representation can be derived as the monogenic extension of a curvature tensor. The 2D spherical harmonics are employed as basis functions to construct the algebraically extended 2D image representation. From this representation, the monogenic signal and the monogenic curvature signal for modeling intrinsically one and two dimensional (i1D/i2D) structures are obtained as special cases. Local features of amplitude, phase and orientation can be extracted at the same time in this unique framework. Compared with the related work, our approach has the advantage of simultaneous estimation of local phase and orientation. The main contribution is the rotationally invariant phase estimation, which enables phase-based processing in many computer vision tasks.

The design of building projects involves several types of resources such as architects, structural engineers, mechanical engineers, electrical engineers, and draftsmen, among others. For design firms to stay in business in this very competitive market, they need to manage their resources in a way that improves productivity and cost effectiveness. This task, however, is not simple and requires thorough analysis of process-level operations, resource use, and productivity. Typically, these operational aspects are the responsibility of the design office manager who assigns available resources to the different design projects to save time and lower design expenses. It is noted that limited studies have been carried out in the literature to model overall organizational operations and behavioral aspects, particularly in firms specialized in the design of building projects. In an effort to simplify the modeling process, a simplified modeling and simulation tool is used in this research. A simulation model representing an actual design office was developed assuming that the office performs designs for small, medium, and large size building projects. The developed model was used to simulate several alternatives and examine various resource assignment strategies. The simulation was conducted over ten years and the resulting productivity and income was measured.

The application of partly decoupled approach by means of continuum mechanics facilitates the calculation of structural responses due to welding. The numerical results demonstrate the ability of a qualitative prediction of welded connections. As it is intended to integrate the local effects of a joint in structural analysis of steel constructions, it is necessary to meet higher approaches towards quality. The wide array of material parameters are presented, which are affecting the thermal, metallurgical and mechanical behavior, and which have to be identified. For that purpose further investigations are necessary to analyze the sensitivity of the models towards different material properties. The experimental determination of every material parameter is not possible due to the extraordinary laborious efforts needed. Besides that, experimentally identified parameters can be applied only for the tested steel quality for measured temperature-time regimes. For that reason alternative approaches for identification of material parameters, such as optimization strategies, have to be applied. After a definition of material parameters a quantitative prediction of welded connections will also be possible. Numerical results show the effect of phase transformation, activated by welding process, on residual stress state. As these phenomena occur in local areas in the range of crystal and grain sizes, the description of microscopic phenomena and their propagation on a macroscopic level due to approaches of homogenization might be expedient. Nevertheless, one should bear in mind, the increasing number of material parameters as well as the complexity of their experimental determination. Thus the microscopic approach should always be investigated under the scope of ability and efficiency of a required prediction. Under certain circumstances a step backwards, adopting a phenomenological approach, also can be beneficial.

The analysis of the response of complex structural systems requires the description of the material constitutive relations by means of an appropriate material model. The level of abstraction of such model may strongly affect the quality of the prognosis of the whole structure. In context to this fact, it is necessary to describe the material in a convenient sense as exact but as simple as possible. All material phenomena of crystalline materials e.g. steel, affecting the behavior of the structure, rely on physical effects which are interacting over spatial scales from subatomic to macroscopic range. Nevertheless, if the material is microscopically heterogenic, it might be appropriate to use phenomenological models for the purpose of civil engineering. Although constantly applied, these models are insufficient for steel materials with microscopic characteristics such as texture, typically occurring in hot rolled steel members or heat affected zones of welded joints. Hence, texture is manifested in crystalline materials as a regular crystallographic structure and crystallite orientation, influencing macroscopic material properties. The analysis of structural response of material with texture (e.g. rolled steel or heat affected zone of a welded joint) obliges the extension of the phenomenological material description of macroscopic scale by means of microscopic information. This paper introduces an enrichment approach for material models based on a hierarchical multiscale methodology. This has been done by describing the grain texture on a mesoscopic scale and coupling it with macroscopic constitutive relations by means of homogenization. Due to a variety of available homogenization methods, the question of an assessment of coupling quality arises. The applicability of the method and the effect of the coupling method on the reliability of the response are presented on an example.

Analysis of the reinforced concrete chimney geometry changes and their influence on the stresses in the chimney mantle was made. All the changes were introduced to a model chimney and compared. Relations between the stresses in the mantle of the chimney and the deformations determined by the change of the chimney's vertical axis geometry were investigated. The vertical axis of chimney was described by linear function (corresponding to the real rotation of the chimney together with the foundation), and by parabolic function (corresponding to the real dislocation of the chimney under the influence of the horizontal forces - wind). The positive stress pattern in the concrete as well as the negative stress pattern in the reinforcing steel have been presented. The two cases were compared. Analysis of the stress changes in the chimney mantle depending on the modification in the thickness of the mantle (the thickness of the chimney mantle was altered in the linear or the abrupt way) was carried out. The relation between the stresses and the chimney's diameter change from the bottom to the top of the chimney was investigated. All the analyses were conducted by means of a specially developed computer program created in Mathematica environment. The program makes it also possible to control calculations and to visualize the results of the calculations at every stage of the calculation process.

TOOL TO CHECK TOPOLOGY AND GEOMETRY FOR SPATIAL STRUCTURES ON BASIS OF THE EXTENDED MAXWELL'S RULE
(2006)

One of the simplest principle in the design of light-weight structures is to avoid bending. This can be achieved by dissolving girders into members acting purely in axial tension or compression. The employment of cables for the tensioned members leads to even lighter structures which are called cable-strut structures. They constitute a subclass of spatial structures. To give fast information about the general feasibility of an architectural concept employing cable-strut structures is a challenging task due to their sophisticated mechanical behavior. In this regard it is essential to control if the structure is stable and if pre-stress can be applied. This paper presents a tool using the spreadsheet software Microsoft (MS) Excel which can give such information. Therefore it is not necessary to purchase special software and the according time consuming training is much lower. The tool was developed on basis of the extended Maxwell's rule, which besides topology also considers the geometry of the structure. For this the rank of the node equilibrium matrix is crucial. Significance and determination of the rank and the implementation of the corresponding algorithms in MS Excel are described in the following. The presented tool is able to support the structural designer in an early stage of the project in finding a feasible architectural concept for cable-strut structures. As examples for the application of the software tool two special cable-strut structures, so called tensegrity structures, were examined for their mechanical behavior.

Projektsteuerung gewinnt als Aufgabenfeld bei der Realisierung von Investitionsprojekten ständig an Bedeutung. Zur Bewältigung der umfangreichen und komplizierten Aufgaben des Projektsteuerers wird zunehmend Software angeboten und genutzt. Dabei ist feststellbar, daß von einer integrierten Projektsteuerung unter Berücksichtigung der Kriterien Leistung, Termin und Kosten nicht gesprochen werden kann, weil wesentliche theoretische und praktische Voraussetzungen fehlen. Im Beitrag werden Defizite der praktizierten Vorgehensweise und Lösungsansätze für eine integrierte rechnergestützte Projektsteuerung vorgestellt. Durch geeignete Formen der Projektstrukturierung und eine Kopplung auf dem Mark befindlicher Standardsoftware können geeignete Lösungen gefunden werden. Problempunkte dabei stellen die Schnittstellen zwischen den Anwenderprogrammen die Datenbeschaffung und -verwaltung sowie geeignete Verfahren zur Kostenermittlung und -verfolgung dar. Im Beitrag werden eine bereits praktizierte und eine in Entwicklung befindliche Lösung vorgestellt.

Nodal integration of finite elements has been investigated recently. Compared with full integration it shows better convergence when applied to incompressible media, allows easier remeshing and highly reduces the number of material evaluation points thus improving efficiency. Furthermore, understanding it may help to create new integration schemes in meshless methods as well. The new integration technique requires a nodally averaged deformation gradient. For the tetrahedral element it is possible to formulate a nodal strain which passes the patch test. On the downside, it introduces non-physical low energy modes. Most of these "spurious modes" are local deformation maps of neighbouring elements. Present stabilization schemes rely on adding a stabilizing potential to the strain energy. The stabilization is discussed within this article. Its drawbacks are easily identified within numerical experiments: Nonlinear material laws are not well represented. Plastic strains may often be underestimated. Geometrically nonlinear stabilization greatly reduces computational efficiency. The article reinterpretes nodal integration in terms of imposing a nonconforming C0-continuous strain field on the structure. By doing so, the origins of the spurious modes are discussed and two methods are presented that solve this problem. First, a geometric constraint is formulated and solved using a mixed formulation of Hu-Washizu type. This assumption leads to a consistent representation of the strain energy while eliminating spurious modes. The solution is exact, but only of theoretical interest since it produces global support. Second, an integration scheme is presented that approximates the stabilization criterion. The latter leads to a highly efficient scheme. It can even be extended to other finite element types such as hexahedrals. Numerical efficiency, convergence behaviour and stability of the new method is validated using linear tetrahedral and hexahedral elements.

Subject of the paper is the realisation of a model based efficiency control system for PV generators using a simulation model. A standard 2-diodes model of PV generator is base of the ColSim model, which is implemented in ANSI C code for flexible code exporting. The algorithm is based on discretisized U-I characteristics, which allows the calculation of string topologies witch parallel and serial PV cells and modules. Shadowing effects can be modelled down to cell configuration using polar horizon definitions. The simulation model was ported to a real time environment, to calculate the efficiency of a PV system. Embedded System technology allows the networked operation and the integration of standard I/O devices. Futher work focus on the adaption of shadowing routine, which will be adapted to get the environment conditions from the real operation.

Die meisten Insolvenzen in Deutschland kommen aus der Bauindustrie. Die Gründe hierfür sind vielschichtig, jedoch kann mittels eines modern ausgerichteten M-I-S und Baustellen-Controllings frühzeitig erkannt werden, wie sich die Baustellenergebnisse entwickeln. Hierzu ist es notwendig, dass die Arbeitskalkulation ständig auf dem Laufenden gehalten wird. Nur wenn dies geschieht, sind monatliche Soll-/ Ist-Vergleiche und eine Betrachtung der cost-to-complete möglich und sinnvoll. Eine monatlich rollierende Prognose des Baustellenergebnisses zum Bauende ermöglicht, dass gravierende Veränderungen des Ergebnisses umgehend aufgedeckt werden. Nur in Kenntnis dieser Entwicklungen kann das Management frühzeitig (im Sinne eines Frühwarnsystems) agieren und Steuerungsmaßnahmen ergreifen. Die Ergebnisprognose zum Bauende ist allein als Steuerungsinstrument nicht ausreichend. Die Finanzsituation der Baustelle muß auch regelmäßig geprüft werden, d.h. der Leistungsstand mit der Rechnungsstellung an den Bauherren abgeglichen sowie die unbezahlten Rechnungen des Bauherren überprüft werden. Das beste Prognoseergebnis ist wertlos, wenn der Bauherr seine bezogenen Leistungen nicht vergütet. Die wirtschaftlichen Daten stehen den Verantwortlichen online im Baustellen-Informations-System (B-I-S) zur Verfügung. Ein Ampelsystem verdeutlicht die wirtschaftliche Lage der Baustelle.

Modern distributed engineering applications are based on complex systems consisting of various subsystems that are connected through the Internet. Communication and collaboration within an entire system requires reliable and efficient data exchange between the subsystems. Middleware developed within the web evolution during the past years provides reliable and efficient data exchange for web applications, which can be adopted for solving the data exchange problems in distributed engineering applications. This paper presents a generic approach for reliable and efficient data exchange between engineering devices using existing middleware known from web applications. Different existing middleware is examined with respect to the suitability in engineering applications. In this paper, a suitable middleware is shown and a prototype implementation simulating distributed wind farm control is presented and validated using several performance measurements.

Hinsichtlich der Integration einzelner Bauwerkslebensphasen und der verschiedenen Beteiligten, insbesondere innerhalb von Bauplanungs- und Revitalisierungsprozessen, bestehen aktuell entscheidende Defizite. Die generelle Zielstellung der in diesem Beitrag vorgestellten Forschungsarbeiten besteht in der Unterstützung und Verbesserung der Integration durch die disziplin- und lebensphasenübergreifende Bereitstellung sämtlicher bauwerksbezogener Informationen. Dies erfordert einerseits geeignete Ansätze zur Modellierung und Integration der vielfältigen disziplinspezifischen Daten, andererseits geeignete Lösungen, die einen globalen Zugriff, Navigation und Recherche im Gesamtdatenbestand ermöglichen. Die Modellierung und Verwaltung bauwerksbezogener Daten ist seit längerem Gegenstand diverser Forschungsarbeiten. Im Rahmen des SFB 524 wurde ein eigener Ansatz basierend auf einem laufzeitdynamischen Partialmodellverbund entwickelt. Dieser wird in den wesentlichen Grundzügen anderen Ansätzen gegenübergestellt. Den Schwerpunkt dieses Beitrags bildet jedoch die Entwicklung einer geeigneten flexiblen Navigations- und Rechercheschicht zu Realisierung projektglobaler Informationsrecherche. Aus der Sicht der Modellierung und Datenverwaltung wie auch aus der Sicht der Informationsrecherche und Informationspräsentation in Planungsprozessen ergeben sich verschiedene Anforderungen an derartige Recherchewerkzeuge, wobei der wesentlichste Grundsatz maximale Flexibilität hinsichtlich verfügbarer Darstellungstechniken und deren freie Kombination mit Techniken formaler Suchanfragen ist. Das entwickelte Systemkonzept basiert auf einem Framework, welches verschiedene Grundtypen von Recherchemodulen und deren Interaktionsprinzipien vorgibt. Einzelne Recherchemodule werden als Ausprägungen dieser Modultypen realisiert und können je nach Bedarf laufzeitdynamisch in die Navigationsschicht integriert werden. Die technische Realisierung des Systems erfolgt im Umfeld vorhandener Prototypen aus vorangegangenen Forschungsaktivitäten. Dieses technische Umfeld gibt verschiedene Rahmenbedingungen vor, welche im Vorfeld prototypischer Implementierungen verschiedene Adaptionen des generellen Systemkonzepts notwendig machen. Der vorliegende Beitrag stellt den aktuellen Entwicklungsstand der Systemlösung aus konzeptioneller und technischer Sicht sowie erste prototypische Realisierungen von Recherchemodulen vor.

The design of safety-critical structures, exposed to cyclic excitations demands for non-degrading or limited-degrading behavior during extreme events. Among others, the structural behavior is mainly determined by the amount of plastic cycles, completed during the excitation. Existing simplified methods often ignore this dependency, or assume/request sufficient cyclic capacity. The paper introduces a new performance based design method that considers explicitly a predefined number of re-plastifications. Hereby approaches from the shakedown theory and signal processing methods are utilized. The paper introduces the theoretical background, explains the steps of the design procedure and demonstrates the applicability with help of an example. This project was supported by German Science Foundation (Deutsche Forschungsgemeinschaft, DFG)

For the dynamic behavior of lightweight structures like thin shells and membranes exposed to fluid flow the interaction between the two fields is often essential. Computational fluid-structure interaction provides a tool to predict this interaction and complement or eventually replace expensive experiments. Partitioned analyses techniques enjoy great popularity for the numerical simulation of these interactions. This is due to their computational superiority over simultaneous, i.e. fully coupled monolithic approaches, as they allow the independent use of suitable discretization methods and modular analysis software. We use, for the fluid, GLS stabilized finite elements on a moving domain based on the incompressible instationary Navier-Stokes equations, where the formulation guarantees geometric conservation on the deforming domain. The structure is discretized by nonlinear, three-dimensional shell elements.
Commonly used sequential staggered coupling schemes may exhibit instabilities due to the so-called artificial added mass effect. As best remedy to this problem subiterations should be invoked to guarantee kinematic and dynamic continuity across the fluid-structure interface. Since iterative coupling algorithms are computationally very costly, their convergence rate is very decisive for their usability. To ensure and accelerate the convergence of this iteration the updates of the interface position are relaxed. The time dependent, 'optimal' relaxation parameter is determined automatically without any user-input via exploiting a gradient method or applying an Aitken iteration scheme.

Steel structural design is an integral part of the building construction process. So far, various methods of design have been applied in practice to satisfy the design requirements. This paper attempts to acquire the Differential Evolution Algorithms in automatization of specific synthesis and rationalization of design process. The capacity of the Differential Evolution Algorithms to deal with continuous and/or discrete optimization of steel structures is also demonstrated. The goal of this study is to propose an optimal design of steel frame structures using built-up I-sections and/or a combination of standard hot-rolled profiles. All optimized steel frame structures in this paper generated optimization solutions better than the original solution designed by the manufacturer. Taking the criteria regarding the quality and efficiency of the practical design into consideration, the produced optimal design with the Differential Evolution Algorithms can completely replace conventional design because of its excellent performance.

A practical framework for generating cross correlated fields with a specified marginal distribution function, an autocorrelation function and cross correlation coefficients is presented in the paper. The contribution promotes a recent journal paper [1]. The approach relies on well known series expansion methods for simulation of a Gaussian random field. The proposed method requires all cross correlated fields over the domain to share an identical autocorrelation function and the cross correlation structure between each pair of simulated fields to be simply defined by a cross correlation coefficient. Such relations result in specific properties of eigenvectors of covariance matrices of discretized field over the domain. These properties are used to decompose the eigenproblem which must normally be solved in computing the series expansion into two smaller eigenproblems. Such decomposition represents a significant reduction of computational effort. Non-Gaussian components of a multivariate random field are proposed to be simulated via memoryless transformation of underlying Gaussian random fields for which the Nataf model is employed to modify the correlation structure. In this method, the autocorrelation structure of each field is fulfilled exactly while the cross correlation is only approximated. The associated errors can be computed before performing simulations and it is shown that the errors happen especially in the cross correlation between distant points and that they are negligibly small in practical situations.

The reduction of oscillation amplitudes of structural elements is necessary not only for maintenance of their durability and longevity but also for elimination of a harmful effect of oscillations on people and technology operations. The dampers are widely applied for this purpose. One of the most widespread models of structural friction forces having piecewise linear relation to displacement was analysed. T The author suggests the application of phase trajectories mapping in plane "acceleration – displacement". Unlike the trajectories mapping in a plane "velocity – displacement", they don't require large number of geometrical constructions for identification of the characteristics of dynamic systems. It promotes improving the accuracy. The analytical assumptions had been verified by numerical modeling. The results show good enough coincide between numerical and analytical estimation of dissipative characteristic.

What is nowadays called (classic) Clifford analysis consists in the establishment of a function theory for functions belonging to the kernel of the Dirac operator. While such functions can very well describe problems of a particle with internal SU(2)-symmetries, higher order symmetries are beyond this theory. Although many modifications (such as Yang-Mills theory) were suggested over the years they could not address the principal problem, the need of a n-fold factorization of the d’Alembert operator. In this paper we present the basic tools of a fractional function theory in higher dimensions, for the transport operator (alpha = 1/2 ), by means of a fractional correspondence to the Weyl relations via fractional Riemann-Liouville derivatives. A Fischer decomposition, fractional Euler and Gamma operators, monogenic projection, and basic fractional homogeneous powers are constructed.

The stress state of a piecewise-homogeneous elastic body, which has a semi-infinite crack along the interface, under in-plane and antiplane loads is considered. One of the crack edges is reinforced by a rigid patch plate on a finite interval adjacent to the crack tip. The crack edges are loaded with specified stresses. The body is stretched at infinity by specified stresses. External forces with a given principal vector and moment act on the patch plate. The problem reduces to a Riemann-Hilbert boundary-value matrix problem with a piecewise-constant coefficient for two complex potentials in the plane case and for one in the antiplane case. The complex potentials are found explicitly using a Gaussian hypergeometric function. The stress state of the body close to the ends of the patch plate, one of which is also simultaneously the crack tip, is investigated. Stress intensity factors near the singular points are determined.

The extended finite element method (XFEM) offers an elegant tool to model material discontinuities and cracks within a regular mesh, so that the element edges do not necessarily coincide with the discontinuities. This allows the modeling of propagating cracks without the requirement to adapt the mesh incrementally. Using a regular mesh offers the advantage, that simple refinement strategies based on the quadtree data structure can be used to refine the mesh in regions, that require a high mesh density. An additional benefit of the XFEM is, that the transmission of cohesive forces through a crack can be modeled in a straightforward way without introducing additional interface elements. Finally different criteria for the determination of the crack propagation angle are investigated and applied to numerical tests of cracked concrete specimens, which are compared with experimental results.

PARAMETER IDENTIFICATION OF MESOSCALE MODELS FROM MACROSCOPIC TESTS USING BAYESIAN NEURAL NETWORKS
(2010)

In this paper, a parameter identification procedure using Bayesian neural networks is proposed. Based on a training set of numerical simulations, where the material parameters are simulated in a predefined range using Latin Hypercube sampling, a Bayesian neural network, which has been extended to describe the noise of multiple outputs using a full covariance matrix, is trained to approximate the inverse relation from the experiment (displacements, forces etc.) to the material parameters. The method offers not only the possibility to determine the parameters itself, but also the accuracy of the estimate and the correlation between these parameters. As a result, a set of experiments can be designed to calibrate a numerical model.

The article presents analysis of stress distribution in the reinforced concrete support beam bracket which is a component of prefabricated reinforced concrete building. The building structure is spatial frame where dilatations were applied. The proper stiffness of its structure is provided by frames with stiff joints, monolithic lift shifts and staircases. The prefabricated slab floors are supported by beam shelves which are shaped as inverted letter ‘T’. Beams are supported by the column brackets. In order to lower the storey height and fulfill the architectural demands at the same time, the designer lowered the height of beam at the support zone. The analyzed case refers to the bracket zone where the slant crack. on the support beam bracket was observed. It could appear as a result of overcrossing of allowable tension stresses in reinforced concrete, in the bracket zone. It should be noted that the construction solution applied, i.e. concurrent support of the “undercut” beam on the column bracket causes local concentration of stresses in the undercut zone where the strongest transverse forces and tangent stresses occur concurrently. Some additional rectangular stresses being a result of placing the slab floors on the lower part of beam shelves sum up with those described above.

The paper contains a description of dynamic effects in the silo wall during the outflow of a stored material. The work allows for determining the danger of construction damage due to resonant vibrations and is of practical importance by determining the influence of cyclic pressures and vibro–creeping during prolonged use of a silo. The paper was devised as a result of tests on silo walls in semi-technical scale. The model is generally applicable and allows for identification of parameters in real- size silos as well.

The paper proposes a new method for general 3D measurement and 3D point reconstruction. Looking at its features, the method explicitly aims at practical applications. These features especially cover low technical expenses and minimal user interaction, a clear problem separation into steps that are solved by simple mathematical methods (direct, stable and optimal with respect to least error squares), and scalability. The method expects the internal and radial distortion parameters of the used camera(s) as inputs, and a plane quadrangle with known geometry within the scene. At first, for each single picture the 3D position of the reference quadrangle (with respect to each camera coordinate frame) is calculated. These 3D reconstructions of the reference quadrangle are then used to yield the relative external parameters of each camera regarding the first one. With known external parameters, triangulation is finally possible. The differences from other known procedures are outlined, paying attention to the stable mathematical methods (no usage of nonlinear optimization) and the low user interaction with good results at the same time.

Since the 90-ties the Pascal matrix, its generalizations and applications have been in the focus of a great amount of publications. As it is well known, the Pascal matrix, the symmetric Pascal matrix and other special matrices of Pascal type play an important role in many scientific areas, among them Numerical Analysis, Combinatorics, Number Theory, Probability, Image processing, Sinal processing, Electrical engineering, etc. We present a unified approach to matrix representations of special polynomials in several hypercomplex variables (new Bernoulli, Euler etc. polynomials), extending results of H. Malonek, G.Tomaz: Bernoulli polynomials and Pascal matrices in the context of Clifford Analysis, Discrete Appl. Math. 157(4)(2009) 838-847. The hypercomplex version of a new Pascal matrix with block structure, which resembles the ordinary one for polynomials of one variable will be discussed in detail.

From the design experiences of arch dams in the past, it has significant practical value to carry out the shape optimization of arch dams, which can fully make use of material characteristics and reduce the cost of constructions. Suitable variables need to be chosen to formulate the objective function, e.g. to minimize the total volume of the arch dam. Additionally a series of constraints are derived and a reasonable and convenient penalty function has been formed, which can easily enforce the characteristics of constraints and optimal design. For the optimization method, a Genetic Algorithm is adopted to perform a global search. Simultaneously, ANSYS is used to do the mechanical analysis under the coupling of thermal and hydraulic loads. One of the constraints of the newly designed dam is to fulfill requirements on the structural safety. Therefore, a reliability analysis is applied to offer a good decision supporting for matters concerning predictions of both safety and service life of the arch dam. By this, the key factors which would influence the stability and safety of arch dam significantly can be acquired, and supply a good way to take preventive measures to prolong ate the service life of an arch dam and enhances the safety of structure.

In this paper proposed the application of two-parameters damage model, based on non-linear finite element approach, to the analysis of masonry panels. Masonry is treated as a homogenized material, for which the material characteristics can be defined by using homogenization technique. The masonry panels subjected to shear loading are studied by using the proposed procedure within the framework of three-dimensional analyses. The nonlinear behaviour of masonry can be modelled using concepts of damage theory. In this case an adequate damage function is defined for taking into account different response of masonry under tension and compression states. Cracking can, therefore, be interpreted as a local damage effect, defined by the evolution of known material parameters and by one or several functions which control the onset and evolution of damage. The model takes into account all the important aspects which should be considered in the nonlinear analysis of masonry structures such as the effect of stiffness degradation due to mechanical effects and the problem of objectivity of the results with respect to the finite element mesh. Finally the proposed damage model is validated with a comparison with experimental results available in the literature.

This paper deals with the modelling and the analysis of masonry vaults. Numerical FEM analyses are performed using LUSAS code. Two vault typologies are analysed (barrel and cross-ribbed vaults) parametrically varying geometrical proportions and constraints. The proposed model and the developed numerical procedure are implemented in a computer analysis. Numerical applications are developed to assess the model effectiveness and the efficiency of the numerical procedure. The main object of the present paper is the development of a computational procedure which allows to define 3D structural behaviour of masonry vaults. For each investigated example, the homogenized limit analysis approach has been employed to predict ultimate load and failure mechanisms. Finally, both a mesh dependence study and a sensitivity analysis are reported. Sensitivity analysis is conducted varying in a wide range mortar tensile strength and mortar friction angle with the aim of investigating the influence of the mechanical properties of joints on collapse load and failure mechanisms. The proposed computer model is validated by a comparison with experimental results available in the literature.

This paper presents results of applying Fuzzy Inference System for estimation of the number of potential Park and Ride users. Usually it is difficult to evaluate the number of users because it depends on human factor and data in the considered system are uncertain. In such situation the traditional mathematical approaches can not take into consideration rough data. Therefore a fuzzy approach can be applied in this case. A fuzzy methodology is treated as a proper way to describe choice of mode of transport, and especially that uncertainty accompanied of choosing process has rather fuzzy character. The proposed approach is based on the Mamdani Fuzzy Inference System and for calculation there is used Matlab software with Fuzzy Logic Toolbox. Mamdani model requires, as an input data, knowledge of the shape of membership function. These functions can be calibrated taking into consideration results of questionnaires conducted among users of Park and Ride system. Due to lack of representative sample of users, one has decided to use results of experts' questionnaires as a input data for calibration the shape of membership functions. Describing factor will be generalized cost of the trip for different modes of transport. Proposed approach consists of two main stages: modeling of share of public/private transport trips and Multimodal model estimating number of Park and Ride users. Verification of presented methodology is treated as an indirect proof. Proposed approach can be applied for estimation of bi-modal split. Then the results are compared with traditional approaches based on logit functions. Comparable results of proposed fuzzy approach with traditional logit models can be treated as a confirmation of chosen methodology.

A concept of non-commutative Galois extension is introduced and binary and ternary extensions are chosen. Non-commutative Galois extensions of Nonion algebra and su(3) are constructed. Then ternary and binary Clifford analysis are introduced for non-commutative Galois extensions and the corresponding Dirac operators are associated.

The aim of this study is to show an application of model robustness measures for soilstructure interaction (henceforth written as SSI) models. Model robustness defines a measure for the ability of a model to provide useful model answers for input parameters which typically have a wide range in geotechnical engineering. The calculation of SSI is a major problem in geotechnical engineering. Several different models exist for the estimation of SSI. These can be separated into analytical, semi-analytical and numerical methods. This paper focuses on the numerical models of SSI specific macro-element type models and more advanced finite element method models using contact description as continuum or interface elements. A brief description of the models used is given in the paper. Following this description, the applied SSI problem is introduced. The observed event is a static loaded shallow foundation with an inclined load. The different partial models to consider the SSI effects are assessed using different robustness measures during numerical application. The paper shows the investigation of the capability to use these measures for the assessment of the model quality of SSI partial models. A variance based robustness and a mathematical robustness approaches are applied. These different robustness measures are used in a framework which allows also the investigation of computational time consuming models. Finally the result shows that the concept of using robustness approaches combined with other model–quality indicators (e.g. model sensitivity or model reliability) can lead to unique model–quality assessment for SSI models.

The sizing of simple resonators like guitar strings or laser mirrors is directly connected to the wavelength and represents no complex optimisation problem. This is not the case with liquid-filled acoustic resonators of non-trivial geometries, where several masses and stiffnesses of the structure and the fluid have to fit together. This creates a scenario of many competing and interacting resonances varying in relative strength and frequency when design parameters change. Hence, the resonator design involves a parameter-tuning problem with many local optima. As its solution evolutionary algorithms (EA) coupled to a forced-harmonic FE simulation are presented. A new hybrid EA is proposed and compared to two state-of-theart EAs based on selected test problems. The motivating background is the search for better resonators suitable for sonofusion experiments where extreme states of matter are sought in collapsing cavitation bubbles.

In this paper we evaluate 2D models for soil-water characteristic curve (SWCC), that incorporate the hysteretic nature of the relationship between volumetric water content θ and suction ψ. The models are based on nonlinear least squares estimation of the experimental data for sand. To estimate the dependent variable θ the proposed models include two independent variables, suction and sensors reading position (depth d in the column test). The variable d represents not only the position where suction and water content are measured but also the initial suction distribution before each of the hydraulic loading test phases. Due to this the proposed 2D regression models acquire the advantage that they: (a) can be applied for prediction of θ for any position along the column and (b) give the functional form for the scanning curves.

Building information modeling offers a huge potential for increasing the productivity and quality of construction planning processes. Despite its promising concept, this approach has not found widespread use. One of the reasons is the insufficient coupling of the structural models with the general building model. Instead, structural engineers usually set up a structural model that is independent from the building model and consists of mechanical models of reduced dimension. An automatic model generation, which would be valuable in case of model revisions is therefore not possible. This can be overcome by a volumetric formulation of the problem. A recent approach employed the p-version of the finite element method to this problem. This method, in conjunction with a volumetric formulation is suited to simulate the structural behaviour of both „thick“ solid bodies and thin-walled structures. However, there remains a notable discretization error in the numerical models. This paper therefore proposes a new approach for overcoming this situation. It sugggests the combination of the Isogeometric analysis together with the volumetric models in order to integrate the structural design into the digital, building model-centered planning process and reduce the discretization error. The concept of the isogeometric analysis consists, roughly, in the application of NURBS functions to represent the geometry and the shape functions of the elements. These functions possess some beneficial properties regarding numerical simulation. Their use, however, leads to some intricacies related to the setup of the stiffness matrix. This paper describes some of these properties.

Optimum technological solutions must take into account the entire life cycle of structures including design procedures as well as quality assurance, inspection, maintenance, and repair strategies. Unfortunately, current design standards do not provide a satisfactory basis to ensure expected structural lifetimes. The latter may vary from only a few years for temporary structures to over a century for bridges, water dams or nuclear repositories. Consistent scientific concepts are urgently required to cover this wide spectrum of lifetimes in structural design and maintenance. This was a motivation for a group of scientists at the Ruhr-University Bochum (RUB) to start a special research program supported by the German Research Foundation (DFG) within the Cooperative Research Center SFB 398 since 1996. Institutes of the University Wuppertal and of the University Essen-Duisburg joined the research group. The goal of the Center is to study sources of damage and deterioration in materials and structures, to develop consistent models and simulation methods, to predict structural lifetimes and finally to integrate this predictions into new lifetime-oriented design strategies.
Research activities in our center are organised in three Project Groups as follows:
- Modelling of lifetime effects
- Methods for lifetime-oriented structural analyses
- Future lifespan-oriented design strategies.

Information technology plays a key role in the everyday operation of buildings and campuses. Many proprietary technologies and methodologies can assist in effective Building Performance Monitoring (BPM) and efficient managing of building resources. The integration of related tools like energy simulator packages, facility, energy and building management systems, and enterprise resource planning systems is of benefit to BPM. However, the complexity to integrating such domain specific systems prevents their common usage. Service Oriented Architecture (SOA) has been deployed successfully in many large multinational companies to create integrated and flexible software systems, but so far this methodology has not been applied broadly to the field of BPM. This paper envisions that SOA provides an effective integration framework for BPM. Service oriented architecture for the ITOBO framework for sustainable and optimised building operation is proposed and an implementation for a building performance monitoring system is introduced.

DIGITAL SUPPORT OF MATERIAL- AND PRODUCT SELECTION IN THE ARCHITECTURAL DESIGN- AND PLANNING PROCESS
(2006)

Architecture is predominantly perceived over the surfaces limiting the space. The used surface materials thereby should support the design intention and have to fulfil various technical and economical requirements. If the architect wants to select the "right" or the "best" material he has to play with very different and sometimes contradicting criteria and must weight these individually for the special purpose. This selection process is supported only insufficiently by today's digital systems. If it would be possible to illustrate all the various parameters by numerical values, the method of multidimensional scaling will offer a solution for architects to find the material which is best fitting on basis of his individual weighting of criteria. By displaying the result of the architect's multidimensional query in a spatial arrangement multidimensional scaling can support an interactive selection process with additional feedback over the applied search strategy.

The paper is devoted to a study of properties of homogeneous solutions of massless field equation in higher dimensions. We first treat the case of dimension 4. Here we use the two-component spinor language (developed for purposes of general relativity). We describe how are massless field operators related to a higher spin analogues of the de Rham sequence - the so called Bernstein-Gel'fand-Gel'fand (BGG) complexes - and how are they related to the twisted Dirac operators. Then we study similar question in higher (even) dimensions. Here we have to use more tools from representation theory of the orthogonal group. We recall the definition of massless field equations in higher dimensions and relations to higher dimensional conformal BGG complexes. Then we discuss properties of homogeneous solutions of massless field equation. Using some recent techniques for decomposition of tensor products of irreducible $Spin(m)$-modules, we are able to add some new results on a structure of the spaces of homogenous solutions of massless field equations. In particular, we show that the kernel of the massless field equation in a given homogeneity contains at least on specific irreducible submodule.

We consider a structural truss problem where all of the physical model parameters are uncertain: not just the material values and applied loads, but also the positions of the nodes are assumed to be inexact but bounded and are represented by intervals. Such uncertainty may typically arise from imprecision during the process of manufacturing or construction, or round-off errors. In this case the application of the finite element method results in a system of linear equations with numerous interval parameters which cannot be solved conventionally. Applying a suitable variable substitution, an iteration method for the solution of a parametric system of linear equations is firstly employed to obtain initial bounds on the node displacements. Thereafter, an interval tightening (pruning) technique is applied, firstly on the element forces and secondly on the node displacements, in order to obtain tight guaranteed enclosures for the interval solutions for the forces and displacements.

IFC-BASED MONITORING INFORMATION MODELING FOR DATA MANAGEMENT IN STRUCTURAL HEALTH MONITORING
(2015)

This conceptual paper discusses opportunities and challenges towards the digital representation of structural health monitoring systems using the Industry Foundation Classes (IFC) standard. State-of-the-art sensor nodes, collecting structural and environmental data from civil infrastructure systems, are capable of processing and analyzing the data sets directly on-board the nodes. Structural health monitoring (SHM) based on sensor nodes that possess so called “on-chip intelligence” is, in this study, referred to as “intelligent SHM”, and the infrastructure system being equipped with an intelligent SHM system is referred to as “intelligent infrastructure”. Although intelligent SHM will continue to grow, it is not possible, on a well-defined formalism, to digitally represent information about sensors, about the overall SHM system, and about the monitoring strategies being implemented (“monitoring-related information”). Based on a review of available SHM regulations and guidelines as well as existing sensor models and sensor modeling languages, this conceptual paper investigates how to digitally represent monitoring-related information in a semantic model. With the Industry Foundation Classes, there exists an open standard for the digital representation of building information; however, it is not possible to represent monitoring-related information using the IFC object model. This paper proposes a conceptual approach for extending the current IFC object model in order to include monitoring-related information. Taking civil infrastructure systems as an illustrative example, it becomes possible to adequately represent, process, and exchange monitoring-related information throughout the whole life cycle of civil infrastructure systems, which is referred to as monitoring information modeling (MIM). However, since this paper is conceptual, additional research efforts are required to further investigate, implement, and validate the proposed concepts and methods.

Due to increasing numbers of wind energy converters, the accurate assessment of the lifespan of their structural parts and the entire converter system is becoming more and more paramount. Lifespan-oriented design, inspections and remedial maintenance are challenging because of their complex dynamic behavior. Wind energy converters are subjected to stochastic turbulent wind loading causing corresponding stochastic structural response and vibrations associated with an extreme number of stress cycles (up to 109 according to the rotation of the blades). Currently, wind energy converters are constructed for a service life of about 20 years. However, this estimation is more or less made by rule of thumb and not backed by profound scientific analyses or accurate simulations. By contrast, modern structural health monitoring systems allow an improved identification of deteriorations and, thereupon, to drastically advance the lifespan assessment of wind energy converters. In particular, monitoring systems based on artificial intelligence techniques represent a promising approach towards cost-efficient and reliable real-time monitoring. Therefore, an innovative real-time structural health monitoring concept based on software agents is introduced in this contribution. For a short time, this concept is also turned into a real-world monitoring system developed in a DFG joint research project in the authors’ institute at the Ruhr-University Bochum. In this paper, primarily the agent-based development, implementation and application of the monitoring system is addressed, focusing on the real-time monitoring tasks in the deserved detail.

The concept is presented of the sensitivity analysis of the limit state of the structure with respect to selected basic variables. The sensitivity is presented in the form of the probability distribution of the limit state of the structure. The analysis is performed by the problem-oriented Monte Carlo simulation procedure. The procedure is based on the problem's definition of the elementary event, as a structural limit state. Thus the sample space consists of limit states of the structure. Defined on the sample space the one-dimensional random multiplier is introduced. This multiplier refers to the dominant basic variable (group of variables) of the problem. Numerical procedure results in the set of random numbers. Normalized relative histogram of this set is an estimator of the PDF of the limit state of the structure. Estimators of reliability, or the probability of failure are statistical characteristics of this histogram. The procedure is illustrated by the example of sensitivity analysis of the serviceability limit state of monumental structure. It is the colonnade of Licheń Basilica, situated in central Poland. Limit state of the structure is examined with reference to the upper deck horizontal deflection. Wind actions are taken as dominant variables. An assumption is made that the wind load intensities acting on the lower and on the upper storey of the colonnade, respectively, are identically distributed, but correlated random variables. Three correlation variants of these variables are considered. Relevant limit state histograms are analysed thereafter. The paper ends with the conclusions referring to the method and some general remarks on the fully probabilistic design.

The Bernstein polynomials are used for important applications in many branches of Mathematics and the other sciences, for instance, approximation theory, probability theory, statistic theory, num- ber theory, the solution of the di¤erential equations, numerical analysis, constructing Bezier curves, q-calculus, operator theory and applications in computer graphics. The Bernstein polynomials are used to construct Bezier curves. Bezier was an engineer with the Renault car company and set out in the early 1960’s to develop a curve formulation which would lend itself to shape design. Engineers may …nd it most understandable to think of Bezier curves in terms of the center of mass of a set of point masses. Therefore, in this paper, we study on generating functions and functional equations for these polynomials. By applying these functions, we investigate interpolation function and many properties of these polynomials.

In order to model and simulate collapses of large scale complex structures, a user-friendly and high performance software system is essential. Because a large number of simulation experiments have to be performed, therefore, next to an appropriate simulation model and high performance computing, efficient interactive control and visualization capabilities of model parameters and simulation results are crucial. To this respect, this contribution is concerned with advancements of the software system CADCE (Computer Aided Demolition using Controlled Explosives) that is extended under particular consideration of computational steering concepts. Thereby, focus is placed on problems and solutions for the collapse simulation of real world large scale complex structures. The simulation model applied is based on a multilevel approach embedding finite element models on a local as well as a near field length scale, and multibody models on a global scale. Within the global level simulation, relevant effects of the local and the near field scale, such as fracture and failure processes of the reinforced concrete parts, are approximated by means of tailor-made multibody subsystems. These subsystems employ force elements representing nonlinear material characteristics in terms of force/displacement relationships that, in advance, are determined by finite element analysis. In particular, enhancements concerning the efficiency of the multibody model and improvements of the user interaction are presented that are crucial for the capability of the computational steering. Some scenarios of collapse simulations of real world large scale structures demonstrate the implementation of the above mentioned approaches within the computational steering.

VARIATION OF ROTATIONAL RESTRAINT IN GRID DECK CONNECTION DUE TO CORROSION DAMAGE AND STRENGTHENING
(2006)

The approach to assessment of rotational restraint of stringer-to-crossbeam connection in a deck of 100-year old steel truss bridge is presented. Sensitivity of rotational restraint coefficient of the connection to corrosion damage and strengthening is analyzed. Two criteria of the assessment of the rotational restraint coefficient are applied: static and kinematic one. The former is based on bending moment distribution in the considered member, the latter one – on the member rotation at the given joint. 2D-element model of finite element method is described: webs and flanges are modeled with shell elements, while rivets in the connection – with system of beam and spring elements. The method of rivet modeling is verified by T-stub connection test results published in literature. FEM analyses proved that recorded extent of corrosion damage does not alter the initial rotational restraint of stringer-to-crossbeam connection. Strengthening of stringer midspan influences midspan bending moment and stringer end rotation in a different way. Usually restoring member load bearing capacity means strengthening its critical regions (where the highest stress levels occur). This alters flexural stiffness distribution over member length and influences rotational restraint at its connection to other members. The impact depends on criterion chosen for rotational restraint coefficient assessment.

CRITICAL STRESS ASSESSMENT IN ANGLE TO GUSSET PLATE BOLTED CONNECTION BY SIMPLIFIED FEM MODELLING
(2010)

Simplified modelling of friction grip bolted connections of steel member – to – gusset plate is often applied in engineering practise. The paper deals with the simplification of pre-tensioned bolt model and simplification of load transfer within connection. Influence on normal strain (and thus stress) distribution at critical cross-section is investigated. Laboratory testing of single-angle or double-angle members – to – gusset plates bolted connections were taken as basis for numerical analysis. FE models were created using 1D and 2D elements. Angles and gusset plates were modelled with shell elements. Two methods of modelling of friction grip bolting were considered: bolt-regarding approach with 1D element systems modelling bolts and two variants of bolt-disregarding approach with special constraints over some part of member and gusset plate surfaces in contact: a) constraints over whole area of contact, b) constraints over the area around each bolt shank (“partially tied”). Modelling of friction grip bolted connections using simplified bolt modelling may be effective, especially in the case of analysis concerning elastic range only. In such a case disregarding bolts and replacing them with “partially tied” modelling seems to be more attractive. It is less time-consuming and provides results of similar accuracy in comparison to analysis utilizing simplified bolt modelling.

The uncertainty existing in the construction industry is bigger than in other industries. Consequently, most construction projects do not go totally as planned. The project management plan needs therefore to be adapted repeatedly within the project lifecycle to suit the actual project conditions. Generally, the risks of change in the project management plan are difficult to be identified in advance, especially if these risks are caused by unexpected events such as human errors or changes in the client preferences. The knowledge acquired from different resources is essential to identify the probable deviations as well as to find proper solutions to the faced change risks. Hence, it is necessary to have a knowledge base that contains known solutions for the common exceptional cases that may cause changes in each construction domain. The ongoing research work presented in this paper uses the process modeling technique of Event-driven Process Chains to describe different patterns of structure changes in the schedule networks. This results in several so called “change templates”. Under each template different types of change risk/ response pairs can be categorized and stored in a knowledge base. This knowledge base is described as an ontology model populated with reference construction process data. The implementation of the developed approach can be seen as an iterative scheduling cycle that will be repeated within the project lifecycle as new change risks surface. This can help to check the availability of ready solutions in the knowledge base for the situation at hand. Moreover, if the solution is adopted, CPSP, “Change Project Schedule Plan „a prototype developed for the purpose of this research work, will be used to make the needed structure changes of the schedule network automatically based on the change template. What-If scenarios can be implemented using the CPSP prototype in the planning phase to study the effect of specific situations without endangering the success of the project objectives. Hence, better designed and more maintainable project schedules can be achieved.

Monitoring und Bewertung sind Hauptaufgaben im Management bzw. der Revitalisierung von Bauwerken. Unterschiedliche Verfahren können bei der Akquisition der erforderlichen geometrischen Information, wie z. B. Größe oder Verformung eines Gebäudes, eingesetzt werden. Da das Potenzial der digitalen Fotografie kontinuierlich wächst, stellt die Industriephotogrammetrie heute eine bedeutende Alternative zu den klassischen Verfahren wie Dehnmessstreifen oder anderen taktilen Sensoren dar. Moderne Industriephotogrammetrie erfasst die Bilder mittels digitaler Systeme. Dies bedeutet, dass die Information digitaler Bilder mit Hilfe der digitalen Bildverarbeitung untersucht werden muss, um die Bildkoordinaten der Messpunkte zu erhalten. Eine der Aufgaben der Bildverarbeitung für photogrammetrische Zwecke besteht somit darin, den Mittelpunkt von kreisförmigen Marken zu lokalisieren. Die modernen Operatoren liefern Subpixelgenauigkeit für die Koordinaten des Punktes. Das optische Messverfahren der Industriephotogrammetrie erfordert hinsichtlich der Hardware in erster Linie hochauflösende digitale Kameras. Dabei lassen sich die Kameras in Videokameras, HighSpeed-Kameras, intelligente Kameras sowie so genannte Consumer und Professionelle Kameras unterscheiden. Die geometrische Auflösung digitaler HighEnd-Kameras liegt heute bei über 10 Megapixel. In punkto Datentransfer zum Rechner sind verschiedene Standards am Markt verfügbar, z. B. USB2.0, GigE-Vision, CameraLink oder Firewire. Die Wahl des Standards hängt immer von der spezifischen Aufgabenstellung ab, da keine der Techniken eine führende Position einnimmt. Die moderne Photogrammetrie bietet viele neue Möglichkeiten für das Monitoring und die Bewertung von Bauwerken. Sie kann ein-, zwei-, drei- oder vierdimensionale Informationen liefern, falls erforderlich auch in Echtzeit. Als berührungsloses Messverfahren ist der Einsatz der Photogrammetrie noch möglich, wenn die taktilen Sensoren z. B. aufgrund ihres Platzbedarfes nicht mehr eingesetzt werden können. Hochauflösende Videokameras erlauben es, selbst dynamische Untersuchungen mit großer Präzision durchzuführen.

Car following models are used to describe the behavior of a number of cars on the road dependent on the distance to the car in front. We introduce a system of ordinary differential equations and perform a theoretical and numerical analysis in order to find solutions that reflect various traffic situations. We present three different variations of the model motivated by reality.

Die Kommunale Wohnungsgesellschaft mbH Erfurt(KoWo) ist mit ihren rund 20.000 Wohnungen in der Landeshauptstadt das größte Wohnungsunternehmen in Thüringen. Der Immobilienbestand ist heterogen in seinem technischen Zustand und im Bezug auf die unterschiedlichen Lagen der Objekte. Bedingt durch Leerstände und unterschiedliche Modernisierungsmaßnahmen und -stände unterscheidet sich die Wirtschaftlichkeit verschiedener Objekte deutlich. Ohne eine einheitliche Einwertung des Immobilienbestandes im Bezug auf die Objektattraktivität, die Standortqualität und die Objektwirtschaftlichkeit fällt eine langfristige strategische Entwicklung des Immobilienportfolios schwer. Über die Schritte der technischen Bestandserfassung, die Einwertung über ein Scorintmodell, die Abbildung in einem Portfoliomodell mit zugehöriger Normstrategie bis hin zur Weiterverarbeitung der Daten in der 20-jährigen Instandsetzungsplanung wird praxisnah aufgezeigt, wie die Vorgehensweise bei der Einwertung des Immobilienportfolios ist.

The numerical simulation of microstructure models in 3D requires, due to enormous d.o.f., significant resources of memory as well as parallel computational power. Compared to homogeneous materials, the material hetrogeneity on microscale induced by different material phases demand for adequate computational methods for discretization and solution process of the resulting highly nonlinear problem. To enable an efficient/scalable solution process of the linearized equation systems the heterogeneous FE problem will be described by a FETI-DP (Finite Element Tearing and Interconnecting - Dual Primal) discretization. The fundamental FETI-DP equation can be solved by a number of different approaches. In our approach the FETI-DP problem will be reformulated as Saddle Point system, by eliminating the primal and Lagrangian variables. For the reduced Saddle Point system, only defined by interior and dual variables, special Uzawa algorithms can be adapted for iteratively solving the FETI-DP saddle-point equation system (FETI-DP SPE). A conjugate gradient version of the Uzawa algorithm will be shown as well as some numerical tests regarding to FETI-DP discretization of small examples using the presented solution technique. Furthermore the inversion of the interior-dual Schur complement operator can be approximated using different techniques building an adequate preconditioning matrix and therewith leading to substantial gains in computing time efficiency.

Quality is one of the most important properties of a product. Providing the optimal quality can reduce costs for rework, scrap, recall or even legal actions while satisfying customers demand for reliability. The aim is to achieve ``built-in'' quality within product development process (PDP). The common approach therefore is the robust design optimization (RDO). It uses stochastic values as constraint and/or objective to obtain a robust and reliable optimal design. In classical approaches the effort required for stochastic analysis multiplies with the complexity of the optimization algorithm. The suggested approach shows that it is possible to reduce this effort enormously by using previously obtained data. Therefore the support point set of an underlying metamodel is filled iteratively during ongoing optimization in regions of interest if this is necessary. In a simple example, it will be shown that this is possible without significant loss of accuracy.

BAUHAUS ISOMETRY AND FIELDS
(2012)

While integration increases by networking, segregation strides ahead too. Most of us fixate our mind on special topics. Yet we are relying on our intuition too. We are sometimes waiting for the inflow of new ideas or valuable information that we hold in high esteem, although we are not entirely conscious of its origin. We may even say the most precious intuitions are rooting in deep subconscious, collective layers of the mind. Take as a simple example the emergence of orientation in paleolithic events and its relation to the dihedral symmetry of the compass. Consider also the extension of this algebraic matter into the operational structures of the mind on the one hand and into the algebra of geometry, Clifford algebra as we use to call it today, on the other. Culture and mind, and even the individual act of creation may be connected with transient events that are subconscious and inaccessible to cognition in principle. Other events causative for our work may be merely invisible too us, though in principle they should turn out attainable. In this case we are just ignorant of the whole creative process. Sometimes we begin to use unusual tools or turn into handicraft enthusiasts. Then our small institutes turn into workshops and factories. All this is indeed joining with the Bauhaus and its spirit. We shall go together into this, and we shall present a record of this session.

Im Bereich der Altbausanierung und der Bestandserfassung im Bauwesen ist es häufig notwendig, bestehende Pläne hinsichtlich des Bauwerkszustandes zu aktualisieren oder, wenn diese Pläne nicht (mehr) zugänglich sind, gänzlich neue Planunterlagen des Ist-Zustandes zu erstellen. Ein komfortabler Weg, diese Bauwerksdaten zu erheben, eröffnet die Technologie der Laservermessung. Der vorliegende Artikel stellt in diesem Zusammenhang Ansätze zur Teilautomatisierung der Generierung eines dreidimensionalen Computermodells eines Bauwerkes vor. Als Ergebnis wird ein Volumenmodell bereitgestellt, in dem zunächst die geometrischen und topologischen Informationen über Flächen, Kanten und Punkte im Sinne eines B-rep Modells beschrieben sind. Die Objekte dieses Volumenmodells werden mit Verfahren aus dem Bereich der künstlichen Intelligenz analysiert und in Bauteilklassen systematisch kategorisiert. Die Kenntnis der Bauteilsemantik erlaubt es somit, aus den Daten ein Bauwerks-Produktmodell abzuleiten und dieses einzelnen Fachplanern – etwa zur Erstellung eines Energiepasses – zugänglich zu machen. Der Aufsatz zeigt den erfolgreichen Einsatz virtueller neuronaler Netze im Bereich der Bestandserfassung anhand eines komplexen Beispiels.

Unconstrained models are very often found in the broad spectrum of different theories of traffic demand models. In these models there are none or only one-sided restrictions influencing the choice of the individual. However in the traffic demand different deciding dependencies of the traffic volume with regard to the specific conditions of the territory structure potentials exist. Kichhoff and Lohse introduced bi- and tri-linearly constrained models to show these dependencies. In principle, the dependencies are described as hard, elastic and open boundary sum criteria. In this article a model is formulated which gets away from these predefined boundary sum criteria and allows a free determination of minimal and maximal boundary sum criteria. The iterative solution algorithm is shown according to a FURNESS procedure at the same time. With the approach of freely selectable minimal and maximal boundary sum criteria the modeling transport planner gets the possibility to show the traffic event even better. Furthermore all common boundary sum criteria can be calculated with this model. Therewith the often necessary and sensible standard and special cases can also be modeled.

Mikroelektronik und Mikrosystemtechnik in Kombination mit Informations- und Kommunikations-technik erlauben es mittlerweile, Rechenleistung und Kommunikationsfähigkeit in kleinsten Formaten, mit geringsten Energien und zu günstigen Preisen nutzbringend in unser privates und berufliches Umfeld einzubringen. Beispiele sind Notebook-PC, PDA, Handy und das Navigationßystem im Auto. Aber auch eingebettete Elektronik in Komponenten, Geräten und Systemen ist nunmehr zur Selbstverständlichkeit geworden. Bekannte Beispiele aus der Haustechnik sind Mikroprozeßoren in Heizungs- und Alarmanlagen und aber auch in Komponenten wie Brand- und Bewegungsmelder. Wir nähern uns dem vor einigen Jahren noch als Vision bezeichneten Zustand der überall vorhandenen elektronischen Rechenleistung (engl. ubiquitous computing) bzw. des von Informationsverarbeitung durchdrungenen täglichen Umfelds (engl. pervasive computing). Werden die TGA-Komponenten genau wie die größeren Computerkomponenten (z.B. PCs, Server) über Datenschnittstellen zu räumlich verteilten Netzwerken verknüpft (z.B. Internet, Intranet) und mit einer systemübergreifenden und adäquaten Intelligenz (Software) programmiert, so können neuartige Funktionalitäten im jeweiligen Anwendungsumfeld (engl. ambient intelligence, kurz AmI, [1]) entstehen. Hier liegt bei Gebäuden und Räumen speziell eine große Chance, die bislang einer ganzheitlichen Systemkonzeption unter Einschluß von Architektur, Gebäudephysik, technischer Gebäudeausrüstung (TGA) und Gebäudeautomation (GA) im Wege stehende Gewerketrennung zu überwinden. Es entstehen für div. Anwendungszwecke systemisch integrierte >smart areas< (nach Prof. Becker, FH Biberach). Im vorliegenden Beitrag erläuterte Beispiele für AmI-Lösungen im Immobilienbereich sind Raumsysteme zur automatischen und sicheren Erkennung von Notfällen, z.B. in Pflegeheimen; sich automatisch an die Nutzung und den Nutzer bzgl. Klima und Beleuchtung adaptierende Raumsysteme im Büro- oder Hotelbereich und die elektronische Aßistenz des Bau- und Betriebsprozeßes von Gebäuden. Im Duisburger inHaus-Innovationszentrum für Intelligente Raum- und Gebäudesysteme der Fraunhofer-Gesellschaft wurden in den letzten Jahren erste Lösungen mit diesem neuartigen Ansatz konzipiert, entwickelt und erprobt. Der Beitrag beschreibt nach einer kurzen Skizzierung des Ambient-Intelligence-Ansatzes an Beispielen Möglichkeiten für den Transfer dieser neuen Technologie in den Raum- und Gebäudebereich. Es folgt eine abschließende Zusammenfaßung und eine Einschätzung der Zukunftspotenziale der Ambient Intelligence in Raum und Bau.

Räume und Gebäude sind heute wegen der enormen Funktionalität der technischen Gebäudeausrüstung (TGA) in Kombination mit der sonstigen Ausstattung und den diversen Anwendungsprozessen und Nutzergruppen ohne innovative Konzepte der integrierten Bedienung kaum noch beherrschbar bzw. optimal nutzbar. Dies gilt sowohl für Wohn- als auch für Zweckimmobilien. Die Gebäudeleittechnik (GLT) und die Gebäudeautomation (GA) können hier unter sinnvoller Integration der Möglichkeiten der Mikroelektronik, Multimedia-, Kommunikations- und Informationstechnik erheblich zu nutzbringenden Innovationen beitragen. Die Automobilindustrie hat in den letzten Jahren gezeigt, wie durch einen integralen Systemansatz und durch Einsatz von Elektronik, Kommunikations- und Informationstechnik eine sinnvolle technische Assistenz der Anwender machbar ist. Genannt sei hier das Konzept des Cockpits mit integrierter Funktionsbündelung und der Informationskonzentration am Armaturenbrett. Im Gegensatz zum Automobil ist der Bereich der technischen Gebäudeausstattung in Wohn- und Nutzimmobilien gekennzeichnet durch eine starke Fragmentierung in unterschiedlichste Gewerke unter Beteiligung vieler oft schlecht koordinierter Akteure. Durch das Duisburger inHaus-Innovationszentrum für Intelligente Raum- und Gebäudesysteme der Fraunhofer-Gesellschaft wurden in den letzten Jahren neuartige Konzepte der Systemintegration heterogener Technik auf der Basis von Middleware-Plattformen und Multimedia-Technologien und -Geräten entwickelt, getestet und in die Anwendung getragen. Einer der ersten Systemanwendungen dieses offenen Infrastrukturkonzepts ist die integrierte Systembedienung mit zum Teil völlig neuen Bedienkonzepten und einer starken Bedienungsvereinfachung auch komplexester Technikausrüstungen in Immobilien. Der Beitrag beschreibt nach einer Analyse der Ausgangslage die technologischen Grundzüge der integrierten Systembedienung. Es folgen einige Anwendungsbeispiele und eine zusammenfassende Bewertung mit einem Ausblick auf weiterführende Aktivitäten.

The topic of structural robustness is covered extensively in current literature in structural engineering. A few evaluation methods already exist. Since these methods are based on different evaluation approaches, the comparison is difficult. But all the approaches have one in common, they need a structural model which represents the structure to be evaluated. As the structural model is the basis of the robustness evaluation, there is the question if the quality of the chosen structural model is influencing the estimation of the structural robustness index. This paper shows what robustness in structural engineering means and gives an overview of existing assessment methods. One is the reliability based robustness index, which uses the reliability indices of a intact and a damaged structure. The second one is the risk based robustness index, which estimates the structural robustness by the usage of direct and indirect risk. The paper describes how these approaches for the evaluation of structural robustness works and which parameters will be used. Since both approaches needs a structural model for the estimation of the structural behavior and the probability of failure, it is necessary to think about the quality of the chosen structural model. Nevertheless, the chosen model has to represent the structure, the input factors and reflect the damages which occur. On the example of two different model qualities, it will be shown, that the model choice is really influencing the quality of the robustness index.

In distributed project organisations and collaboration there is a need for integrating unstructured self-contained text information with structured project data. We consider this a process of text integration in which various text technologies can be used to externalise text content and consolidate it into structured information or flexibly interlink it with corresponding information bases. However, the effectiveness of text technologies and the potentials of text integration greatly vary with the type of documents, the project setup and the available background knowledge. The goal of our research is to establish text technologies within collaboration environments to allow for (a) flexibly combining appropriate text and data management technologies, (b) utilising available context information and (c) the sharing of text information in accordance to the most critical integration tasks. A particular focus is on Semantic Service Environments that leverage on Web service and Semantic Web technologies and adequately support the required systems integration and parallel processing of semi-structured and structured information. The paper presents an architecture for text integration that extends Semantic Service Environments with two types of integration services. Backbone to the Information Resource Sharing and Integration Service is a shared environment ontology that consolidates information on the project context and the available model, text and general linguistic resources. It also allows for the configuration of Semantic Text Analysis and Annotation Services to analyse the text documents as well as for capturing the discovered text information and sharing it through semantic notification and retrieval engines. A particular focus of the paper is the definition of the overall integration process configuring a complementary set of analyses and information sharing components.

In this paper we evaluate 2D models for soil-water characteristic curve (SWCC), that incorporate the hysteretic nature of the relationship between volumetric water content Θ and suction Ψ. The models are based on nonlinear least squares estimation of the experimental data for sand. To estimate the dependent variable Θ the proposed models include two independent variables, suction and sensors reading position (depth d in the column test). The variable d represents not only the position where suction and water content are measured but also the initial suction distribution before each of the hydraulic loading test phases. Due to this the proposed 2D regression models acquire the advantage that they: (a) can be applied for prediction of Θ for any position along the column and (b) give the functional form for the scanning curves.

The use of virtual reality techniques in the development of educational applications brings new perspectives to the teaching of subjects related to the field of civil construction in Civil Engineering domain. In order to obtain models, which would be able to visually simulate the construction process of two types of construction work, the research turned to the techniques of geometric modelling and virtual reality. The applications developed for this purpose are concerned with the construction of a cavity wall and a bridge. These models make it possible to view the physical evolution of the work, to follow the planned construction sequence and to visualize details of the form of every component of the works. They also support the study of the type and method of operation of the equipment necessary for these construction procedures. These models have been used to distinct advantage as educational aids in first-degree courses in Civil Engineering. Normally, three-dimensional geometric models, which are used to present architectural and engineering works, show only their final form, not allowing the observation of their physical evolution. The visual simulation of the construction process needs to be able to produce changes to the geometry of the project dynamically. In the present study, two engineering construction work models were created, from which it was possible to obtain three-dimensional models corresponding to different states of their form, simulating distinct stages in their construction. Virtual reality technology was applied to the 3D models. Virtual reality capacities allow the interactive real-time viewing of 3D building models and facilitate the process of visualizing, evaluating and communicating.

The main aim of the research project in progress is to develop virtual models as tools to support decision-making in the planning of construction maintenance. The virtual models gives the capacity to allow them to transmit, visually and interactively, information related to the physical behaviour of materials, components of given infrastructures, defined as a function of the time variable. The interactive application allows decisions to be made on conception options in the definition of plans for maintenance, conservation or rehabilitation. The first virtual prototype that is now in progress concerns just lamps. It allows the examination of the physical model, visualizing, for each element modelled in 3D and linked to a database, the corresponding technical information concerned with the wear and tear aspects of the material, calculated for that period of time. In addition, the analysis of solutions for repair work or substitution and inherent cost are predicted, the results being obtained interactively and visualized in the virtual environment itself. The aim is that the virtual model should be able to be applied directly over the 3D models of new constructions, in situations of rehabilitation. The practical usage of these models is directed, then, towards supporting decision-making in the conception phase and the planning of maintenance. In further work other components will be analysed and incorporated into the virtual system.

Die effektive Kooperation aller beteiligten Fachplaner im Bauplanungsprozess ist die Voraussetzung für wirtschaftliches und qualitativ hochwertiges Bauen. Bauprojektorganisationen bestehen in der Regel aus zahlreichen unabhängigen Planungspartnern, die örtlich verteilt spezifische Planungsaufgaben bearbeiten und die Ergebnisse in Teilproduktmodellen ablegen. Da Planungsprozesse im Bauwesen stark arbeitsteilig ablaufen, sind die Teilproduktmodelle der einzelnen Fachplanungen in hohem Maße voneinander abhängig. Ziel des hier vorgestellten Ansatzes ist die Integration der Teilproduktmodelle der Gebäudeplanung in einem netzwerkbasierten Modellverbund am Beispiel der Brandschutzplanung. Im Beitrag werden die Probleme der Verteiltheit und insbesondere der semantischen Heterogenität der involvierten Teilproduktmodelle betrachtet. Der verteilte Zugriff wird mithilfe mobiler Software-Agenten realisiert. Die Agenten können sich dabei frei im netzwerkbasierten Planungsverbund bewegen und agieren als Vertreter der Fachplaner. Das Problem der semantischen Heterogenität der Teilproduktmodelle wird auf der Basis von Ontologien gelöst. Dazu werden erstens Domänenontologien entwickelt, die Objekte der realen Welt einer abgeschlossenen Domäne, hier des Brandschutzes, abbilden. Zweitens werden Applikationsontologien entwickelt, die die einzelnen proprietären Datenhaltungen (im Sinne von Teilproduktmodellen) der jeweiligen Fachplanungen repräsentieren. Beide Ontologien werden mit einem regelbasierten Ansatz verknüpft. Im vorgestellten Anwendungsfall Brandschutz dient die Domänenontologie als einheitliche Schnittstelle für den Zugriff auf die verteilten Modelle und abstrahiert dabei von deren Datenbankspezifika und proprietären Schemata. Mithilfe von mobilen Agenten und semantischen Technologien kann so eine Plattform zur Verfügung gestellt werden, die erstens die dynamische Integration von Ressourcen in den Planungsverbund erlaubt und zweitens auf deren Basis unabhängig von der Verteiltheit und Heterogenität der eingebundenen Ressourcen ingenieurgerechte Verarbeitungsmethoden realisiert werden können.

The changed global security situation in the last eight years has shown the importance of emergency management plans in public buildings. Therefore, the use of computer simulators for surveying fire safety design and evacuation process is increasing. The aim of these simulators is to have more realistic evacuation simulations. The challenge is, firstly, to realize the virtual simulation environment based on geometrical and material boundary conditions, secondly, to considerate the mutual interaction effects between different parameters and, finally, to have a realistic visualization of the simulated results. In order to carry out this task, an especial new software method on a BIM-platform has to be developed which can integrate all required simulations and will be able to have an immersive output BIM ISEE (Immersive Safety Engineering Environment). The new BIM-ISEE will integrate the Fire Dynamics Simulator (FDS) for fire and evacuation simulation in the Autodesk Revit which is a BIM-platform and will represent the simulation results in the immersive virtual environment at the institute (CES-Lab). With BIM-ISEE the fire safety engineer will be able to obtain more realistic visualizations in the immersive environment, to modify his concept more effectively, to evaluate the simulation results more accurately and to visualize the various simulation results. It can also give the rescue staff the opportunity to perform and evaluate emergency evacuation trainings.

The growing competition pressure in the building industry increases the demands on the design and construction processes in respect to economical, technical and time aspects. These demands require efficient improvements of the value-added chain, which can be realized mainly with the usage of innovative information- and communication-technologies. To support the collaboration of all participants involved in a certain building project the Workflow-Management-System “BauKom-Online” has been developed. In the focus of the system is to support the coordination of the participants and their information exchange. Such a software-method is well suited to ensure a high quality planning process. The modelling of business-processes enables a better self-comprehension of the participants work and helps to enhance the project performance. The system architecture of BauKom-Online contains two basic components: the processmodelling tool and the workflow-engine. The process-model contains of activities and states of the planning and construction processes and their relations. These connected processes compose the workflow. Such a process-model for engineering purposes has to satisfy several needs, e.g., the consideration of planning and building alternatives, dynamic changes of the model during execution of the project and the linkage to further technical objects like costs, building structure, specifications and documentmanagement. Furthermore, the scheduling of the project can be done within the process-model and can be visualized as a Gantt-diagram. ...

As a result of the pilot-project “Grundwasser-Online” the supervision and the active controlling of large monitoring- and catchment-areas are realised by a co-operative integration of all related institutions, a synchronisation-process to combine all distributed data into one central server database, and a high-level eGovernment-Service to provide evaluated information over the internet. Based on this software system the local authorities are able to supervise the groundwater-levels and to find adequate decisions, which finally result in official permissions for the usage of groundwater-reservoirs by the water supply companies.

Durch die Betrachtung des Produktions-Prozesses als zentrales Transformationselement wird die Struktur der Bauproduktion realitätsnah gefasst. Die Integration der prozessorientierten Kostendefinition setzt relevante Kostenparameter und Produktionsfaktoren so in Beziehung, dass sie im Einklang mit der realen Kostenstruktur und Kostendynamik einer Baustelle stehen. Die Beziehung zwischen Bauzeit und Kosten wird direkt erfasst und ausgewertet. Der hohen Dynamik der Bauproduktion zwischen kapazitätsbeschränkten Einsatzmitteln und Produktionsprozessen wurde durch das Poolmodell und der Simulation als Berechnungsmethode Rechnung getragen. Eine einfache Modellierung von sich zyklusartig wiederholenden Arbeitsvorgängen (Taktplanung) ist möglich. Die Taktbildung vollzieht sich bei der Simulation durch Kapazitätsbeschränkungen ohne Zutun des Benutzers. Durch eine Optimierungsmethode kann automatisiert nach der kostengünstigsten oder zeitlich schnellsten Produktionsvariante gesucht werden

MODELLING THE PLASTIC HINGE IN THE STATICALLY INDETERMINABLE REINFORCED CONCRETE BAR ELEMENTS
(2006)

The paper presents the example numerical model to calculate the reinforced concrete bar structures. Usually applied methods of structure dimensioning do not include the case of plastic hinges occurrence under the limit load of construction. The model represented by A. Borcz is based on the differential equation of deflection line of the beam and it includes the effects of rearrangement of the internal forces and reological effects. The experimental parameters obtained in earlier investigations describe effects resulting from the rise of plastic hinges in the proposed equation.

We present the way of calculation of displacement in the bent reinforced concrete bar elements where rearrangement of internal forces and plastic hinge occurred. The described solution is based on prof. Borcz’s mathematical model. It directly takes into consideration the effects connected with the occurrence of plastic hinge, such as for example a crack, by means of a differential equation of axis of the bent reinforced concrete beam. The EN Eurocode 2 makes it possible to consider the influence of plastic hinge on the values of the reinforced concrete structures. This influence can also be assumed using other analytical methods. However, the results obtained by the application of Eurocode 2 are higher from those received in testing. Just comparably big error level occurs when calculations are made by means of Borcz’s method, but in the latter case, the results depend on the assumptions made beforehand. This method makes it possible to apply the experimental results using parameters r1 i r0. When the experimental results are taken into account, one could observe the compatibility between the calculations and actual deflections of the structure.

The Laguerre polynomials appear naturally in many branches of pure and applied mathematics and mathematical physics. Debnath introduced the Laguerre transform and derived some of its properties. He also discussed the applications in study of heat conduction and to the oscillations of a very long and heavy chain with variable tension. An explicit boundedness for some class of Laguerre integral transforms will be present.

In photogrammetry and computer vision the trifocal tensor is used to describe the geometric relation between projections of points in three views. In this paper we analyze the stability and accuracy of the metric trifocal tensor for calibrated cameras. Since a minimal parameterization of the metric trifocal tensor is challenging, the additional constraints of the interior orientation are applied to the well-known projective 6-point and 7-point algorithms for three images. The experimental results show that the linear 7-point algorithm fails for some noise-free degenerated cases, whereas the minimal 6-point algorithm seems to be competitive even with realistic noise.

Several results concerning the distribution of the headway of busses in the flow behind a traffic signal are presented. In the main focus of interest is the description of analytical models, which are verified by the results of Monte-Carlo-Methods. The advantage of analytical models (verified, but not derived by simulation methods) is their flexibility with respect to possible generalizations. For instance, several random distributions of the flow incoming to the traffic signal can be compared. The attention will be directed at the question, how the primary headway H (analyzed in front of the traffic signal) is mapped to the headway H’ analyzed behind of the traffic signal and how the random distribution of H is mapped to that of H’. For the traffic flow in front of the traffic signal several models will be discussed. The first case considers the situation, that busses operate on a common lane with the individual motor car traffic and the traffic flow is saturated. In the second situation, busses operate on a separated bus lane. Moreover, a mixed situation is discussed to model as close to reality as possible.

This research focuses on the Case-based Reasoning paradigm in architectural design (CBD) and education. Initial point for further exploring this only seemingly comprehensive investigated field of research constitutes the finding that promising looking concepts exist but that they do not play a role in daily routine of designing architects or in university education. In search of reasons for this limited success a critical review of the CBR approach to architectural education and design was performed. The aim was to identify gaps in the CBD research and to discover potential fields of research within CBR research in architectural education and design to improve acceptance and practical suitability. Two major shortcomings could be identified. In the first place the way retrieval mechanisms of systems under investigation relate to the needs of architectural designers and students. At second: Successful CBD systems rely on the work of third-parties in sharing their experiences with others and filling the databases with relevant cases. Therefore two questions remain unanswered: The question of which projects become part of the database and how get existing projects not only described but evaluated. This is an essential task and prerequisite to meet the requirements of the underlying theory of CBR.

By the use of numerical methods and the rapid development of computer technology in the recent years, a large variety, complexity, refinement and capability of partial models have been achieved. This can be noticed in the evaluation of the reliability of structures, e.g. the increased use of spatial structural systems. For the different fields of civil engineering, well developed partial models already exist. Because these partial models are most often used separately, the general view is not entirely illustrated. Until now, there has been no common methodology for evaluating the efficiency of models; the trust in the prediction of a special engineering model has generally relied on the engineer’s experience. In this paper the basics of evaluation of simple models and coupled partial models of frame structures will be discussed using sustainable numerical methods. Furthermore, quality classes (levels) of design tasks will be defined based on their practical relevance. In addition, analysis methods will be systemized. After analysis of different published assessment methods, it may be noted, that the Efficiency Indicator Method (EWM) is most suitable for the observed evaluation problem. Therefore, the EWM was modified to the Model Efficiency Analysis (MEA) for the purpose of a holistic evaluation. The criteria are characterized by two groups, benefit and expenditure, and it is possible by calculating the quotient (benefit/expenditure) to make a statement about the efficiency of the observed models. Presently, the expenditure value is not a subject of investigation, and so the model efficiency is calculated only by the benefit value. This paper also contains the associated criteria catalog, different normalization methods, as well as weighting possibilities.

In the paper presented, reinforced concrete shells of revolution are analyzed in both meridional and circumferential directions. Taking into account the physical non-linearity of the material, the internal forces and the deflections of the shell as well as the strain distribution at the cross-sections are calculated. The behavior of concrete under compression is described by linear and non-linear stress-strain relations. The description of the behavior of concrete under tension must account for tension stiffening effects. A tri-linear function is used to formulate the material law of reinforcement. The problem cannot be solved analytically due to the physical non-linearity. Thus a numerical solution is formulated by means of the LAGRANGE Principle of the minimum of the total potential energy. The kinematically admissible field of deformation is defined by the displacements u in the meridional and w in the radial direction. These displacements must satisfy the equations of compatibility and the kinematical boundary conditions of the shell. The strains are linearly distributed across the wall thickness. The strain energy depends on the specific of the material behavior. Using integral formulations of the material law [1], the strain energy of each part of the cross-section is defined as a function of the strains at the boundaries of the cross-sections. The shell is discretised in the meridional direction. Various methods of numerical differentiation and numerical integration are applied in order to determine the deformations and the strain energy. The unknown displacements u and w are calculated by a non-restricted extremum problem based on the minimum of the total potential energy. From mathematical point of view, the objective function is a convex function, thus the minimum can be determined without difficulty. The advantage of this formulation is that unlike non-linear methods with path-following algorithms the calculation does not have to account for changing stiffness and load increments. All iterations necessary to find the solution are integrated into the “Solver”. The model presented provides many ways of investigating the influence of various material parameters on the stresses and deformations of the entire shell structure.

The presented method for an physically non-linear analysis of stresses and deformations of composite cross-sections and members based on energy principles and their transformation to non-linear optimisation problems. From the LAGRANGE principle of minimum of total potential energy a kinematic formulation of the mechanical problem can be developed, which has the general advantage that pre-deformations excited by shrinkage, temperature, residual deformations after unloading et al., can be considered directly. Thus the non-linear analysis of composite cross-sections with layers of different mechanical properties and different preloading becomes possible and cracks in concrete, stiffness degradation and other specifics of the material behaviour can be taken into account without cardinal modification of the mathematical model. The impact of local defects on the bearing capacity of an entire element can also be analysed in this principle way. Standard computational systems for mathematical optimisation or general programs for spreadsheet analysis enable an uncomplicated implementation of the developed models and an effective non-linear analysis for composite cross-sections and elements.

A new approach to the non-linear analysis of cross-sections loaded by normal forces and bending moments is presented in the paper. The mechanical model is based on the LAGRANGE principle of minimum of total potential energy. Deformations, stresses and limit load parameters are obtained by solving a non-linear optimisation problem. The mathematical model is independent of the specifics of material. In addition to the stress strain relation and the specific strain energy W(ε) two further functions F(ε) and Φ(ε) are introduced to describe the material behaviour. Thus cracks in concrete, non-linearity of material etc. can be taken into account without basic modification of the numerical algorithm. For polygonal cross-sections the GAUSS' integral theorem is used. Numerical solutions of the non-linear optimisation problems can be found by application of standard software. Thus the analysis of reinforced concrete cross-sections or more general composite cross-sections with non-linear behaviour of material is as simple as in the case of linear elasticity. The application of the method is demonstrated for polygonal cross-sections. Pre-stresses or pre-strains can easily be included in the mathematical model.

An energy method based on the LAGRANGE Principle of the minimum of total potential en-ergy is presented to calculate the stresses and strains of composite cross-sections. The stress-strain relation of each partition of the cross-section can be an arbitrary piecewise continuous function. The strain energy is transformed into a line integral by GAUSS’s integral theorem. The total strain of each partition of the cross-section is split into load-dependent strain and pre-strain. Pre-strains have to be taken into account when the cross-section is pre-stressed, retrofit-ted or influenced by shrinkage, temperature etc. The unconstrained minimum problem can be solved for each load combination using standard software. The application of the method presented in the paper is demonstrated by means of examples.

As numerical techniques for solving PDE or integral equations become more sophisticated, treatments of the generation of the geometric inputs should also follow that numerical advancement. This document describes the preparation of CAD data so that they can later be applied to hierarchical BEM or FEM solvers. For the BEM case, the geometric data are described by surfaces which we want to decompose into several curved foursided patches. We show the treatment of untrimmed and trimmed surfaces. In particular, we provide prevention of smooth corners which are bad for diffeomorphism. Additionally, we consider the problem of characterizing whether a Coons map is a diffeomorphism from the unit square onto a planar domain delineated by four given curves. We aim primarily at having not only theoretically correct conditions but also practically efficient methods. As for FEM geometric preparation, we need to decompose a 3D solid into a set of curved tetrahedra. First, we describe some method of decomposition without adding too many Steiner points (additional points not belonging to the initial boundary nodes of the boundary surface). Then, we provide a methodology for efficiently checking whether a tetrahedral transfinite interpolation is regular. That is done by a combination of degree reduction technique and subdivision. Along with the method description, we report also on some interesting practical results from real CAD data.

Major problems of applying selective sensitivity to system identification are requirement of precise knowledge about the system parameters and realization of the required system of forces. This work presents a procedure which is able to deriving selectively sensitive excitation by iterative experiments. The first step is to determine the selectively sensitive displacement and selectively sensitive force patterns. These values are obtained by introducing the prior information of system parameters into an optimization which minimizes the sensitivities of the structure response with respect to the unselected parameters while keeping the sensitivities with respect to the selected parameters as a constant. In a second step the force pattern is used to derive dynamic loads on the tested structure and measurements are carried out. An automatic control ensures the required excitation forces. In a third step, measured outputs are employed to update the prior information. The strategy is to minimize the difference between a predicted displacement response, formulated as function of the unknown parameters and the measured displacements, and the selectively sensitive displacement calculated in the first step. With the updated values of the parameters a re-analysis of selective sensitivity is performed and the experiment is repeated until the displacement response of the model and the actual structure are conformed. As an illustration a simply supported beam made of steel, vibrated by harmonic excitation is investigated, thereby demonstrating that the adaptive excitation can be obtained efficiently.

The paper presents a linear static analysis on continuous orthotropic thin-walled shell structures simply supported at the transverse ends with a random deformable contour of the cross section. The external loads can be random as well. The class of this structures involves most of the bridges, scaffold bridges, some roof structures etc. A numerical example of steel continuous structures on five spans with an open contour of the cross-section has been solved. The examination of the structure has used the following two computation models: a prismatic structure consisting of isotropic strips, a plates and ribs, with considering their real interaction, and a smooth orthotropic plate equivalent to the structure in the first model. The displacements and forces of the structure characterizing its stressed and deformed condition have been determined. The results obtained from the two solutions have been analyzed. The study on the structure is made with the force method in combination with the analytical finite strip method (AFSM) in displacements. The basic system is obtained by separating the superstructure from the understructure at the places of intermediate supports and consists of two parts. The first part is a single span thin-walled prismatic shell structure; the second part presents supports (columns, space frames etc.). The connection between the superstructure and intermediate supports is made under random supporting conditions. The forces at the supporting points in the direction of the connections removed are assumed to be the basic unknowns of the force method. The solution of the superstructure has been accomplished by the AFSM in displacements. The structure is divided in only one (transverse) direction into a finite number of plain strips connected to each other in longitudinal linear nodes. The three displacements of the points on the node lines and the rotation around those lines have been assumed to be the basic unknown in each node. The boundary conditions of each strip of the basic system correspond to the simply support along the transverse ends and the restraint along the longitudinal ones. The particular strip of the basic system has been solved by the method of the single trigonometric series. The method is reduced to solving a discrete structure in displacements and restoring its continuity at the places of the sections made in respect to both the displacements and forces. The two parts of the basic system have been solved in sequence under the action of single values of each of the basic unknowns and with the external load. The solution of the support part is accomplished using software for analyzing structures by the FEM. The basic unknown forces have been determined from system of canonic equations, the conditions of the deformations continuity on the places of the removed connections under superstructure and intermediate supports. The final displacements and forces at a random point of a continuous superstructure have been determined using the principle of superposition. The computations have been carried by software developed with Visual Fortran version 5.0 for PC.

Let the information of a civil engineering application be decomposed into objects of a given set of classes. Then the set of objects forms the data base of the application. The objects contain attributes and methods. Properties of the objects are stored in the attributes. Algorithms which the objects perform are implemented in the methods of the objects. If objects are modified by a user, the consistency of data in the base is destroyed. The data base must be modified in an update to restore its consistency. The sequence of the update operations is not arbitrary, but is governed by dependence between the objects. The situation can be described mathematically with graph theory. The available algorithms for the determination of the update sequence are not suitable when the data base is large. A new update algorithm for large data bases has been developed and is presented in this paper.

The paper is dedicated to decidability exploration of market segmentation problem with the help of linear convolution algorithms. Mathematical formulation of this problem represents interval task of bipartite graph cover by stars. Vertices of the first partition correspond to types of commodities, vertices of the second – to customers groups. Appropriate method is offered for interval problem reduction to two-criterion task that has one implemented linear convolution algorithm. Unsolvability with the help of linear convolution algorithm of multicriterion, and consequently interval, market segmentation problem is proved.

Integrated Engineering Workflow focused on the Structural Engineering in the Industrial Environment
(2004)

The engineering and construction industry has been slow to exploit the full potential of information technology. The industry is highly fragmented, price sensitive, risk-adverse, and profit margins are small. Each project is unique with a small amount of technological innovation opportunities to capitalise on from one project to the next. Technological innovations that have been taking place are just simulating the old traditional paper workflow. Engineering information in digital form is being conveyed using traditional paper representations, which have to be interpreted by humans before the information can be used in other applications, thereby creating ‘islands of information’. It can be seen that poorly implemented IT strategies are duplicating paperwork, rather than reducing or eliminating it (Crowley et al., 2000). This paper will introduce the Integrated Engineering Workflow (IEW) concept to re-organise a structural discipline working on multi-disciplinary projects so as to maximise the advantages offered by new information technology.

Water resources development and management is a complex problem. It includes the design and operation of single system components, often as part of larger interrelated systems and usually on the basis of river basins. While several decades ago the dominant objective was the maximization of economic benefit, other objectives have evolved as part of the sustainable development envisaged. Today, planning and operation of larger water resources systems is practically impossible without adequate computer tools, normally being one or several models, increasingly combined with data bank management systems and multi criteria assessment procedures in decision support systems. The use of models in civil engineering already has a long history when structural engineering is considered. These design support models, however, must rather be seen as expert systems made to support the engineer with his daily work. They often have no direct link to stakeholders and the decision makers community. The scale of investigation is often much larger in water resources engineering than in structural engineering which is related to different stakeholders and decision making procedures. Still, several similarities are obvious which can be summarized as the search for a compromise solution on a complex, i.e. multiobjective and interdisciplinary decision problem. While in structural engineering e.g. aestetics, stability and energy consumption might be important evaluation criteria in addition to construction and maintenance cost other or additional criteria have to be considered in water resources planning such as political, environmental and social criteria. In this respect civil engineers tend to overemphasize technical criteria. For the future the existing expert systems should be embedded into an improved decision support shell, keeping in mind that decision makers are hardly interested in numerical modelling results. The paper will introduce into the problem and demonstrate the state of the art by means of an example.

Nonlinear analyses are characterised by approximations of the fundamental equations in different quality. Starting with a general description of nonlinear finite element formulation the fundamental equations are derived for plane truss elements. Special emphasis is placed on the determination of internal and external system energy as well as influence of different quality approaches for the displacement-strain relationship on solution quality. To simplify the solution procedure the nonlinear function describing the kinematics is expanded into a Taylor series and truncated after the n-th series term. The different kinematics influence speed of convergence as well as exactness of solution. On a simple truss structure this influence is shown. To assess the quality of different formulations concerning the nonlinear kinematic equation three approaches are discussed. First the overall internal and external energy is compared for different kinematical models. In a second step the energy content related to single terms describing displacement-strain relationship is investigated and used for quality control following two different paths. Based on single ε-terms an adaptive scheme is used to change the kinematical model depending on increasing nonlinearity of the structure. The solution quality has turned out satisfactory compared to the exact result. More detailed investigations are necessary to find criteria for the threshold values for the iterative process as well as for decision on number and step size of incremental load steps.

This study contributes to the identification of coupled THM constitutive model parameters via back analysis against information-rich experiments. A sampling based back analysis approach is proposed comprising both the model parameter identification and the assessment of the reliability of identified model parameters. The results obtained in the context of buffer elements indicate that sensitive parameter estimates generally obey the normal distribution. According to the sensitivity of the parameters and the probability distribution of the samples we can provide confidence intervals for the estimated parameters and thus allow a qualitative estimation on the identified parameters which are in future work used as inputs for prognosis computations of buffer elements. These elements play e.g. an important role in the design of nuclear waste repositories.

A four-node quadrilateral shell element with smoothed membrane-bending based on Mindlin-Reissner theory is proposed. The element is a combination of a plate bending and membrane element. It is based on mixed interpolation where the bending and membrane stiffness matrices are calculated on the boundaries of the smoothing cells while the shear terms are approximated by independent interpolation functions in natural coordinates. The proposed element is robust, computationally inexpensive and free of locking. Since the integration is done on the element boundaries for the bending and membrane terms, the element is more accurate than the MITC4 element for distorted meshes. This will be demonstrated for several numerical examples.

Sand-bentonite mixtures are well recognized as buffer and sealing material in nuclear waste repository constructions. The behaviour of compacted sand-bentonite mixture needs to be well understood in order to guarantee the safety and the efficiency of the barrier construction. This paper presents numerical simulations of swelling test and coupled thermo-hydro-mechanical (THM) test on compacted sand-bentonite mixture in order to reveal the influence of the temperature and hydraulic gradients on the distribution of temperature, mechanical stress and water content in such materials. Sensitivity analysis is carried out to identify the parameters which influence the most the response of the numerical model. Results of back analysis of the model parameters are reported and critically assessed.

Monogenic functions play a role in quaternion analysis similarly to that of holomorphic functions in complex analysis. A holomorphic function with nonvanishing complex derivative is a conformal mapping. It is well-known that in Rn+1, n ≥ 2 the set of conformal mappings is restricted to the set of Möbius transformations only and that the Möbius transformations are not monogenic. The paper deals with a locally geometric mapping property of a subset of monogenic functions with nonvanishing hypercomplex derivatives (named M-conformal mappings). It is proved that M-conformal mappings orthogonal to all monogenic constants admit a certain change of solid angles and vice versa, that change can characterize such mappings. In addition, we determine planes in which those mappings behave like conformal mappings in the complex plane.

A numerical analysis of the mode of deformation of the main load-bearing components of a typical frame sloping shaft headgear was performed. The analysis was done by a design model consisting of plane and solid finite elements, which were modeled in the program «LIRA». Due to the numerical results, the regularities of local stress distribution under a guide pulley bearing were revealed and parameters of a plane stress under both emergency and normal working loads were determined. In the numerical simulation, the guidelines to improve the construction of the joints of guide pulleys resting on sub-pulley frame-type structures were established. Overall, the results obtained are the basis for improving the engineering procedures of designing steel structures of shaft sloping headgear.