Refine
Document Type
- Conference Proceeding (296) (remove)
Institute
- Professur Informatik im Bauwesen (129)
- In Zusammenarbeit mit der Bauhaus-Universität Weimar (98)
- Professur Theorie und Geschichte der modernen Architektur (21)
- Graduiertenkolleg 1462 (18)
- Institut für Strukturmechanik (ISM) (10)
- Professur Angewandte Mathematik (7)
- Bauhaus-Institut für Geschichte und Theorie der Architektur und Planung (3)
- Institut für Konstruktiven Ingenieurbau (IKI) (2)
- Junior-Professur Computational Architecture (2)
- Professur Stochastik und Optimierung (2)
Keywords
- Computerunterstütztes Verfahren (149)
- Architektur <Informatik> (99)
- Angewandte Informatik (83)
- Angewandte Mathematik (83)
- CAD (67)
- Computer Science Models in Engineering; Multiscale and Multiphysical Models; Scientific Computing (41)
- Verteiltes System (25)
- Building Information Modeling (21)
- Data, information and knowledge modeling in civil engineering; Function theoretic methods and PDE in engineering sciences; Mathematical methods for (robotics and) computer vision; Numerical modeling in engineering; Optimization in engineering applications (21)
- Modellierung (20)
Practical examples show that the improvement in cost flow and total amount of money spend in construction and further use may be cut significantly. The calculation is based on spreadsheets calculation, very easy to develop on most PC´s now a days. Construction works, are a field where the evaluation of Cash Flow can be and should be applied. Decisions about cash flow in construction are decisions with long-term impact and long-term memory. Mistakes from the distant past have a massive impact on situations in the present and into the far economic future of economic activities. Two approaches exist. The Just-in-Time (JIT) approach and life cycle costs (LCC) approach. The calculation example shows the dynamic results for the production speed in opposition to stable flow of production in duration of activities. More sophisticated rescheduling in optimal solution might bring in return extra profit. In the technologies and organizational processes for industrial buildings, railways and road reconstruction, public utilities and housing developments there are assembly procedures that are very appropriate for the given purpose, complicated research-, development-, innovation-projects are all very good aspects of these kinds of applications. The investors of large investments and all public invested money may be spent more efficiently if an optimisation speed-strategy can be calculated.
The promise of lower costs for sensors that can be used for construction inspection means that inspectors will continue to have new choices to consider in creating inspection plans. However, these emerging inspection methods can require different activities, resources, and decisions such that it can be difficult to compare the emerging methods with other methods that satisfy the same inspection needs. Furthermore, the context in which inspection is performed can significantly influence how well certain inspection methods are suited for a given set of goals for inspection. Context information, such as weather, security, and the regulatory environment, can be used to understand what information about a component should be collected and how an inspection should be performed. The research described in this paper is aimed at developing an approach for comparing and selecting inspection plans. This approach consists of (1) refinement of given goals for inspection, if necessary, in order to address any additional information needs due to a given context and in order to reach a level of detail that can be addressed by an inspection activity; (2) development of constraints to describe how an inspection should be achieved; (3) matching of goals to available inspection methods, and generation of activities and resource plans in order to address the goals; and (4) selection of an inspection plan from among the possible plans that have been identified. The authors illustrate this approach with observations made at a local construction site.
This paper is a report of Radio Frequency Identification (RFID) technology and its potential applications in the commercial construction industry. RFID technology offers wireless communication between RFID tags and readers with non line-of-sight readability. These fundamental properties eliminate manual data entry and introduce the potential for automated processes to increase project productivity, construction safety, and project cost efficiency. Construction contractors, owners, and material suppliers that believe technology can further develop methods and processes in construction should feel obligated to participate in RFID studies for the advancement of the construction industry as a whole.
We describe the database requirements of SEED (Software Environment to Support the Early Phases in Building Design). The requirements are typical for a database that intends to support a heterogeneous design support environment consisting of independent software modules with diverse internal design models, requirements not met by any commercial database system. The design and implementation of this database is an integral part of the overall software engineering effort. We describe the SEED approach that integrates external and in-house software based on a shared information model specified in the modeling language SPROUT, which allows for the specification of domains, and classes, relationship types and their behavior, and multiple classifications. The SPROUT run-time system organizes and coordinates the communication between the software modules and the database
In this study we introduce a concept of discrete Laplacian on the plane lattice and consider its iteration dynamical system. At first we discuss some basic properties on the dynamical system to be proved. Next making their computer simulations, we show that we can realize the following phenomena quite well:(1) The crystal of waters (2) The designs of carpets, embroideries (3) The time change of the numbers of families of extinct animals, and (4) The echo systems of life things. Hence we may expect that we can understand the evolutions and self organizations by use of the dynamical systems. Here we want to make a stress on the following fact: Although several well known chaotic dynamical systems can describe chaotic phenomena, they have difficulties in the descriptions of the evolutions and self organizations.
The worldwide growth of communication networks and associated technologies provide the basic infrastructure for new ways of executing the engineering process. Collaboration amongst team members seperated in time and location is of particular importance. Two broad themes can be recognized in research pertaining to distributed collaboration. One theme focusses on the technical and technological aspects of distributed work, while the other emphasises human aspects thereof. The case of finite element structural analysis in a distributed collaboratory is examined in this paper. An approach is taken which has its roots in human aspects of the structural analysis task. Based on experience of how structural engineers currently approach and execute this task while utilising standard software designed for use on local workstations only, criteria are stated for a software architechture that could support collaborative structural analysis. Aspects of a pilot application and the results of qualitative performance measurements are discussed.
CRITICAL STRESS ASSESSMENT IN ANGLE TO GUSSET PLATE BOLTED CONNECTION BY SIMPLIFIED FEM MODELLING
(2010)
Simplified modelling of friction grip bolted connections of steel member – to – gusset plate is often applied in engineering practise. The paper deals with the simplification of pre-tensioned bolt model and simplification of load transfer within connection. Influence on normal strain (and thus stress) distribution at critical cross-section is investigated. Laboratory testing of single-angle or double-angle members – to – gusset plates bolted connections were taken as basis for numerical analysis. FE models were created using 1D and 2D elements. Angles and gusset plates were modelled with shell elements. Two methods of modelling of friction grip bolting were considered: bolt-regarding approach with 1D element systems modelling bolts and two variants of bolt-disregarding approach with special constraints over some part of member and gusset plate surfaces in contact: a) constraints over whole area of contact, b) constraints over the area around each bolt shank (“partially tied”). Modelling of friction grip bolted connections using simplified bolt modelling may be effective, especially in the case of analysis concerning elastic range only. In such a case disregarding bolts and replacing them with “partially tied” modelling seems to be more attractive. It is less time-consuming and provides results of similar accuracy in comparison to analysis utilizing simplified bolt modelling.
In this contribution, the design of an analysis environment is presented, that supports an analyst to come to a decision within a gradual collaborative planning process. An analyst represents a project manager, planner or any other person, involved in the planning process. Today, planning processes are managed by several geographically distributed planners and project managers. Thus, complexity of such a process rises even more. Prediction of consequences of many planning decisions is not possible, in particular since assessment of a planning advance is not trivial. There have to be considered several viewpoints, that depend on individual perceptions. In the following, methods are presented to realize planning decision support.
The uncertainty existing in the construction industry is bigger than in other industries. Consequently, most construction projects do not go totally as planned. The project management plan needs therefore to be adapted repeatedly within the project lifecycle to suit the actual project conditions. Generally, the risks of change in the project management plan are difficult to be identified in advance, especially if these risks are caused by unexpected events such as human errors or changes in the client preferences. The knowledge acquired from different resources is essential to identify the probable deviations as well as to find proper solutions to the faced change risks. Hence, it is necessary to have a knowledge base that contains known solutions for the common exceptional cases that may cause changes in each construction domain. The ongoing research work presented in this paper uses the process modeling technique of Event-driven Process Chains to describe different patterns of structure changes in the schedule networks. This results in several so called “change templates”. Under each template different types of change risk/ response pairs can be categorized and stored in a knowledge base. This knowledge base is described as an ontology model populated with reference construction process data. The implementation of the developed approach can be seen as an iterative scheduling cycle that will be repeated within the project lifecycle as new change risks surface. This can help to check the availability of ready solutions in the knowledge base for the situation at hand. Moreover, if the solution is adopted, CPSP, “Change Project Schedule Plan „a prototype developed for the purpose of this research work, will be used to make the needed structure changes of the schedule network automatically based on the change template. What-If scenarios can be implemented using the CPSP prototype in the planning phase to study the effect of specific situations without endangering the success of the project objectives. Hence, better designed and more maintainable project schedules can be achieved.
This paper deals with the modelling and the analysis of masonry vaults. Numerical FEM analyses are performed using LUSAS code. Two vault typologies are analysed (barrel and cross-ribbed vaults) parametrically varying geometrical proportions and constraints. The proposed model and the developed numerical procedure are implemented in a computer analysis. Numerical applications are developed to assess the model effectiveness and the efficiency of the numerical procedure. The main object of the present paper is the development of a computational procedure which allows to define 3D structural behaviour of masonry vaults. For each investigated example, the homogenized limit analysis approach has been employed to predict ultimate load and failure mechanisms. Finally, both a mesh dependence study and a sensitivity analysis are reported. Sensitivity analysis is conducted varying in a wide range mortar tensile strength and mortar friction angle with the aim of investigating the influence of the mechanical properties of joints on collapse load and failure mechanisms. The proposed computer model is validated by a comparison with experimental results available in the literature.
A lot of real-life problems lead frequently to the solution of a complicated (large scale, multicriteria, unstable, nonsmooth etc.) nonlinear optimization problem. In order to cope with large scale problems and to develop many optimum plans a hiearchical approach to problem solving may be useful. The idea of hierarchical decision making is to reduce the overall complex problem into smaller and simpler approximate problems (subproblems) which may thereupon treated independently. One way to break a problem into smaller subproblems is the use of decomposition-coordination schemes. For finding proper values for coordination parameters in convex programming some rapidly convergent iterative methods are developed, their convergence properties and computational aspects are examined. Problems of their global implementation and polyalgorithmic approach are discussed as well.
This paper concerns schedule synchronization problems in public transit networks. In particular, it consists of three main parts. In the first the subject area is introduced, the terms are defined and framework for optimal synchronization in the form of problem representation and formulation is proposed. The second part is devoted to transfer synchronization problem when passengers changing transit lines at transfer points. The intergrated Tabu Search and Genetic solution method is developed with respect to this specific problem. The third part deals with headways harmonization problem i.e. synchronization of different transit lines schedules on a common segments of routes. For the solution of this problem a new bilevel optimization method is proposed with zones harmonization at the bottom level and co-ordination of zones, by time buffers assigned to timing points, at the upper level. Finally, the synchronization problems are numerically illustrated by real-life examples of the public transport lines in Cracow.
BAUHAUS ISOMETRY AND FIELDS
(2012)
While integration increases by networking, segregation strides ahead too. Most of us fixate our mind on special topics. Yet we are relying on our intuition too. We are sometimes waiting for the inflow of new ideas or valuable information that we hold in high esteem, although we are not entirely conscious of its origin. We may even say the most precious intuitions are rooting in deep subconscious, collective layers of the mind. Take as a simple example the emergence of orientation in paleolithic events and its relation to the dihedral symmetry of the compass. Consider also the extension of this algebraic matter into the operational structures of the mind on the one hand and into the algebra of geometry, Clifford algebra as we use to call it today, on the other. Culture and mind, and even the individual act of creation may be connected with transient events that are subconscious and inaccessible to cognition in principle. Other events causative for our work may be merely invisible too us, though in principle they should turn out attainable. In this case we are just ignorant of the whole creative process. Sometimes we begin to use unusual tools or turn into handicraft enthusiasts. Then our small institutes turn into workshops and factories. All this is indeed joining with the Bauhaus and its spirit. We shall go together into this, and we shall present a record of this session.
Most of the existing seismic resistant design codes are based on the response spectrum theory. The influence of inelastic deformations can be evaluated by considering inelastic type of resisting force and then the inelastic spectrum is considerably different from the elastic one. Also, the influence of stiffness degradation and strength deterioration can be accounted for by including more precise models from material point of view. In some recent papers the corresponding changes in response spectra due to the P- Ä effect are discussed. The experience accumulated from the recent earthquakes indicates that structural pounding may considerably influence the response of structures and should be taken into account in design procedures. The most convenient way to do that is to predict the influence of the pounding on the response spectra for accelerations, velocities and displacements. Generally speaking the contact problems such as pounding are characterized by large extent of nonlinearity and slow convergence of the computational procedures. Thus obtaining spectra where the contact problem is accounted for seems very attractive from engineering point of view because could easy be implemented into the design procedures. However it is worth nothing that there is not rigorous mathematical proof that the original system can be decomposed into single equations related to single degree of freedom systems. It is the porpose of the paper to study the influence of the pounding on the response spectra and to evaluate the amplification due to the impact. For this purpose two adjacent SDOF systems are considered that are able to interact during the vibration process. This problem is solved versus the elastic stiffness ratio, which appears to be very important for such assemblage. The contact between masses is numerically simulated using opening gap elements as links. Comparisons between calculated response spectra and linear response spectra are made in order to derive analytical relationships to simply obtain the contribution of pounding. The results are graphically illustrated in response spectra format and the influence of the stiffness ratio is clarified.
The cost of keeping large area urban computer aided architectural design (CAAD) models up to date justifies wider use and access. This paper reviews the potential for collaborative groupwork creation and maintenance of such models and suggests an approach to data entry, data management and generation of appropriate levels of detail models from a Geographic Information System (GIS). Staff at the University of the West of England (UWE) modelled a large area of Bristol to demonstrate millennium landmark proposals. It became swiftly apparent that continued amendment of the model to keep it an accurate reflection of changes on the ground was a major data management problem. Piecing in new CAAD models received from Architectural Practices to visualise them in context as part of the planning negotiation process has often taken staff several days of work for each instance. The model is so complex and proprietary that Bristol City operates a specialist visualisation bureau service. UWE later modelled the environs of the Tower of London to support bids for funding and to provide the context for judging the visual impact of iterative design development. Further research continued to develop more effective approaches to. Data conversion and amalgamation from all the diverse sources was the major impediment to effective group working to create the models. It became apparent that a GIS would assist retrieving all the appropriate data that described the part of the model under creation. It was possible to predict that management of many historic part models stepping back through time, allowing for different expert interpretations to co-exist would be in itself a major task requiring a spatial database/GIS. UWE started afresh from the original source data, to explore the collaborative use of GIS and Virtual Reality Modelling Language (VRML) to integrate models and interventions from various sources and to generate an overall navigable interactive whole. Current exploration of the combination of event driven behaviours and Structured Query Language is seeking to define how appropriately to modify objects in the VRML model on demand. This is beginning to realise the potential for use of this process for: asynchronous group modelling on the lines of a collaborative virtual design studio; historic building maintenance management; visitor management; interpretation of historic sites to visitors and public planning information.
Nodal integration of finite elements has been investigated recently. Compared with full integration it shows better convergence when applied to incompressible media, allows easier remeshing and highly reduces the number of material evaluation points thus improving efficiency. Furthermore, understanding it may help to create new integration schemes in meshless methods as well. The new integration technique requires a nodally averaged deformation gradient. For the tetrahedral element it is possible to formulate a nodal strain which passes the patch test. On the downside, it introduces non-physical low energy modes. Most of these "spurious modes" are local deformation maps of neighbouring elements. Present stabilization schemes rely on adding a stabilizing potential to the strain energy. The stabilization is discussed within this article. Its drawbacks are easily identified within numerical experiments: Nonlinear material laws are not well represented. Plastic strains may often be underestimated. Geometrically nonlinear stabilization greatly reduces computational efficiency. The article reinterpretes nodal integration in terms of imposing a nonconforming C0-continuous strain field on the structure. By doing so, the origins of the spurious modes are discussed and two methods are presented that solve this problem. First, a geometric constraint is formulated and solved using a mixed formulation of Hu-Washizu type. This assumption leads to a consistent representation of the strain energy while eliminating spurious modes. The solution is exact, but only of theoretical interest since it produces global support. Second, an integration scheme is presented that approximates the stabilization criterion. The latter leads to a highly efficient scheme. It can even be extended to other finite element types such as hexahedrals. Numerical efficiency, convergence behaviour and stability of the new method is validated using linear tetrahedral and hexahedral elements.
In earlier research, generalized multidimensional Hilbert transforms have been constructed in m-dimensional Euclidean space, in the framework of Clifford analysis. Clifford analysis, centred around the notion of monogenic functions, may be regarded as a direct and elegant generalization to higher dimension of the theory of the holomorphic functions in the complex plane. The considered Hilbert transforms, usually obtained as a part of the boundary value of an associated Cauchy transform in m+1 dimensions, might be characterized as isotropic, since the metric in the underlying space is the standard Euclidean one. In this paper we adopt the idea of a so-called anisotropic Clifford setting, which leads to the introduction of a metric dependent m-dimensional Hilbert transform, showing, at least formally, the same properties as the isotropic one. The Hilbert transform being an important tool in signal analysis, this metric dependent setting has the advantage of allowing the adjustment of the co-ordinate system to possible preferential directions in the signals to be analyzed. A striking result to be mentioned is that the associated anisotropic (m+1)-dimensional Cauchy transform is no longer uniquely determined, but may stem from a diversity of (m+1)-dimensional "mother" metrics.
The evident advances of the computational power of the digital computers enable the modeling of the total system of structures. Such modeling demands compatible representations of the couplings of different structural subsystems. Therefore, models of dynamic interaction between the vehicle and the bridge and models of a bridge bearing, a coupling element between the bridge's superstructure and substructure, are of interest and discussed within this paper. The vehicle-bridge interaction may be described as a function connecting two sets of behavior. In this case, the coupling is embodied by mutual parameters that affect both systems, such as the frequency content of the bridge and the vehicle. Whereas the bridge bearings are elements used specifically to couple, in such elements the deformation and the transferred loads are used in characterizing the coupling The nature of these couplings and their influence on the bridge response is different. However, the need to assess the amount of dynamic response transferred by or within these couplings is a common argument.
The development of a consistent material model for textile reinforced concrete requires the formulation and calibration of several sub-models on different resolution scales. Each of these models represents the material structure at the corresponding scale. While the models at the micro-level are able to capture the fundamental failure and damage mechanisms of the material components (e.g. filament rupture and debonding from the matrix) their computational costs limit their application to the small size representative unit cells of the material structure. On the other hand, the macro-level models provide a sufficient performance at the expense of limited range of applicability. Due to the complex structuring of the textile reinforced concrete at several levels (filament - yarn - textile - matrix) it is a non-trivial task to develop a multiscale model from scratch. It is rather more effective to develop a set of conceptually related sub-models for each structural level covering the selected phenomena of the material behavior. The homogenized effective material properties obtained at the lower level may be verified and validated using experiments and models at the higher level(s). In this paper the development of a consistent material model for textile reinforced concrete is presented. Load carrying and failure mechanisms at the micro, meso and macro scales are described and models with the focus on the specified scales are introduced. The models currently being developed in the framework of the collaborative research center are classified and evaluated with respect to the failure mechanisms being captured. The micromechanical modeling of the yarn and bonding behavior is discussed in detail and the correspondence with the experiments focused on the selected failure and interaction mechanisms is shown. The example of modeling the bond layer demonstrates the application of the presented strategy.
Humans are able to think, to feel, and to sense. We are also able to compute but not very well. In contrast, computers are giants in computing. Yet, they can not do anything else besides computing. Appropriate combinations of the different gifts and strengths of human and computer may result in impressive performances. In the 3-Hirn approach one human and two computers are involved. On the computers different programs are running. The human starts the machines and inspects the solutions they propose. He compares these candidate solutions and finally decides for one of the alternatives. So, the human makes the final choice from a small number of computer proposals. In performance-oriented chess, 3-Hirn combinations consisting of an amateur player and commer-cial software have reached world class level. 3-Hirn is a Decision Support System with Multiple Choice Structure. Such Multiple Choice Systems will be exhibited and discussed.
This work presents a concept of interactive machine learning in a human design process. An urban design problem is viewed as a multiple-criteria optimization problem. The outlined feature of an urban design problem is the dependence of a design goal on a context of the problem. We model the design goal as a randomized fitness measure that depends on the context. In terms of multiple-criteria decision analysis (MCDA), the defined measure corresponds to a subjective expected utility of a user. In the first stage of the proposed approach we let the algorithm explore a design space using clustering techniques. The second stage is an interactive design loop; the user makes a proposal, then the program optimizes it, gets the user’s feedback and returns back the control over the application interface.
This article discusses how the failures of construction projects could be prevented. The focus is on using IT tools in communication. The article is based on the findings of the PROLABproject. The project seeks to find solutions for how the information can effectively be used in project management, specially in construction projects, what kind of procedures help the management of the knowledge and how the obstacles for efficient ways of administrating the information can be removed. These obstacles can be either related to use of new technology or to organization culture. The paper discusses e-mail as a medium of communication with the media richness framework (Fulk & Boyd 1991) based on the interviews done with the participants of construction projects. We found out that the emails were used in both rich way and not so rich way. Obstacles of rich use of IT communication were partly technical and partly psychological. Technology acceptance model (TAM) is used to study the actual use of emails in communication. Obstacles of the use of emails were mostly individual differences like the skills to use the system or external variables like they did not have proper software to open the attachments. Also people talked about information overload. Emails often push too much information. Keywords: IT tools, Communication, Knowledge management, Construction
The analysis of the response of complex structural systems requires the description of the material constitutive relations by means of an appropriate material model. The level of abstraction of such model may strongly affect the quality of the prognosis of the whole structure. In context to this fact, it is necessary to describe the material in a convenient sense as exact but as simple as possible. All material phenomena of crystalline materials e.g. steel, affecting the behavior of the structure, rely on physical effects which are interacting over spatial scales from subatomic to macroscopic range. Nevertheless, if the material is microscopically heterogenic, it might be appropriate to use phenomenological models for the purpose of civil engineering. Although constantly applied, these models are insufficient for steel materials with microscopic characteristics such as texture, typically occurring in hot rolled steel members or heat affected zones of welded joints. Hence, texture is manifested in crystalline materials as a regular crystallographic structure and crystallite orientation, influencing macroscopic material properties. The analysis of structural response of material with texture (e.g. rolled steel or heat affected zone of a welded joint) obliges the extension of the phenomenological material description of macroscopic scale by means of microscopic information. This paper introduces an enrichment approach for material models based on a hierarchical multiscale methodology. This has been done by describing the grain texture on a mesoscopic scale and coupling it with macroscopic constitutive relations by means of homogenization. Due to a variety of available homogenization methods, the question of an assessment of coupling quality arises. The applicability of the method and the effect of the coupling method on the reliability of the response are presented on an example.
Site superintendents performing project management tasks on construction sites need to access project documents and need to collect information that they observe while inspecting the site. Often, information that is observed on a construction site needs to be integrated into electronic documents or project control systems. In the future, we expect integrated product and process models to be the medium for storing and handling construction project management information. Even though mobile computing devices today are already capable of storing and handling such integrated product and process data models, the user interaction with such large and complex models is difficult and not adequately addressed in the existing research. In this paper, we introduce a system that supports project management tasks on construction sites effectively and efficiently by making integrated product and process models accessible. In order to effectively and efficiently enter or access information, site superintendents need visual representations of the project data that are flexible with respect to the level of detail, the decomposition structure, and the type of visual representation. Based on this understanding of the information and data collection needs, we developed the navigational model framework and the application Site Data Collection System (SiDaCoS), which implements that framework. The navigational model framework allows site superintendents to create customized representations of information contained in a product and process model that correspond to their data access and data collection needs on site.
Collaboration in AEC Design : Web-enabling Applications using Peer-to-Peer Office Communicator
(2004)
A market analysis conducted by Gartner Dataquest in August 2001 has shown the typical characteristics of the AEC design process. High volatility in membership of AEC design groups and members dispersed over several external offices is the common collaboration scenario. Membership is most times short lived, compared to the overall duration of the process. A technical solution has to take that into account by making joining and leaving a collaborative work group very easy. The modelling of roles of collaboration between group members must be based on a commonly understood principle like the publisher / subscriber model, where the individual that is responsible for the distribution of vital information is clear. Security issues and trust in the confidentiality of the system is a central concern for the acceptance of the system. Therefore, keeping the subset of data that will be published under the absolute control of the publisher is a must. This is not the case with server-based scenarios, sometimes even due to psychological reasons. A loosely bound Peer-to-Peer network offers advantages over a server-based solution, because of less administrative overhead and simple installation procedures. In a peer-to-peer environment, a publish/subscribe role model can be more easily implemented. The publish/subscribe model matches the way AEC processes are modelled in real world scenarios today, where legal proof of information exchange between external offices is of high importance. Workflow management systems for small to midsize companies of the AEC industry may adopt the peer-to-peer approach to collaboration in the future. Further investigations are being made on the research level (WINDS) by integrating the viewer and redlining application Collaborate! into a collaborative environment.
Most retaining walls and box culverts built for arterial road construction are simple, and the design process of these structures is often repetitive and labor-intensive because they are so similar in structural configuration. Although some integrated design automation systems developed for retaining walls and box culverts have expedited the design process of these structures, the process of collecting and distributing the resultant engineering documents has not been fully integrated with the computer applications. We have been developing a Web-based design automation system to manage the resultant documents as well as to speed up the repetitive design process. Manipulation of engineering drawings in the Web page is one of the critical functions needed for Web-based design automation. eXtensible Markup Language (XML) and XML-based vector graphics are expected to facilitate the representation of engineering drawings in the Web page. In this paper, we present how we used XML and Scalable Vector Graphics (SVG) to compose engineering drawings and represent them in the Web page. XML Data Island we designed to define drawing components turned out effective in manipulating the engineering drawings in the Web page.
Former achievements for integrated information management have concentrated on interoperability of applications like e.g. CAD, structural analysis or facility management, based on product models introducing additional application independent model layers (core models). In the last years it has become clear, that besides interoperability of autonomous applications, the concurrent processes of model instantiation and evolution have to be modeled, including the relationship to available project resources, persons, legal requirements and communication infrastructure. This paper discusses some basic concepts for an emerging methodology relating the fields of product modeling, project management and workflow systems by elaborating the concept of a process model, which gives a decomposition of the project goals into executable activities. Integrated information management systems should be related to process models to detect pending activities, deadlocks and alternatives of execution. According to the heterogeneous nature of project communication processes, a method for dynamic classification of ad-hoc activities is suggested, that complements predefined highlevel process definitions. In a brief outline of the system architecture, we show how sophisticated information management systems can be broadly made available by using conventional Internet technologies.
The paper contains a description of dynamic effects in the silo wall during the outflow of a stored material. The work allows for determining the danger of construction damage due to resonant vibrations and is of practical importance by determining the influence of cyclic pressures and vibro–creeping during prolonged use of a silo. The paper was devised as a result of tests on silo walls in semi-technical scale. The model is generally applicable and allows for identification of parameters in real- size silos as well.
Three-dimensional models of urban objects play an important role in the urban applications such as urban planning, environmental concerning, or urban disaster mitigations. While the modeling of urban objects is time consuming and storage costing. This paper presents solutions for this. Buildings with regular shapes and plane roofs are constructed into computer models by identifying of graphic elements from the digital maps of urban area to get building base plane and building heights. Buildings with irregular shapes and non-plane roofs are constructed into computer models by employment of a specific system developed by the authors. Road objects and topologies are constructed into computer models by employment of specific algorithms. The solutions presented in this paper has been used in the development of urban disaster mitigation system for Shantou, China.
EXTRA-STATECRAFT
(2011)
Keller Easterling is an architect, urbanist, and writer. Her latest book, Enduring Innocence: Global Architecture and Its Political Masquerades (MIT, 2005), researches familiar spatial products that have landed in difficult or hyperbolic political situations around the world. Her previous book, Organization Space: Landscapes, Highways and Houses in America, applies network theory to a discussion of American infrastructure and development formats. A forthcoming book, Extrastatecraft, researches global infrastructure as a medium of polity. Ms. Easterling is also the author of Call It Home, a laser disc history of suburbia, and American Town Plans. She has recently completed two research installations on the Web: “Wildcards: A Game of Orgman” and “Highline: Plotting NYC.” Her work has been widely published in journals such as Grey Room, Volume, Cabinet, Assemblage, Log, Praxis, Harvard Design Magazine, Perspecta, Metalocus, and ANY. Her work is also included as chapters in numerous publications. She has lectured widely in the United States as well as internationally. Ms. Easterling’s work has been exhibited at the Queens Museum, the Architectural League, the Municipal Arts Society, and the Wexner Center. Easterling is a professor at Yale’s School of Architecture.
Euclidean Clifford analysis is a higher dimensional function theory offering a refinement of classical harmonic analysis. The theory is centered around the concept of monogenic functions, i.e. null solutions of a first order vector valued rotation invariant differential operator called the Dirac operator, which factorizes the Laplacian. More recently, Hermitean Clifford analysis has emerged as a new and successful branch of Clifford analysis, offering yet a refinement of the Euclidean case; it focusses on the simultaneous null solutions, called Hermitean (or h-) monogenic functions, of two Hermitean Dirac operators which are invariant under the action of the unitary group. In Euclidean Clifford analysis, the Clifford-Cauchy integral formula has proven to be a corner stone of the function theory, as is the case for the traditional Cauchy formula for holomorphic functions in the complex plane. Previously, a Hermitean Clifford-Cauchy integral formula has been established by means of a matrix approach. This formula reduces to the traditional Martinelli-Bochner formula for holomorphic functions of several complex variables when taking functions with values in an appropriate part of complex spinor space. This means that the theory of Hermitean monogenic functions should encompass also other results of several variable complex analysis as special cases. At present we will elaborate further on the obtained results and refine them, considering fundamental solutions, Borel-Pompeiu representations and the Teoderescu inversion, each of them being developed at different levels, including the global level, handling vector variables, vector differential operators and the Clifford geometric product as well as the blade level were variables and differential operators act by means of the dot and wedge products. A rich world of results reveals itself, indeed including well-known formulae from the theory of several complex variables.
The general motivation of this research is to develop software to support the handling of the increased complexity of architectural design. In this paper we describe a system providing general support during the whole process. Instead of only developing design tools we are also addressing the problem of the operating environment of these tools. We conclude that design tools have to be integrated in an open, modular, distributed, user friendly and efficient environment. Two major fields have to be addressed - the development of design tools and the realisation of an integrated system as their operation environment. We will briefly focus on the latter by discussing known technologies in the field of information technology and other design disciplines that can be used to realise such an environment. Regarding the first subject we have to state the need of a detailed tool specification. As a solution we suggest a strategy where the tool functions are specified on the basis of a transformation, where a hierarchical process model is mapped into specifications of different design tools realising appropriate support for all sub-processes of architectural design. Using this strategy the main steps to develop such a support system are: implementation of a framework as basis for the integrated design system decision whether the tool specification are already implemented in available tools in this case these tools can be integrated using known methods for tool coupling otherwise new design tools have to be developed according to the framework
In the superelliptic shell joined to a circular cylinder bending stresses are absent when it is subjected to uniform pressure.Some geometrical characteristics have been found. Expressions for determining stresses in the shell crest(in the singular point of plane type) are suggested. The problem of a theoretical critical buckling load of an elongated shell supported by frames is studied. A critical buckling load for two shells with different specifications was found experimentally.
The subject of this talk is the problem of surface design based upon a mesh that may contain both triangular and quadrangular domains. We investigate the cases when such a combined mesh occurs more preferable for bivariate data interpolation than a pure triangulation. First we describe a modification of the well-known flipping algorithm that constructs a locally optimal combined mesh with a predefined quality criterion. Then we introduce two quality measures for triangular and quadrangular domains and present the results of a computational experiment that compares integral interpolation errors and errors in gradients caused by the piecewise surface models produced by the flipping algorithm with the introduced quality measures. The experiment shows that triangular meshes with the Delaunay quality measure provide better interpolation accuracy only if the interpolated function is strictly convex, as well as a saddle-shaped function is better interpolated by bilinear patches within a combined mesh. For a randomly shaped function combined meshes demonstrate smaller error values and better stability in compare with pure triangulations. At the end we consider other resources for mesh improvement, such as excluding >bad< points from the input set for the mesh generating procedure. Because the function values at these points should not be lost, some linear or bilinear patches are replaced by nonlinear patches that pass through the excluded points.
Urban design played a central role for the European dictatorships during the 20th century, it served to legitimate the regime, to produce agreement, to demonstrate power, efficiency and speed, it communicated the social, as well as design projects, of the dictatorial regimes domestically and internationally, it tied old experts, as well as new, to the regime. Dictatorial urban design also played an important role after the fall of the dictatorships: It became the object of structural and verbal handling strategies: of demolition, of transformation, of reconstruction, of forgetting, of suppressing, of re-interpretation and of glorification. The topic area is, therefore, both historical and relevant to the present day. The discussion of the topic area is, like it or not, always embedded in the present state of societal engagement with dictatorships.
In order to even be able to discuss all of these aspects, different conceptual decisions are necessary. In retrospect, these may seem to many as self-evident, although they are anything but. Our thesis is that there are three methodological imperatives, especially, which allow an expanded approach to the topic area “urban design and dictatorship”. First and above all, the tunnel view, focused on individual dictatorships and neglecting the international dimension, must be overcome. Second, the differences in urban design over the course of a dictatorship, through an appropriate periodisation, should be emphasised. Third, we must strive for an open, flexible, but complex concept of urban design. The main focus lies on the urban design of the most influential dictatorships of the first half of the 20th century: Soviet Union, Fascist Italy and Nazi Germany, including the urban design of the autarky periods in Portugal and Spain.
After all, urban design is not just a product of specific historic circumstances. It is a form that continues to have long-term effects, which demonstrates its usefulness and adaptability throughout this process. The urban design products undoubtedly still recall the dictatorial rule under which they were created. However, they are more than a memory space. They are also a living space of the present. They can and should be discussed with respect to their spatial and functional utility for today and tomorrow. Such a perspective is a given for the citizens of a city, but also for city marketing, having marvellous consequences. Only when we do not exclude this dimension a priori, even in academic discussions, can we do justice to the products of dictatorships.
And finally, the view of the urban design of dictatorships can and must contribute to the questioning of simplified and naive conceptions of dictatorships. With urban design in mind, we can observe how dictatorships work and how they were able to prevail. In Europe, these questions are of the highest actuality.
Ideally, multiple computational building evaluation routines (particularly simulation tools) should be coupled in real-time to the representational design model to provide timely performance feed-back to the system user. In this paper we demonstrate how this can be achieved effectively and conveniently via homology-based mapping. We consider two models as homologous if they entail isomorphic topological information. If the general design representation (i.e., a shared object model) is generated in a manner so as to include both the topological building information and pointers to the semantic information base, it can be used to directly derive the domain representations (>enriched< object models with detailed configurational information and filtered semantic data) needed for evaluation purposes. As a proof of concept, we demonstrate a computational design environment that dynamically links an object-oriented space-based design model, with structurally homologous object models of various simulation routines.
In construction engineering, a schedule’s input data, which is usually not exactly known in the planning phase, is considered deterministic when generating the schedule. As a result, construction schedules become unreliable and deadlines are often not met. While the optimization of construction schedules with respect to costs and makespan has been a matter of research in the past decades, the optimization of the robustness of construction schedules has received little attention. In this paper, the effects of uncertainties inherent to the input data of construction schedules are discussed. Possibilities are investigated to improve the reliability of construction schedules by considering alternative processes for certain tasks and by identifying the combination of processes generating the most robust schedule with respect to the makespan of a construction project.
It's not uncommon that analysis and simulation methods are used mainly to evaluate finished designs and to proof their quality. Whereas the potential of such methods is to lead or control a design process from the beginning on. Therefore, we introduce a design method that move away from a “what-if” forecasting philosophy and increase the focus on backcasting approaches. We use the power of computation by combining sophisticated methods to generate design with analysis methods to close the gap between analysis and synthesis of designs. For the development of a future-oriented computational design support we need to be aware of the human designer’s role. A productive combination of the excellence of human cognition with the power of modern computing technology is needed. We call this approach “cognitive design computing”. The computational part aim to mimic the way a designer’s brain works by combining state-of-the-art optimization and machine learning approaches with available simulation methods. The cognition part respects the complex nature of design problems by the provision of models for human-computation interaction. This means that a design problem is distributed between computer and designer. In the context of the conference slogan “back to command”, we ask how we may imagine the command over a cognitive design computing system. We expect that designers will need to let go control of some parts of the design process to machines, but in exchange they will get a new powerful command on complex computing processes. This means that designers have to explore the potentials of their role as commanders of partially automated design processes. In this contribution we describe an approach for the development of a future cognitive design computing system with the focus on urban design issues. The aim of this system is to enable an urban planner to treat a planning problem as a backcasting problem by defining what performance a design solution should achieve and to automatically query or generate a set of best possible solutions. This kind of computational planning process offers proof that the designer meets the original explicitly defined design requirements. A key way in which digital tools can support designers is by generating design proposals. Evolutionary multi-criteria optimization methods allow us to explore a multi-dimensional design space and provide a basis for the designer to evaluate contradicting requirements: a task urban planners are faced with frequently. We also reflect why designers will give more and more control to machines. Therefore, we investigate first approaches learn how designers use computational design support systems in combination with manual design strategies to deal with urban design problems by employing machine learning methods. By observing how designers work, it is possible to derive more complex artificial solution strategies that can help computers make better suggestions in the future.
Procedures of a construction of general solutions for some classes of partial differential equations (PDEs) are proposed and a symmetry operators approach to the raising the orders of the polynomial solutions to linear PDEs are develops. We touch upon an ''operator analytic function theory'' as the solution of a frequent classes of the equations of mathematical physics, when its symmetry operators forms vast enough space. The MAPLE© package programs for the building the operator variables is elaborated also.
The use of process models in the analysis, optimization and simulation of processes has proven to be extremely beneficial in the instances where they could be applied appropriately. However, the Architecture/Engineering/Construction (AEC) industries present unique challenges that complicate the modeling of their processes. A simple Engineering process model, based on the specification of Tasks, Datasets, Persons and Tools, and certain relations between them, have been developed, and its advantages over conventional techniques have been illustrated. Graph theory is used as the mathematical foundation mapping Tasks, Datasets, Persons and Tools to vertices and the relations between them to edges forming a directed graph. The acceptance of process modeling in AEC industries not only depends on the results it can provide, but the ease at which these results can be attained. Specifying a complex AEC process model is a dynamic exercise that is characterized by many modifications over the process model's lifespan. This article looks at reducing specification complexity, reducing the probability for erroneous input and allowing consistent model modification. Furthermore, the problem of resource leveling is discussed. Engineering projects are often executed with limited resources and determining the impact of such restrictions on the sequence of Tasks is important. Resource Leveling concerns itself with these restrictions caused by limited resources. This article looks at using Task shifting strategies to find a near-optimal sequence of Tasks that guarantees consistent Dataset evolution while resolving resource restrictions.
We consider a structural truss problem where all of the physical model parameters are uncertain: not just the material values and applied loads, but also the positions of the nodes are assumed to be inexact but bounded and are represented by intervals. Such uncertainty may typically arise from imprecision during the process of manufacturing or construction, or round-off errors. In this case the application of the finite element method results in a system of linear equations with numerous interval parameters which cannot be solved conventionally. Applying a suitable variable substitution, an iteration method for the solution of a parametric system of linear equations is firstly employed to obtain initial bounds on the node displacements. Thereafter, an interval tightening (pruning) technique is applied, firstly on the element forces and secondly on the node displacements, in order to obtain tight guaranteed enclosures for the interval solutions for the forces and displacements.
By the use of numerical methods and the rapid development of computer technology in the recent years, a large variety, complexity, refinement and capability of partial models have been achieved. This can be noticed in the evaluation of the reliability of structures, e.g. the increased use of spatial structural systems. For the different fields of civil engineering, well developed partial models already exist. Because these partial models are most often used separately, the general view is not entirely illustrated. Until now, there has been no common methodology for evaluating the efficiency of models; the trust in the prediction of a special engineering model has generally relied on the engineer’s experience. In this paper the basics of evaluation of simple models and coupled partial models of frame structures will be discussed using sustainable numerical methods. Furthermore, quality classes (levels) of design tasks will be defined based on their practical relevance. In addition, analysis methods will be systemized. After analysis of different published assessment methods, it may be noted, that the Efficiency Indicator Method (EWM) is most suitable for the observed evaluation problem. Therefore, the EWM was modified to the Model Efficiency Analysis (MEA) for the purpose of a holistic evaluation. The criteria are characterized by two groups, benefit and expenditure, and it is possible by calculating the quotient (benefit/expenditure) to make a statement about the efficiency of the observed models. Presently, the expenditure value is not a subject of investigation, and so the model efficiency is calculated only by the benefit value. This paper also contains the associated criteria catalog, different normalization methods, as well as weighting possibilities.
ARCHITECTURE AND ATMOSPHERE
(2011)
Nathalie Bredella is an architect. She was educated at the TU Berlin and Cooper Union, New York. She received a PhD in Architectural Theory. She taught architectural design at the TU Berlin. She ist the author of Architekturen des Zuschauens. Imaginäre und reale Räume im Film (transcript-verlag). The work is based on an interdisciplinary approach incorporating architecture, film theory and philosophy. Her interests in architectural practice focus on the relationship between spatial strategies, film and media on an urban and architectural scale.
The concept is presented of the sensitivity analysis of the limit state of the structure with respect to selected basic variables. The sensitivity is presented in the form of the probability distribution of the limit state of the structure. The analysis is performed by the problem-oriented Monte Carlo simulation procedure. The procedure is based on the problem's definition of the elementary event, as a structural limit state. Thus the sample space consists of limit states of the structure. Defined on the sample space the one-dimensional random multiplier is introduced. This multiplier refers to the dominant basic variable (group of variables) of the problem. Numerical procedure results in the set of random numbers. Normalized relative histogram of this set is an estimator of the PDF of the limit state of the structure. Estimators of reliability, or the probability of failure are statistical characteristics of this histogram. The procedure is illustrated by the example of sensitivity analysis of the serviceability limit state of monumental structure. It is the colonnade of Licheń Basilica, situated in central Poland. Limit state of the structure is examined with reference to the upper deck horizontal deflection. Wind actions are taken as dominant variables. An assumption is made that the wind load intensities acting on the lower and on the upper storey of the colonnade, respectively, are identically distributed, but correlated random variables. Three correlation variants of these variables are considered. Relevant limit state histograms are analysed thereafter. The paper ends with the conclusions referring to the method and some general remarks on the fully probabilistic design.
The paper deals with the simulation of the non-linear and time dependent behaviour of complex structures in engineering. Such simulations have to provide high accuracy in the prediction of deformations and stability, by taking into account the long term influences of the non-linear behaviour of the material as well as the large deformation and contact conditions. The limiting factors of the computer simulation are the computer run time and the memory requirement during solving large scale problems. To overcome these problems we use a dynamic-explicit time integration procedure for the solution of the semi-discrete equations of motion, which is very suited for parallel processing. In the paper at first we give a brief review of the theoretical background of the mechanical modelling and the dynamic-explicit technique for the solution of the semi-discrete equations of motion. Then the concept of parallel processing will be discussed . A test example concludes the paper.
This paper introduces the current development of electronic catalogues of construction products. The common system architecture of Web-based electronic products catalogues are discussed, followed by a discussion on construction products information standardization and latest distributed-systems technologies for communicating construction products information. Finally, this paper presents an implementation of Web service in e-commerce systems to enable sharing of construction products information.
Lara Schrijver is an assistant professor at the Faculty of Architecture of the TU Delft. She is one of three program leaders for a new research program in the department of architecture, ‘The Architectural Project and its Foundations’. Schrijver holds degrees in architecture from Princeton University and the TU Delft. She received her Ph.D. from the TU Eindhoven in 2005. Schrijver has taught design and theory courses, and contributed to conferences in the Netherlands as well as abroad. She was an editor for OASE, journal for architecture, for ten years, and was co-organizer of the 2006 conference ‘The Projective Landscape’. Her current work revolves around the role of architecture in the city, and its responsibility in defining the public domain. Her first book, Radical Games, on the influence of the 1960s on contemporary discourse, is forthcoming in the spring of 2009.
In spite of the extensive research in dynamic soil-structure interaction (SSI), there still exist miscon-ceptions concerning the role of SSI in the seismic performance of structures, especially the ones founded on soft soil. This is due to the fact that current analytical SSI models that are used to evaluate the influence of soil on the overall structural behavior are approximate models and may involve creeds and practices that are not always precise. This is especially true in the codified approaches which in-clude substantial approximations to provide simple frameworks for the design. As the direct numerical analysis requires a high computational effort, performing an analysis considering SSI is computationally uneconomical for regular design applications. This paper outlines the set up some milestones for evaluating SSI models. This will be achieved by investigating the different assumptions and involved factors, as well as varying the configurations of R/C moment-resisting frame structures supported by single footings which are subject to seismic excita-tions. It is noted that the scope of this paper is to highlight, rather than fully resolve, the above subject. A rough draft of the proposed approach is presented in this paper, whereas a thorough illustration will be carried out throughout the presentation in the course of the conference.