Refine
Document Type
- Article (238) (remove)
Institute
- Professur Theorie und Geschichte der modernen Architektur (62)
- Professur Informatik im Bauwesen (59)
- Institut für Strukturmechanik (ISM) (27)
- In Zusammenarbeit mit der Bauhaus-Universität Weimar (15)
- Professur Bauphysik (11)
- Bauhaus-Institut für zukunftsweisende Infrastruktursysteme (b.is) (5)
- Professur Bauchemie und Polymere Werkstoffe (5)
- Junior-Professur Computational Architecture (4)
- Junior-Professur Organisation und vernetzte Medien (4)
- Professur Stahl- und Hybridbau (4)
Keywords
- Bauhaus-Kolloquium (62)
- Weimar (62)
- Architektur (40)
- 2003 (30)
- Raum (23)
- Digitalisierung (20)
- 2007 (19)
- Bild (19)
- Design (19)
- Bautechnik (18)
Temporary changes in precipitation may lead to sustained and severe drought or massive floods in different parts of the world. Knowing the variation in precipitation can effectively help the water resources decision-makers in water resources management. Large-scale circulation drivers have a considerable impact on precipitation in different parts of the world. In this research, the impact of El Niño-Southern Oscillation (ENSO), Pacific Decadal Oscillation (PDO), and North Atlantic Oscillation (NAO) on seasonal precipitation over Iran was investigated. For this purpose, 103 synoptic stations with at least 30 years of data were utilized. The Spearman correlation coefficient between the indices in the previous 12 months with seasonal precipitation was calculated, and the meaningful correlations were extracted. Then, the month in which each of these indices has the highest correlation with seasonal precipitation was determined. Finally, the overall amount of increase or decrease in seasonal precipitation due to each of these indices was calculated. Results indicate the Southern Oscillation Index (SOI), NAO, and PDO have the most impact on seasonal precipitation, respectively. Additionally, these indices have the highest impact on the precipitation in winter, autumn, spring, and summer, respectively. SOI has a diverse impact on winter precipitation compared to the PDO and NAO, while in the other seasons, each index has its special impact on seasonal precipitation. Generally, all indices in different phases may decrease the seasonal precipitation up to 100%. However, the seasonal precipitation may increase more than 100% in different seasons due to the impact of these indices. The results of this study can be used effectively in water resources management and especially in dam operation.
A phantom-node method is developed for three-node shell elements to describe cracks. This method can treat arbitrary cracks independently of the mesh. The crack may cut elements completely or partially. Elements are overlapped on the position of the crack, and they are partially integrated to implement the discontinuous displacement across the crack. To consider the element containing a crack tip, a new kinematical relation between the overlapped elements is developed. There is no enrichment function for the discontinuous displacement field. Several numerical examples are presented to illustrate the proposed method.
Paraffin Nanocomposites for Heat Management of Lithium-Ion Batteries: A Computational Investigation
(2016)
Lithium-ion (Li-ion) batteries are currently considered as vital components for advances in mobile technologies such as those in communications and transport. Nonetheless, Li-ion batteries suffer from temperature rises which sometimes lead to operational damages or may even cause fire. An appropriate solution to control the temperature changes during the operation of Li-ion batteries is to embed batteries inside a paraffin matrix to absorb and dissipate heat. In the present work, we aimed to investigate the possibility of making paraffin nanocomposites for better heat management of a Li-ion battery pack. To fulfill this aim, heat generation during a battery charging/discharging cycles was simulated using Newman’s well established electrochemical pseudo-2D model. We couple this model to a 3D heat transfer model to predict the temperature evolution during the battery operation. In the later model, we considered different paraffin nanocomposites structures made by the addition of graphene, carbon nanotubes, and fullerene by assuming the same thermal conductivity for all fillers. This way, our results mainly correlate with the geometry of the fillers. Our results assess the degree of enhancement in heat dissipation of Li-ion batteries through the use of paraffin nanocomposites. Our results may be used as a guide for experimental set-ups to improve the heat management of Li-ion batteries.
Meshfree methods (MMs) such as the element free Galerkin (EFG)method have gained popularity because of some advantages over other numerical methods such as the finite element method (FEM). A group of problems that have attracted a great deal of attention from the EFG method community includes the treatment of large deformations and dealing with strong discontinuities such as cracks. One efficient solution to model cracks is adding special enrichment functions to the standard shape functions such as extended FEM, within the FEM context, and the cracking particles method, based on EFG method. It is well known that explicit time integration in dynamic applications is conditionally stable. Furthermore, in enriched methods, the critical time step may tend to very small values leading to computationally expensive simulations. In this work, we study the stability of enriched MMs and propose two mass-lumping strategies. Then we show that the critical time step for enriched MMs based on lumped mass matrices is of the same order as the critical time step of MMs without enrichment. Moreover, we show that, in contrast to extended FEM, even with a consistent mass matrix, the critical time step does not vanish even when the crack directly crosses a node.
The release of the large language model-based chatbot ChatGPT 3.5 in November 2022 has brought considerable attention to the subject of artificial intelligence, not only to the public. From the perspective of higher education, ChatGPT challenges various learning and assessment formats as it significantly reduces the effectiveness of their learning and assessment functionalities. In particular, ChatGPT might be applied to formats that require learners to generate text, such as bachelor theses or student research papers. Accordingly, the research question arises to what extent writing of bachelor theses is still a valid learning and assessment format. Correspondingly, in this exploratory study, the first author was asked to write his bachelor’s thesis exploiting ChatGPT. For tracing the impact of ChatGPT methodically, an autoethnographic approach was used. First, all considerations on the potential use of ChatGPT were documented in logs, and second, all ChatGPT chats were logged. Both logs and chat histories were analyzed and are presented along with the recommendations for students regarding the use of ChatGPT suggested by a common framework. In conclusion, ChatGPT is beneficial for thesis writing during various activities, such as brainstorming, structuring, and text revision. However, there are limitations that arise, e.g., in referencing. Thus, ChatGPT requires continuous validation of the outcomes generated and thus fosters learning. Currently, ChatGPT is valued as a beneficial tool in thesis writing. However, writing a conclusive thesis still requires the learner’s meaningful engagement. Accordingly, writing a thesis is still a valid learning and assessment format. With further releases of ChatGPT, an increase in capabilities is to be expected, and the research question needs to be reevaluated from time to time.
A technique for using object-oriented technologies to write structural analysis software has been developed. The structural design information of an individual building is stored in an object-oriented database. A global database provides general design values as material data and safety factors. A class library for load elements has been evolved to model the transfer of loads in a building. This class library is the basis for the development of further classes for other structural elements such as beams, columns or slabs. A software has been developed to monitor the forces transferred from one structural member to another in a building for load cases and combinations according to Eurocode 1. The results of the analysis are stored in the projects database from which a structural design report may be generated. The software was developed under Microsoft Visual C++. The Microsoft Foundation Class Library (MFC) was used to program the Graphical User Interface (GUI). Object Linking and Embedding (OLE) technology is useful to include any type of OLE server objects for example texts written with a word processor or CAD drawings in the structural design report. The Object-Oriented Database Management System (OODBMS) ObjectStore provides services to store the large amount of objects.
Object-Oriented Damage Information Modeling Concepts and Implementation for Bridge Inspection
(2022)
Bridges are designed to last for more than 50 years and consume up to 50% of their life-cycle costs during their operation phase. Several inspections and assessment actions are executed during this period. Bridge and damage information must be gathered, digitized, and exchanged between different stakeholders. Currently, the inspection and assessment practices rely on paper-based data collection and exchange, which is time-consuming and error-prone, and leads to loss of information. Storing and exchanging damage and building information in a digital format may lower costs and errors during inspection and assessment and support future needs, for example, immediate simulations regarding performance assessment, automated maintenance planning, and mixed reality inspections. This study focused on the concept for modeling damage information to support bridge reviews and structural analysis. Starting from the definition of multiple use cases and related requirements, the data model for damage information is defined independently from the subsequent implementation. In the next step, the implementation via an established standard is explained. Functional tests aim to identify problems in the concept and implementation. To show the capability of the final model, two example use cases are illustrated: the inspection review of the entire bridge and a finite-element analysis of a single component. Main results are the definition of necessary damage data, an object-oriented damage model, which supports multiple use cases, and the implementation of the model in a standard. Furthermore, the tests have shown that the standard is suitable to deliver damage information; however, several software programs lack proper implementation of the standard.
In machine learning, if the training data is independently and identically distributed as the test data then a trained model can make an accurate predictions for new samples of data. Conventional machine learning has a strong dependence on massive amounts of training data which are domain specific to understand their latent patterns. In contrast, Domain adaptation and Transfer learning methods are sub-fields within machine learning that are concerned with solving the inescapable problem of insufficient training data by relaxing the domain dependence hypothesis. In this contribution, this issue has been addressed and by making a novel combination of both the methods we develop a computationally efficient and practical algorithm to solve boundary value problems based on nonlinear partial differential equations. We adopt a meshfree analysis framework to integrate the prevailing geometric modelling techniques based on NURBS and present an enhanced deep collocation approach that also plays an important role in the accuracy of solutions. We start with a brief introduction on how these methods expand upon this framework. We observe an excellent agreement between these methods and have shown that how fine-tuning a pre-trained network to a specialized domain may lead to an outstanding performance compare to the existing ones. As proof of concept, we illustrate the performance of our proposed model on several benchmark problems.
Assessing Essential Qualities of Urban Space with Emotional and Visual Data Based on GIS Technique
(2016)
Finding a method to evaluate people’s emotional responses to urban spaces in a valid and objective way is fundamentally important for urban design practices and related policy making. Analysis of the essential qualities of urban space could be made both more effective and more accurate using innovative information techniques that have become available in the era of big data. This study introduces an integrated method based on geographical information systems (GIS) and an emotion-tracking technique to quantify the relationship between people’s emotional responses and urban space. This method can evaluate the degree to which people’s emotional responses are influenced by multiple urban characteristics such as building shapes and textures, isovist parameters, visual entropy, and visual fractals. The results indicate that urban spaces may influence people’s emotional responses through both spatial sequence arrangements and shifting scenario sequences. Emotional data were collected with body sensors and GPS devices. Spatial clustering was detected to target effective sampling locations; then, isovists were generated to extract building textures. Logistic regression and a receiver operating characteristic analysis were used to determine the key isovist parameters and the probabilities that they influenced people’s emotion. Finally, based on the results, we make some suggestions for design professionals in the field of urban space optimization.
In this paper, an artificial neural network is implemented for the sake of predicting the thermal conductivity ratio of TiO2-Al2O3/water nanofluid. TiO2-Al2O3/water in the role of an innovative type of nanofluid was synthesized by the sol–gel method. The results indicated that 1.5 vol.% of nanofluids enhanced the thermal conductivity by up to 25%. It was shown that the heat transfer coefficient was linearly augmented with increasing nanoparticle concentration, but its variation with temperature was nonlinear. It should be noted that the increase in concentration may cause the particles to agglomerate, and then the thermal conductivity is reduced. The increase in temperature also increases the thermal conductivity, due to an increase in the Brownian motion and collision of particles. In this research, for the sake of predicting the thermal conductivity of TiO2-Al2O3/water nanofluid based on volumetric concentration and temperature functions, an artificial neural network is implemented. In this way, for predicting thermal conductivity, SOM (self-organizing map) and BP-LM (Back Propagation-Levenberq-Marquardt) algorithms were used. Based on the results obtained, these algorithms can be considered as an exceptional tool for predicting thermal conductivity. Additionally, the correlation coefficient values were equal to 0.938 and 0.98 when implementing the SOM and BP-LM algorithms, respectively, which is highly acceptable. View Full-Text
Tall buildings have become an integral part of cities despite all their pros and cons. Some current tall buildings have several problems because of their unsuitable location; the problems include increasing density, imposing traffic on urban thoroughfares, blocking view corridors, etc. Some of these buildings have destroyed desirable views of the city. In this research, different criteria have been chosen, such as environment, access, social-economic, land-use, and physical context. These criteria and sub-criteria are prioritized and weighted by the analytic network process (ANP) based on experts’ opinions, using Super Decisions V2.8 software. On the other hand, layers corresponding to sub-criteria were made in ArcGIS 10.3 simultaneously, then via a weighted overlay (map algebra), a locating plan was created. In the next step seven hypothetical tall buildings (20 stories), in the best part of the locating plan, were considered to evaluate how much of theses hypothetical buildings would be visible (fuzzy visibility) from the street and open spaces throughout the city. These processes have been modeled by MATLAB software, and the final fuzzy visibility plan was created by ArcGIS. Fuzzy visibility results can help city managers and planners to choose which location is suitable for a tall building and how much visibility may be appropriate. The proposed model can locate tall buildings based on technical and visual criteria in the future development of the city and it can be widely used in any city as long as the criteria and weights are localized.
For many purposes geometric information about existing buildings is necessary, e.g. planing of conservation or reconstruction. Architectural photogrammetry is a technique to acquire 3D geometric data of buildings for a CAD model from images. In this paper the state of the art in architectural photogrammetry and some developments towards automation are described. The photogrammetric process consists of image acquisition, orientation and restitution. Special attention is put on digital methods, from digital image acquisition to restitution methods, supported by digital image processing. There are a few field of development towards automation, e.g. feature extraction, extraction of edges and lines and the detection of corresponding points. The acquired data may be used in a CAD environment or for visualization in Virtual Reality Models, using digital orthoimages for texture mapping.
Effective knowledge management is increasingly considered as a cornerstone of sustainable business success. Knowledge management systems are strategically valuable for both ensuring consistency and continuous improvement of various aspects such as quality delivery, productivity and competitiveness. The small and medium enterprises (SMEs) in the construction industry are mostly operating under tighter timeframes, narrower profit margins and more constrained resources. Hence the recently commenced SMILE-SMC (Strategic Management with Information Leveraged Excellece for Small and Medium Contractors) project aims to support the information and knowledge management needs of the small and medium contractors in Hong Kong. This paper presents some snapshots on the SMILE-SMC project, and its conceptualized deliverables with some highlights of recent developments.
Biofeedback constitutes a well-established, non-invasive method to voluntary interfere in emotional processing by means of cognitive strategies. However, treatment durations exhibit strong inter-individual variations and first successes can often be achieved only after a large number of sessions. Sham feedback constitutes a rather untapped approach by providing feedback that does not correspond to the participant’s actual state. The current study aims to gain insights into mechanisms of sham feedback processing in order to support new techniques in biofeedback therapy. We carried out two experiments and applied different types of sham feedback on skin conductance responses and pupil size changes during affective processing. Results indicate that standardized but context-sensitive sham signals based on skin conductance responses exert a stronger influence on emotional regulation compared to individual sham feedback from ongoing pupil dynamics. Also, sham feedback should forego unnatural signal behavior to avoid irritation and skepticism among participants. Altogether, a reasonable combination of stimulus features and sham feedback characteristics enables to considerably reduce the actual bodily responsiveness already within a single session.
The paper summarizes a part of research carried out in ICCI project and provides a current review of ICT infrastructures supporting collaboration. It covers taxonomies, protocols, standards, components, typical subsystems as well as future trends and recommendation for two most important technologies with applications in AEC: (1) EIP (Enterprise information portal) – a single gateway to a company's information, knowledge base, and applications for all actors; (2) RTC (Real-Time Communication and Collaboration technologies) that provide means for asynchronous communication between geographically dislocated people using ICT. Proposed future developments are: orientation towards web services - with building information models, business intelligence, personalization, AEC information retrieval, p2p workspaces and grids.
Current disaster management procedures rely primarily on heuristics which result in their strategies being very cautious and sub-optimum in terms of saving life, minimising damage and returning the building to its normal function. Also effective disaster management demands decentralized, dynamic, flexible, short term and across domain resource sharing, which is not well supported by existing distributing computing infrastructres. The paper proposes a conceptual framework for emergency management in the built environment, using Semantic Grid as an integrating platform for different technologies. The framework supports a distributed network of specialists in built environment, including structural engineers, building technologists, decision analysts etc. It brings together the necessary technology threads, including the Semantic Web (to provide a framework for shared definitions of terms, resources and relationships), Web Services (to provide dynamic discovery and integration) and Grid Computing (for enhanced computational power, high speed access, collaboration and security control) to support rapid formation of virtual teams for disaster management. The proposed framework also make an extensive use of modelling and simulation (both numerical and using visualisations), data mining (to find resources in legacy data sets) and visualisation. It also include a variety of hardware instruments with access to real time data. Furthermore the whole framework is centred on collaborative working by the virtual team. Although focus of this paper is on disaster management, many aspects of the discussed Grid and Visualisation technologies will be useful for any other forms of collaboration. Conclusions are drawn about the possible future impact on the built environment.
Zu den Städtebautheorien von R. Unwin bis E. May : Anmerkungen zum geschichtlichen Hintergrund
(1990)
Wissenschaftliches Kolloquium vom 27. bis 30. Juni 1989 in Weimar an der Hochschule für Architektur und Bauwesen zum Thema: ‚Produktivkraftentwicklung und Umweltgestaltung. Sozialer und wissenschaftlich-technischer Fortschritt in ihren Auswirkungen auf Architektur und industrielle Formgestaltung in unserer Zeit. Zum 100. Geburtstag von Hannes Meyer'
Many construction and facilities management Web sites can be found on the Internet. The interested parties on construction and facilities management Web sites can find databases of best practices, calculators, analyzers, software, expert and decision support systems, neural networks, etc. Technological innovation mainly through changes in the availability of information and communication technology inclusive databases of best practices, calculators, analyzers, software, neural networks, decision support and expert systems that have been provided by a variety of new services developed by the construction and facilities management sectors. Most of all calculators, analyzers, software, decision support and expert systems, neural networks and on-line systems seek to find out how to make the most economic decisions and most of all these decisions are intended only for economic objectives. Alternatives under evaluation have to be evaluated not only from the economic position, but take into consideration qualitative, technical, technological and other characteristics as well. Based on the analysis of the existing calculators, analyzers, information, expert and decision support systems, neural networks and in order to determine most efficient versions of best practices a Decision Support Web-Based System for Construction Innovation (IDSS) was developed by Vilnius Gediminas Technical University.
Multi-criteria decision analysis (MCDA) is an established methodology to support the decision-making of multi-objective problems. For conducting an MCDA, in most cases, a set of objectives (SOO) is required, which consists of a hierarchical structure comprised of objectives, criteria, and indicators. The development of an SOO is usually based on moderated development processes requiring high organizational and cognitive effort from all stakeholders involved. This article proposes elementary interactions as a key paradigm of an algorithm-driven development process for an SOO that requires little moderation efforts. Elementary interactions are self-contained information requests that may be answered with little cognitive effort. The pairwise comparison of elements in the well-known analytical hierarchical process (AHP) is an example of an elementary interaction. Each elementary interaction in the development process presented contributes to the stepwise development of an SOO. Based on the hypothesis that an SOO may be developed exclusively using elementary interactions (EIs), a concept for a multi-user platform is proposed. Essential components of the platform are a Model Aggregator, an Elementary Interaction Stream Generator, a Participant Manager, and a Discussion Forum. While the latter component serves the professional exchange of the participants, the first three components are intended to be automatable by algorithms. The platform concept proposed has been evaluated partly in an explorative validation study demonstrating the general functionality of the algorithms outlined. In summary, the platform concept suggested demonstrates the potential to ease SOO development processes as the platform concept does not restrict the application domain; it is intended to work with little administration moderation efforts, and it supports the further development of an existing SOO in the event of changes in external conditions. The algorithm-driven development of SOOs proposed in this article may ease the development of MCDA applications and, thus, may have a positive effect on the spread of MCDA applications.
Multi-criteria decision analysis (MCDA) is an established methodology to support the decision-making of multi-objective problems. For conducting an MCDA, in most cases, a set of objectives (SOO) is required, which consists of a hierarchical structure comprised of objectives, criteria, and indicators. The development of an SOO is usually based on moderated development processes requiring high organizational and cognitive effort from all stakeholders involved. This article proposes elementary interactions as a key paradigm of an algorithm-driven development process for an SOO that requires little moderation efforts. Elementary interactions are self-contained information requests that may be answered with little cognitive effort. The pairwise comparison of elements in the well-known analytical hierarchical process (AHP) is an example of an elementary interaction. Each elementary interaction in the development process presented contributes to the stepwise development of an SOO. Based on the hypothesis that an SOO may be developed exclusively using elementary interactions (EIs), a concept for a multi-user platform is proposed. Essential components of the platform are a Model Aggregator, an Elementary Interaction Stream Generator, a Participant Manager, and a Discussion Forum. While the latter component serves the professional exchange of the participants, the first three components are intended to be automatable by algorithms. The platform concept proposed has been evaluated partly in an explorative validation study demonstrating the general functionality of the algorithms outlined. In summary, the platform concept suggested demonstrates the potential to ease SOO development processes as the platform concept does not restrict the application domain; it is intended to work with little administration moderation efforts, and it supports the further development of an existing SOO in the event of changes in external conditions. The algorithm-driven development of SOOs proposed in this article may ease the development of MCDA applications and, thus, may have a positive effect on the spread of MCDA applications.
Evaporation is a very important process; it is one of the most critical factors in agricultural, hydrological, and meteorological studies. Due to the interactions of multiple climatic factors, evaporation is considered as a complex and nonlinear phenomenon to model. Thus, machine learning methods have gained popularity in this realm. In the present study, four machine learning methods of Gaussian Process Regression (GPR), K-Nearest Neighbors (KNN), Random Forest (RF) and Support Vector Regression (SVR) were used to predict the pan evaporation (PE). Meteorological data including PE, temperature (T), relative humidity (RH), wind speed (W), and sunny hours (S) collected from 2011 through 2017. The accuracy of the studied methods was determined using the statistical indices of Root Mean Squared Error (RMSE), correlation coefficient (R) and Mean Absolute Error (MAE). Furthermore, the Taylor charts utilized for evaluating the accuracy of the mentioned models. The results of this study showed that at Gonbad-e Kavus, Gorgan and Bandar Torkman stations, GPR with RMSE of 1.521 mm/day, 1.244 mm/day, and 1.254 mm/day, KNN with RMSE of 1.991 mm/day, 1.775 mm/day, and 1.577 mm/day, RF with RMSE of 1.614 mm/day, 1.337 mm/day, and 1.316 mm/day, and SVR with RMSE of 1.55 mm/day, 1.262 mm/day, and 1.275 mm/day had more appropriate performances in estimating PE values. It was found that GPR for Gonbad-e Kavus Station with input parameters of T, W and S and GPR for Gorgan and Bandar Torkmen stations with input parameters of T, RH, W and S had the most accurate predictions and were proposed for precise estimation of PE. The findings of the current study indicated that the PE values may be accurately estimated with few easily measured meteorological parameters.
Cyber security has become a major concern for users and businesses alike. Cyberstalking and harassment have been identified as a growing anti-social problem. Besides detecting cyberstalking and harassment, there is the need to gather digital evidence, often by the victim. To this end, we provide an overview of and discuss relevant technological means, in particular coming from text analytics as well as machine learning, that are capable to address the above challenges. We present a framework for the detection of text-based cyberstalking and the role and challenges of some core techniques such as author identification, text classification and personalisation. We then discuss PAN, a network and evaluation initiative that focusses on digital text forensics, in particular author identification.
According to Eurocode, the computation of bending strength for steel cantilever beams is a straightforward process. The approach is based on an Ayrton-Perry formula adaptation of buckling curves for steel members in compression, which involves the computation of an elastic critical buckling load for considering the instability. NCCI documents offer a simplified formula to determine the critical bending moment for cantilevers beams with symmetric cross-section. Besides the NCCI recommendations, other approaches, e.g. research literature or Finite-Element-Analysis, may be employed to determine critical buckling loads. However, in certain cases they render different results. Present paper summarizes and compares the abovementioned analytical and numerical approaches for determining critical loads and it exemplarily analyses corresponding cantilever beam capacities using numerical approaches based on plastic zones theory (GMNIA).
Particle Simulation and Evaluation of Personal Exposure to Contaminant Sources in an Elevation Space
(2004)
An elevator, which figures a small volume, is normally used by everyone for a short period of time and equipped with simple ventilation system..Any contaminant released within it may cause serious problem. This research adapt a fire and smoke simulation software (FDS) into non-fire indoor airflow scario. Differently from previous research, particles are chosen as a risk evalution unit. A personal and multi-personal exposure model is proposed. The model takes the influence of the human thermal boundary, coughing, inhalation, exhalation, standing position, and the fan factor into account. The model is easy-to-use and suitable for the design of elevator system in practice.
Hannah and her sisters
(2003)
Wissenschaftliches Kolloquium vom 24. bis 27. April 2003 in Weimar an der Bauhaus-Universität zum Thema: ‚MediumArchitektur - Zur Krise der Vermittlung'
Wissenschaftliches Kolloquium vom 19. bis 22. April 2007 in Weimar an der Bauhaus-Universität zum Thema: ‚Die Realität des Imaginären. Architektur und das digitale Bild'
Formal Mutations
(2008)
Wissenschaftliches Kolloquium vom 19. bis 22. April 2007 in Weimar an der Bauhaus-Universität zum Thema: ‚Die Realität des Imaginären. Architektur und das digitale Bild'
Methods with the convergence order p 2 (Newton`s, tangent hyperbolas, tangent parabolas etc.) and their approximate variants are studied. Conditions are presented under which the approximate variants preserve their convergence rate intrinsic to these methods and some computational aspects (possibilities to organize parallel computation, globalization of a method, the solution of the linear equations versus the matrix inversion at every iteration etc.) are discussed. Polyalgorithmic computational schemes (hybrid methods) combining the best features of various methods are developed and possibilities of their application to numerical solution of two-point boundary-value problem in ordinary differential equations and decomposition-coordination problem in convex programming are analyzed.
The purpose of this paper is to review model for finite element techniques for non-linear crack analysis of reinforced concrete beams and slabs. The non-linear behaviour of concrete and steel were described. Some calculations of >self-stress< for concrete and reinforced concrete beam was made. Current computational aspects are discussed. Several remarks for future studies are also given. The numerical model of the concrete and reinforced concrete was described. The paper shows the results of calculations on a reinforced concrete plane stress panel with cracks. The non-linear, numerical model of calculations of reinforced concrete was assumed. Using finite elements method some calculations were made. The results of calculations like displacements, stresses and cracking are shown on diagrams. They were compared with experimental results and other finding. Some conclusions about the described model and results of calculation are shown.
Earthquake is among the most devastating natural disasters causing severe economical, environmental, and social destruction. Earthquake safety assessment and building hazard monitoring can highly contribute to urban sustainability through identification and insight into optimum materials and structures. While the vulnerability of structures mainly depends on the structural resistance, the safety assessment of buildings can be highly challenging. In this paper, we consider the Rapid Visual Screening (RVS) method, which is a qualitative procedure for estimating structural scores for buildings suitable for medium- to high-seismic cases. This paper presents an overview of the common RVS methods, i.e., FEMA P-154, IITK-GGSDMA, and EMPI. To examine the accuracy and validation, a practical comparison is performed between their assessment and observed damage of reinforced concrete buildings from a street survey in the Bingöl region, Turkey, after the 1 May 2003 earthquake. The results demonstrate that the application of RVS methods for preliminary damage estimation is a vital tool. Furthermore, the comparative analysis showed that FEMA P-154 creates an assessment that overestimates damage states and is not economically viable, while EMPI and IITK-GGSDMA provide more accurate and practical estimation, respectively.
Skinless architecture
(2003)
Wissenschaftliches Kolloquium vom 24. bis 27. April 2003 in Weimar an der Bauhaus-Universität zum Thema: ‚MediumArchitektur - Zur Krise der Vermittlung'
Strain measurement is important in mechanical testing. A wide variety of techniques exists for measuring strain in the tensile test; namely the strain gauge, extensometer, stress and strain determined by machine crosshead motion, Geometric Moire technique, optical strain measurement techniques and others. Each technique has its own advantages and disadvantages. The purpose of this study is to quantitatively compare the strain measurement techniques. To carry out the tensile test experiments for S 235, sixty samples were cut from the web of the I-profile in longitudinal and transverse directions in four different dimensions. The geometry of samples are analysed by 3D scanner and vernier caliper. In addition, the strain values were determined by using strain gauge, extensometer and machine crosshead motion. Three techniques of strain measurement are compared in quantitative manner based on the calculation of mechanical properties (modulus of elasticity, yield strength, tensile strength, percentage elongation at maximum force) of structural steel. A statistical information was used for evaluating the results. It is seen that the extensometer and strain gauge provided reliable data, however the extensometer offers several advantages over the strain gauge and crosshead motion for testing structural steel in tension. Furthermore, estimation of measurement uncertainty is presented for the basic material parameters extracted through strain measurement.
The paper presents a general map-based approach to prototyping of products in virtual reality environments. Virtual prototyping of products is considered as a consistent simulation and visualization process mapping the source product model into its target visual representations. The approach enables to interrelate formally the product and visual information models with each other by defining mapping rules, to specify a prototyping scenario as a composition of map instances, and then to explore particular product models in virtual reality environments by interpreting the composed scenario. Having been realized, the proposed approach provides for the strongly formalized method and the common software framework to build virtual prototyping applications. As a result, the applications gain in expressiveness, reusability and reliability, as well as take on additional runtime flexibility...
Civil engineering decision support systems (construction, building life cycle, refurbishment, total quality management, innovation, etc.) created in Lithuania are described in this paper. The above decision support systems comprise of the following constituent parts: data (database and its management system), models (model base and its management system) and a user interface. Presentation of information in databases may be in conceptual (digital, textual, graphical, photographic, video) and quantitative forms. Quantitative information presented involves criteria systems and subsystems, units of measurement, values and initial weight fully defining the variants provided. The databases were developed providing a multiple criteria analysis of alternatives from economical, infrastructure, technical, technological, qualitative, legislative, social and other perspectives. This information is provided in a user-oriented way. Since the analysis of alternatives is usually performed by taking into account economical, infrastructure, technical, technological, qualitative and other factors, a model-base include models which enable a decision maker to carry out a comprehensive analysis of the variants available and make a proper choice. These systems, related questions and practical case study were analysed the paper.
Integrated structural engineering system usually consists of large number of design objects that may be distributed across different platforms. These design objects need to communicate data and information among each other. For efficient communication among design objects a common communication protocol need to be defined. This paper presents the elements of a communication protocol that uses a mediator agent to facilitate communication among design objects. This protocol is termed the Mediative Communication Protocol (MCP). The protocol uses certain design communication performatives and the semantics of an Agent Communication language (ACL) mainly the Knowledge and Query Manipulation Language (KQML) to implement its steps. Details of a Mediator Agent, that will facilitate the communication among design objects, is presented. The Unified Modeling Language (UML) is used to present the Meditative protocol and show how the mediator agent can be use to execute the steps of the meditative communication protocol. An example from structural engineering application is presented to demonstrate and validate the protocol. It is concluded that the meditative protocol is a viable protocol to facilitate object-to-object communication and also has potential to facilitate communication among the different project participants at the higher level of integrated structural engineering systems.
An architecture of a distributed planning system for the building industry has been developed. The emphasis is on highly collaborative environments in steelwork, timber construction etc. where designers concurrently handle 3D models. The overall system connects local design systems by the so-called Design Framework DFW. This framework consists of the definition of distributed components and protocols which make the collaborative design work. The process of collaborative design has been formalized on an abstract level. This paper describes how this has been done. A sample is given to illustrate the mapping of concrete scenarios of the ‘real design world’ to an abstract scenario level. This work is funded by the Deutsche Forschungsgemeinschaft DFG as part of the project SPP1103 (Meißner et al. 2003).
Wissenschaftliches Kolloquium vom 24. bis 27. April 2003 in Weimar an der Bauhaus-Universität zum Thema: ‚MediumArchitektur - Zur Krise der Vermittlung'
A Product Model of a Road
(1997)
Many errors and delays frequently appear when data is exchanged between particular tasks in the lifecycle of the road. Inter-task connections are therefore of great importance for the quality of the final product. The article describes a product model of a road wich is the kernel of an integrated information system intended to support all important stages of the road lifecycle: design, evaluation (through different analysis procedures), construction, and maintainance. Since particular tasks are often executed at different places and in different companies, the interconnections are supported by a special metafile which contains all specific data of the product model. The concept of the integrated system is object and component oriented. Additionally, existing conventional program packages are included to support some common tasks (methods). A conventional relational database system as well as an open spatial database system with the relevant GIS functionality are included to support the data structures of the model.
The paper analyses the influence of the effect of inertia on the reliability of production systems. Systems inertia represents the phenomenon of continuing work for some time after the breakdown of one of the former phases. In our considerations, inertia is treated as the time elapsed from the onset of breakdown till the system's inability to work. A special method had to be devised to investigate the effect of inertia in order to evaluate the reliability of production systems and to attempt algorithmization to control the reliability of production system by means of inertia or reserving. The method of reliability analysis is presented only in an inform aspect. The possibilities of increasing reliability of production systems are listed. A comparison of the redundancy method and inertia method is presented. The results of this comparison and simulated investigations of influence of inertia on reliability of system are essential scope of the paper. Selected conclusions are as follows: when inertia approaches the last phase in the system, its influence on the shape of the distribution of the system's ability increases; an increase in inertia causes an increase in the availability of the system which approaches a certain border value; dependence of the average of a system's disability on inertia has a saddle-like character whereas dependence of the number of breakdowns (stoppages) in the system has the nature of an S-curve.
Antimicrobial resistances (AMR) are ranked among the top ten threats to public health and societal development worldwide. Toilet wastewater contained in domestic wastewater is a significant source of AMR entering the aquatic environment. The current commonly implemented combined sewer systems at times cause overflows during rain events, resulting in the discharge of untreated wastewater into the aquatic environment, thus promoting AMR. In this short research article, we describe an approach to transform combined sewer systems into source separation-modified combined sewer systems that separately treat toilet wastewater. We employ simulations for demonstrating that source separation-modified combined sewer systems reduce the emission of AMR- causing substances by up to 11.5 logarithm levels. Thus, source separation- modified combined sewer systems are amongst the most effective means of combating AMR.
KEYWORDS
Efforts to define standards for representing AEC/FM data have been fairly successful. However defining a standard reference process model has not met with the same success. Yet almost every conceptual modelling or software development project starts by defining the business processes to be supported and the related requirements to be satisfied. This paper describes a new process-centred methodology for user requirements capture developed in the ICCI project (IST-2001-33022). Its essence is in recognising user requirements and use cases in the context of the real construction process, identifying the actors and roles for each individual activity and associating these activities with information, communication and standardisation requirements on the basis of a formalised specification, named the Process Matrix. In the paper we outline the history of process matrix development, introduce the basic structure of the matrix and show how it can be further extended and refined. We present also a web-based software implementation of the developed approach, describe how it has been used in ICCI and outline further perspectives.
Image Analysis Using Human Body Geometry and Size Proportion Science for Action Classification
(2020)
Gestures are one of the basic modes of human communication and are usually used to represent different actions. Automatic recognition of these actions forms the basis for solving more complex problems like human behavior analysis, video surveillance, event detection, and sign language recognition, etc. Action recognition from images is a challenging task as the key information like temporal data, object trajectory, and optical flow are not available in still images. While measuring the size of different regions of the human body i.e., step size, arms span, length of the arm, forearm, and hand, etc., provides valuable clues for identification of the human actions. In this article, a framework for classification of the human actions is presented where humans are detected and localized through faster region-convolutional neural networks followed by morphological image processing techniques. Furthermore, geometric features from human blob are extracted and incorporated into the classification rules for the six human actions i.e., standing, walking, single-hand side wave, single-hand top wave, both hands side wave, and both hands top wave. The performance of the proposed technique has been evaluated using precision, recall, omission error, and commission error. The proposed technique has been comparatively analyzed in terms of overall accuracy with existing approaches showing that it performs well in contrast to its counterparts.
Wissenschaftliches Kolloquium vom 19. bis 22. April 2007 in Weimar an der Bauhaus-Universität zum Thema: ‚Die Realität des Imaginären. Architektur und das digitale Bild'
Plastic structural analysis may be applied without any difficulty and with little effort for structural member verifications with regard to lateral torsional buckling of doubly symmetric rolled I sections. Suchlike analyses can be performed based on the plastic zone theory, specifically using finite beam elements with seven degrees of freedom and 2nd order theory considering material nonlinearity. The existing Eurocode enables these approaches and the coming-up generation will provide corresponding regulations in EN 1993-1-14. The investigations allow the determination of computationally accurate limit loads, which are determined in the present paper for selected structural systems with different sets of parameters, such as length, steel grade and cross section types. The results are compared to approximations gained by more sophisticated FEM analyses (commercial software Ansys Workbench applying solid elements) for reasons of verification/validation. In this course, differences in the results of the numerical models are addressed and discussed. In addition, results are compared to resistances obtained by common design regulations based on reduction factors χlt including regulations of EN 1993-1-1 (including German National Annex) as well as prEN 1993-1-1: 2020-08 (proposed new Eurocode generation). Concluding, correlations of results and their advantages as well as disadvantages are discussed.
Wissenschaftliches Kolloquium vom 19. bis 22. April 2007 in Weimar an der Bauhaus-Universität zum Thema: ‚Die Realität des Imaginären. Architektur und das digitale Bild'
The complexity of the relationships between the actors of a building project requires high efficiency in communication. Among other things, data sharing is crucial. The exchange of data is made possible by interfaces between expert programs, which rely on product models. The latter are neutral standards with formal definitions of building objects and their attributes. This paper deals with the state of the art and the research activities concerning product models in the steel construction domain and the advantages provided by this technology for the sector.
Chemical glass frosting processes are widely used to create visual attractive glass surfaces. A commonly used frosting bath mainly contains ammonium bifluoride (NH4HF2) mixed with hydrochloric acid (HCl). The frosting process consists of several baths. Firstly, the preliminary bath to clean the object. Secondly, the frosting bath which etches the rough light scattering structure into the glass surface. Finally, the washing baths to clean the frosted object. This is where the constituents of the preceding steps accumulate and have to be filtered from the sewage. In the present contribution, phosphoric acid (H3PO4) was used as a substitute for HCl to reduce the amount of ammonium (NH4+) and chloride (Cl−) dissolved in the waste water. In combination with magnesium carbonate (MgCO3), it allows the precipitation of ammonium within the sewage as ammonium magnesium phosphate (MgNH4PO4). However, a trivial replacement of HCl by H3PO4 within the frosting process causes extensive frosting errors, such as inhomogeneous size distributions of the structures or domains that are not fully covered by these structures. By modifying the preliminary bath composition, it was possible to improve the frosting result considerably. To determine the optimal composition of the preliminary bath, a semi-automatic evaluation method has been developed. This method renders the objective comparison of the resulting surface quality possible.
Wissenschaftliches Kolloquium vom 19. bis 22. April 2007 in Weimar an der Bauhaus-Universität zum Thema: ‚Die Realität des Imaginären. Architektur und das digitale Bild'
This paper presents an innovative software platform OpenSTEP intended to build advanced distributed integrated systems and to conduct multidisciplinary collaborative projects in both academy and industry. The paper discusses an open system architecture, methodology, component library and CASE toolkit enabling the developers to build a wide range of interoperable applications and systems compliant with STEP and, particularly, with IFC becoming the increasingly important standard for information integration in architecture, engineering and construction.
Digital Surfacing
(2008)
Wissenschaftliches Kolloquium vom 19. bis 22. April 2007 in Weimar an der Bauhaus-Universität zum Thema: ‚Die Realität des Imaginären. Architektur und das digitale Bild'
A comprehensive framework of information management system for construction projects in China has been established through extensive literature survey and field investigation. It utilizes the potential information technologies and covers the practical management patterns as well as the major aspects of construction project management. It can be used to guide and evaluate the design of the information management systems for construction projects in order to make the system to be applicable to a wide variety of construction projects and survive the changes in project management.
Compiling and disseminating information about incidents and disasters are key to disaster management and relief. But due to inherent limitations of the acquisition process, the required information is often incomplete or missing altogether. To fill these gaps, citizen observations spread through social media are widely considered to be a promising source of relevant information, and many studies propose new methods to tap this resource. Yet, the overarching question of whether and under which circumstances social media can supply relevant information (both qualitatively and quantitatively) still remains unanswered. To shed some light on this question, we review 37 disaster and incident databases covering 27 incident types, compile a unified overview of the contained data and their collection processes, and identify the missing or incomplete information. The resulting data collection reveals six major use cases for social media analysis in incident data collection: (1) impact assessment and verification of model predictions, (2) narrative generation, (3) recruiting citizen volunteers, (4) supporting weakly institutionalized areas, (5) narrowing surveillance areas, and (6) reporting triggers for periodical surveillance. Furthermore, we discuss the benefits and shortcomings of using social media data for closing information gaps related to incidents and disasters.
While Public-Private Partnership (PPP) is widely adopted across various sectors, it raises a question on its meagre utilisation in the housing sector. This paper, therefore, gauges the perspective of the stakeholders in the building industry towards the application of PPP in various building sectors together with housing. It assesses the performance reliability of PPP for housing by learning possible take-aways from other sectors. The role of key stakeholders in the industry becomes highly responsible for an informed understanding and decision-making. To this end, a two-tier investigation was conducted including surveys and expert interviews, with several stakeholders in the PPP industry in Europe, involving the public sector, private sector, consultants, as well as other community/user representatives.
The survey results demonstrated the success rate with PPPs, major factors important for PPPs such as profitability or end-user acceptability, the prevalent practices and trends in the PPP world, and the majority of support expressed in favour of the suitability of PPP for housing. The interviews added more detailed dimensions to the understanding of the PPP industry, its functioning and enabling the formation of a comprehensive outlook. The results present the perspective, approaches, and experiences of stakeholders over PPP practices, current trends and scenarios and their take on PPP in housing. It shall aid in understanding the challenges prevalent in the PPP approach for implementation in housing and enable the policymakers and industry stakeholders to make provisions for higher uptake to accelerate housing provision.
Many problems related to data integration in AEC can be better tackled by an approach that takes into account the heterogeneity of tasks, models and applications but does not require continuous consistency of the evolving design data, at each data management operation. Such an approach must incorporate adequate services that can facilitate reintegration of concurrently modified data at reasonably selected coordination points. In this paper we present a set of methods which, used in combination, can achieve that goal. After a description of the principal envisaged cooperative work scenario each of these methods is discussed in detail and current observations drawn from their software realisation are given. Whilst the suggested approach is valid for any EXPRESS-based data model, the practical focus of work has been on facilitating IFC-driven integration.
Wissenschaftliches Kolloquium vom 24. bis 27. April 2003 in Weimar an der Bauhaus-Universität zum Thema: ‚MediumArchitektur - Zur Krise der Vermittlung'
Wissenschaftliches Kolloquium vom 24. bis 27. April 2003 in Weimar an der Bauhaus-Universität zum Thema: ‚MediumArchitektur - Zur Krise der Vermittlung'
Wissenschaftliches Kolloquium vom 24. bis 27. April 2003 in Weimar an der Bauhaus-Universität zum Thema: ‚MediumArchitektur - Zur Krise der Vermittlung'
Overheating is a major problem in many modern buildings due to the utilization of lightweight constructions with low heat storing capacity. A possible answer to this problem is the emplacement of phase change materials (PCM), thereby increasing the thermal mass of a building. These materials change their state of aggregation within a defined temperature range. Useful PCM for buildings show a phase transition from solid to liquid and vice versa. The thermal mass of the materials is increased by the latent heat. A modified gypsum plaster and a salt mixture were chosen as two materials for the study of their impact on room temperature reduction. For realistic investigations, test rooms were erected where measurements were carried out under different conditions such as temporary air change, alternate internal heat gains or clouding. The experimental data was finally reproduced by dint of a mathematical model.
The classical Internet of things routing and wireless sensor networks can provide more precise monitoring of the covered area due to the higher number of utilized nodes. Because of the limitations in shared transfer media, many nodes in the network are prone to the collision in simultaneous transmissions. Medium access control protocols are usually more practical in networks with low traffic, which are not subjected to external noise from adjacent frequencies. There are preventive, detection and control solutions to congestion management in the network which are all the focus of this study. In the congestion prevention phase, the proposed method chooses the next step of the path using the Fuzzy decision-making system to distribute network traffic via optimal paths. In the congestion detection phase, a dynamic approach to queue management was designed to detect congestion in the least amount of time and prevent the collision. In the congestion control phase, the back-pressure method was used based on the quality of the queue to decrease the probability of linking in the pathway from the pre-congested node. The main goals of this study are to balance energy consumption in network nodes, reducing the rate of lost packets and increasing quality of service in routing. Simulation results proved the proposed Congestion Control Fuzzy Decision Making (CCFDM) method was more capable in improving routing parameters as compared to recent algorithms.
This paper presents an evaluation system for steel structures of hydroelectric power stations, including hydraulic gates and penstocks, based on Fault Tree Analyasis (FTA) and performance maps. This system consists of fault tree diagrams of FTA, performance maps, design and analysis systems, and engineerin databases. These four modules are integrated by appropriate hyperlinks so that the user of this system can use it easily and seamlessly. A well developed system was applied to some illustrative example cases, and they showed that the developed methodology and system worked well and the users found the system useful and effective for their maintenance tasks at powerstations.
The paper analyses the application of 3D gaming technologies in the simulation of processes associated with human resources and machinery on construction sites in order to determine process costs. It addresses the problem of detailing in process simulation. The authors outline special boundary conditions for the simulation of cost-relevant resource processes on virtual construction sites. The approach considers different needs for detailing in process simulation during the planning and building phase. For simulation of process costs on a construction site (contractors view) the level of detail has to be set to high. A prototype for determination of process durations (and hereby process costs) developed at the Bauhaus University Weimar is presented as a result of ongoing researches on detailing in process simulation. It shows the method of process cost determination on a high level of detail (game between excavator and truck) through interaction with the virtual environment of the site.
Few studies have investigated how search behavior affects complex writing tasks. We analyze a dataset of 150 long essays whose authors searched the ClueWeb09 corpus for source material, while all querying, clicking, and writing activity was meticulously recorded. We model the effect of search and writing behavior on essay quality using path analysis. Since the boil-down and build-up writing strategies identified in previous research have been found to affect search behavior, we model each writing strategy separately. Our analysis shows that the search process contributes significantly to essay quality through both direct and mediated effects, while the author's writing strategy moderates this relationship. Our models explain 25–35% of the variation in essay quality through rather simple search and writing process characteristics alone, a fact that has implications on how search engines could personalize result pages for writing tasks. Authors' writing strategies and associated searching patterns differ, producing differences in essay quality. In a nutshell: essay quality improves if search and writing strategies harmonize—build-up writers benefit from focused, in-depth querying, while boil-down writers fare better with a broader and shallower querying strategy.
Wissenschaftliches Kolloquium vom 24. bis 27. April 2003 in Weimar an der Bauhaus-Universität zum Thema: ‚MediumArchitektur - Zur Krise der Vermittlung'
Wissenschaftliches Kolloquium vom 24. bis 27. April 2003 in Weimar an der Bauhaus-Universität zum Thema: ‚MediumArchitektur - Zur Krise der Vermittlung'
SLang - the Structural Language : Solving Nonlinear and Stochastic Problems in Structural Mechanics
(1997)
Recent developments in structural mechanics indicate an increasing need of numerical methods to deal with stochasticity. This process started with the modeling of loading uncertainties. More recently, also system uncertainty, such as physical or geometrical imperfections are modeled in probabilistic terms. Clearly, this task requires close connenction of structural modeling with probabilistic modeling. Nonlinear effects are essential for a realistic description of the structural behavior. Since modern structural analysis relies quite heavily on the Finite Element Method, it seems to be quite reasonable to base stochastic structural analysis on this method. Commercially available software packages can cover deterministic structural analysis in a very wide range. However, the applicability of these packages to stochastic problems is rather limited. On the other hand, there is a number of highly specialized programs for probabilistic or reliability problems which can be used only in connection with rather simplistic structural models. In principle, there is the possibility to combine both kinds of software in order to achieve the goal. The major difficulty which then arises in practical computation is to define the most suitable way of transferring data between the programs. In order to circumvent these problems, the software package SLang (Structural Language) has been developed. SLang is a command interpreter which acts on a set of relatively complex commands. Each command takes input from and gives output to simple data structures (data objects), such as vectors and matrices. All commands communicate via these data objects which are stored in memory or on disk. The paper will show applications to structural engineering problems, in particular failure analysis of frames and shell structures with random loads and random imperfections. Both geometrical and physical nonlinearities are taken into account.
Global structural analyses in civil engineering are usually performed considering linear-elastic material behavior. However, for steel structures, a certain degree of plasticization depending on the member classification may be considered. Corresponding plastic analyses taking material nonlinearities into account are effectively realized using numerical methods. Frequently applied finite elements of two and three-dimensional models evaluate the plasticity at defined nodes using a yield surface, i.e. by a yield condition, hardening rule, and flow rule. Corresponding calculations are connected to a large numerical as well as time-consuming effort and they do not rely on the theoretical background of beam theory, to which the regulations of standards mainly correspond. For that reason, methods using beam elements (one-dimensional) combined with cross-sectional analyses are commonly applied for steel members in terms of plastic zones theories. In these approaches, plasticization is in general assessed by means of axial stress only. In this paper, more precise numerical representation of the combined stress states, i.e. axial and shear stresses, is presented and results of the proposed approach are validated and discussed.
Wissenschaftliches Kolloquium vom 24. bis 27. April 2003 in Weimar an der Bauhaus-Universität zum Thema: ‚MediumArchitektur - Zur Krise der Vermittlung'
Media anthropology is a new and interdisciplinary field of research with very different subjects and methods that seems to be already heavily informed by a comparatively narrow understanding of media as mass media (e.g. TV, Internet, social web, etc.). Therefore, most theories in this field, at least implicitly, employ a hierarchical and often dichotomic preconception of the two poles of media-human relations, by analysing the operationalities and ontologies of the human and the media independently from one another. This article deviates from this line of thought by advocating an expanded, symmetrical and relational understanding of the terms media and human, taking them as always already intermingled facets of a broader dynamic configuration. Starting from a consideration of the historically powerful, yet overlooked media of the so-called habitat diorama, the heuristic concept of “anthropomediality” is to be developed. Eventually, this relational approach may open up a new, interesting field for interrogation of (media-)anthropological analysis in general.
In current AEC practice client requirements are typically recorded in a building program, which, depending on the building type, covers various aspects from the overall goals, activities and spatial needs to very detailed material and condition requirements. This documentation is used as the starting point of the design process, but as the design progresses, it is usually left aside and changes are made incrementally based on the previous design solution. These incremental small changes can lead to a solution that may no longer meet the original requirements. In addition, design is by nature an iterative process and the proposed solutions often also cause evolution in the client requirements. However, the requirements documentation is usually not updated accordingly. Finding the latest updates and evolution of the requirements from the documentation is very difficult, if not impossible. This process can lead to an end result, which is significantly different from the documented requirements. Some important requirements may not be satisfied, and even if the design process was based on agreed-upon changes in the scope and requirements, differences in the requirements documents and in the completed building can lead to well-justified doubts about the quality of the design and construction process...
Physicochemical forces are responsible for the swelling pressure development in saturated bentonites. In this paper, the swelling pressures of several compacted bentonite specimens for a range of dry density of 1.10–1.73 Mg/m3 were measured experimentally. The clay used was a divalent-rich Ca-Mg-bentonite with 12% exchangeable Na+ ions. The theoretical swelling pressure–dry density relationship for the bentonite was determined from the Gouy-Chapman diffuse double-layer theory. A comparison of experimental and theoretical results showed that the experimental swelling pressures are either smaller or greater than their theoretical counterparts within different dry density ranges. It is shown that for dry density of the clay less than about 1.55 Mg/m3, a possible dissociation of ions from the surface of the clay platelets contributed to the diffuse double-layer repulsion. At higher dry densities, the adsorptive forces due to surface and ion hydration dominated the swelling pressures of the clay. A comparison of the modified diffuse double-layer theory equations proposed in the literature to determine the swelling pressures of compacted bentonites and the experimental results for the clay in this study showed that the agreement between the calculated and experimental swelling pressure results is very good for dry densities less than 1.55 Mg/m3, whereas at higher dry densities the use of the equations was found to be limited.
The seismic vulnerability assessment of existing reinforced concrete (RC) buildings is a significant source of disaster mitigation plans and rescue services. Different countries evolved various Rapid Visual Screening (RVS) techniques and methodologies to deal with the devastating consequences of earthquakes on the structural characteristics of buildings and human casualties. Artificial intelligence (AI) methods, such as machine learning (ML) algorithm-based methods, are increasingly used in various scientific and technical applications. The investigation toward using these techniques in civil engineering applications has shown encouraging results and reduced human intervention, including uncertainties and biased judgment. In this study, several known non-parametric algorithms are investigated toward RVS using a dataset employing different earthquakes. Moreover, the methodology encourages the possibility of examining the buildings’ vulnerability based on the factors related to the buildings’ importance and exposure. In addition, a web-based application built on Django is introduced. The interface is designed with the idea to ease the seismic vulnerability investigation in real-time. The concept was validated using two case studies, and the achieved results showed the proposed approach’s potential efficiency
Many researchers are working on developing robots into adequate partners, be it at the working place, be it at home or in leisure activities, or enabling elder persons to lead a self-determined, independent life. While quite some progress has been made in e.g. speech or emotion understanding, processing and expressing, the relations between humans and robots are usually only short-term. In order to build long-term, i.e. social relations, qualities like empathy, trust building, dependability, non-patronizing, and others will be required. But these are just terms and as such no adequate starting points to “program” these capacities even more how to avoid the problems and pitfalls in interactions between humans and robots. However, a rich source for doing this is available, unused until now for this purpose: artistic productions, namely literature, theater plays, not to forget operas, and films with their multitude of examples. Poets, writers, dramatists, screen-writers, etc. have studied for centuries the facets of interactions between persons, their dynamics, and the related snags. And since we wish for human-robot relations as master-servant relations - the human obviously being the master - the study of these relations will be prominent. A procedure is proposed, with four consecutive steps, namely Selection, Analysis, Categorization, and Integration. Only if we succeed in developing robots which are seen as servants we will be successful in supporting and helping humans through robots.
The Carbon journal is pleased to introduce a themed collection of recent articles in the area of computational carbon nanoscience. This virtual special issue was assembled from previously published Carbon articles by Guest Editors Quan Wang and Behrouz Arash, and can be accessed as a set in the special issue section of the journal website homepage: www.journals.elsevier.com/carbon. The article below by our guest editors serves as an introduction to this virtual special issue, and also a commentary on the growing role of computation as a tool to understand the synthesis and properties of carbon nanoforms and their behavior in composite materials.
Wissenschaftliches Kolloquium vom 19. bis 22. April 2007 in Weimar an der Bauhaus-Universität zum Thema: ‚Die Realität des Imaginären. Architektur und das digitale Bild'
In this paper the results of the investigations of the free oscillations of the pre-stressed flexible structure elements are presented . Two cases of the central preliminary stress are investigated : without intermediate fastening of the tie to the flexible element and with the intermediate fastening in the middle of the element length. The given physical model can be applied to the flexible sloping shells and arches, membranes, large space antenna fields (besides flexible elements). The peculiarity of these systems is the possibility of the non-adjacent equilibrium form existence at the definite relations of the physical parameters . The transition from one stable equilibrium form to another, non-adjacent form, may be treated as jump. In this case they are called systems with buckling or the systems with two potential «gaps». These systems commenced the new section of the mathematical physics - the theory of chaos and strange attractors. The analysis of the solutions confirms the received for the first time by the author and given in effect of the oscillation period doubling of the system during the transition from the «small» oscillations relatively center to the >large< relatively all three equilibrium conditions. The character of the frequency (period) dependence on the free oscillation amplitudes of the non-linear system also confirms the received earlier result of the duality of the system behaviour : >small< oscillations possess the qualities of soft system; >large< oscillations possess the qualities of rigid system. The >small< oscillation natural frequency changing, depending on the oscillation amplitudes, is in the internal . Here the frequency takes zero value at the amplitude values Aa and Ad (or Aa and Ae ); the frequency takes maximum value at the amplitude value near point b .The >large< oscillation natural frequency changes in the interval . Here is also observed . The influence of the tie intermediate fastening doesn't introduce qualitative changes in the behaviour of the investigated system. It only increases ( four times ) the critical value of the preliminary tension force
Wissenschaftliches Kolloquium vom 27. bis 30. Juni 1996 in Weimar an der Bauhaus-Universität zum Thema: ‚Techno-Fiction. Zur Kritik der technologischen Utopien'
Priority-rule methods for approximately minimizing the duration of a project subject to minimal and maximal time lags between the activities of the project and limited availability of renewable resources are considered. Such a project can be modelled by a cyclic activity-on-node network. Two generation schemes for constructing feasible schedules are discussed: the serial and parallel schemes. Two different kinds of heuristic procedures are proposed. The sequential or direct method processes the activities or respectively nodes of the project network one after another without considering the strong components separately. The contraction method uses a bottom-up technique. First, a feasible subschedule is determined for each strong component. Second, each strong component is replaced by a single node and the resulting acyclic network is treated by the direct method. In conclusion, some results from an experimental performance analysis of the heuristics are given using a new network generator.
As an optimization that starts from a randomly selected structure generally does not guarantee reasonable optimality, the use of a systemic approach, named the ground structure, is widely accepted in steel-made truss and frame structural design. However, in the case of reinforced concrete (RC) structural optimization, because of the orthogonal orientation of structural members, randomly chosen or architect-sketched framing is used. Such a one-time fixed layout trend, in addition to its lack of a systemic approach, does not necessarily guarantee optimality. In this study, an approach for generating a candidate ground structure to be used for cost or weight minimization of 3D RC building structures with included slabs is developed. A multiobjective function at the floor optimization stage and a single objective function at the frame optimization stage are considered. A particle swarm optimization (PSO) method is employed for selecting the optimal ground structure. This method enables generating a simple, yet potential, real-world representation of topologically preoptimized ground structure while both structural and main architectural requirements are considered. This is supported by a case study for different floor domain sizes.
Image and the Space of the Modern City in Erich Mendelsohn's Amerika: Bilderbuch eines Architekten
(2008)
Wissenschaftliches Kolloquium vom 19. bis 22. April 2007 in Weimar an der Bauhaus-Universität zum Thema: ‚Die Realität des Imaginären. Architektur und das digitale Bild'
Physically Based Modeling and Multi-Physical Simulation System for Wood Structure Fire Performance
(2004)
This research is devoted to promoting the performance-based engineering in wood structure fire. It looks into the characteristic of the material, structural composing and collapse detecting to find out the main factors in the wood structure collapse in fire. The aim of the research is to provide an automatic simulation platform for the complicated circulation. A physically based model for slim member for beams and columns and a frame of multi-physical simulation are provided to implement the system. The physically based model contains material model, structural mechanics model, material mechanics model, as well as geometry model for the compositive simulation. The multi-physical simulation is built on the model and has the capacity to carry out a simulation combining structural, fire (thermal, CFD) and material degradation simulation. The structural and fire simulation rely on two sophisticated software respectively, ANSYS (an FEA software) and FDS (with a core of CFD). Researchers of the paper develop system by themselves to combine the two existing ones. The system has the capability to calculate the wood char to find out the loss of cross-section and to detect the collapse caused in different ways. The paper gives a sample of Chinese traditional house to show how this simulation system works.
Renovation's peculiarities of industrial enterprises in conditions of economic selfsufficiency
(1997)
Probleme of recrienfation of building complex, to the sharp increase of share of reconstruction works, capital repair and modernisation of in-dustrial plants are concidered in this work. The conception of develop-ment and creation of unitified system of expluatation and renovation of industrial plants are worded out. This system is based on date-computer technology and taking into conciderations of real economic relations.
Development and Analysis of Sparse Matrix Concepts for Finite Element Approximation on general Cells
(2004)
In engineering and computing, the finite element approximation is one of the most well-known computational solution techniques. It is a great tool to find solutions for mechanic, fluid mechanic and ecological problems. Whoever works with the finite element method will need to solve a large system of linear equations. There are different ways to find a solution. One way is to use a matrix decomposition technique such as LU or QR. The other possibility is to use an iterative solution algorithm like Conjugate Gradients, Gauß-Seidel, Multigrid Methods, etc. This paper will focus on iterative solvers and the needed storage techniques...
This paper presents an agent-based software, Virtual Administrator System (VAS) for the smallscale maintenance of school buildings. VAS is capable of handling a heavy load of routine, lowtech maintenance jobs. It assigns a different priority to each job application according to its significance and urgency, and automatically adjusts schedules for maintenance engineers when on-site supervision is needed. The system can help ease off the burden of routine small-scale maintenance work, making it more cost-effective and efficient in the overall management of school building maintenance. VAS posts jobs on the Web in a multi-media format and classified all applications into four categories: the on-call maintenance contract, the term maintenance contract, the guaranty maintenance contract, and the regular maintenance contract. It then estimates their urgency level and passes the information to maintenance engineers who will decide whether on-site inspection is needed. Based on the engineers’ feedback, VAS automatically implements the scheduling for inspection as well as sends out real-time or batch notifications to contractors. All these activities are recorded in a database to allow continuous research and data mining and the analysis and diagnosis of specific jobs for followup maintenance plans.
Wissenschaftliches Kolloquium vom 24. bis 27. April 2003 in Weimar an der Bauhaus-Universität zum Thema: ‚MediumArchitektur - Zur Krise der Vermittlung'
Architecture and association
(2003)
Wissenschaftliches Kolloquium vom 24. bis 27. April 2003 in Weimar an der Bauhaus-Universität zum Thema: ‚MediumArchitektur - Zur Krise der Vermittlung'
Individual views on a building product of people involved in the design process imply different models for planning and calculation. In order to interpret these geometrical, topological and semantical data of a building model we identify a structural component graph, a graph of room faces, a room graph and a relational object graph as aids and we explain algorithms to derive these relations. The application of the technique presented is demonstrated by the analysis and discretization of a sample model in the scope of building energy simulation.
The synchronous distributed processing of common source code in the software development process is supported by well proven methods. The planning process has similarities with the software development process. However, there are no consistent and similarly successful methods for applications in construction projects. A new approach is proposed in this contribution.
In this paper we present a computer aided method supporting co-operation between different project partners, such as architects and engineers, on the basis of strictly three-dimensional models. The center of our software architecture is a product model, described by the Industry Foundation Classes (IFC) of the International Alliance for Interoperability (IAI). From this a geometrical model is extracted and automatically transferred to a computational model serving as a basis for various simulation tasks. In this paper the focus is set on the advantage of the fully three-dimensional structural analysis performed by p-version of the finite element analysis. Other simulation methods are discussed in a separate contribution of this Volume (Treeck 2004). The validity of this approach will be shown in a complex example.
In Germany, bridges have an average age of 40 years. A bridge consumes between 0.4% and 2% of its construction cost per year over its entire life cycle. This means that up to 80% of the construction cost are additionally needed for operation, inspection, maintenance, and destruction. Current practices rely either on paperbased inspections or on abstract specialist software. Every application in the inspection and maintenance sector uses its own data model for structures, inspections, defects, and maintenance. Due to this, data and properties have to be transferred manually, otherwise a converter is necessary for every data exchange between two applications. To overcome this issue, an adequate model standard for inspections, damage, and maintenance is necessary. Modern 3D models may serve as a single source of truth, which has been suggested in the Building Information Modeling (BIM) concept. Further, these models offer a clear visualization of the built infrastructure, and improve not only the planning and construction phases, but also the operation phase of construction projects. BIM is established mostly in the Architecture, Engineering, and Construction (AEC) sector to plan and construct new buildings. Currently, BIM does not cover the whole life cycle of a building, especially not inspection and maintenance. Creating damage models needs the building model first, because a defect is dependent on the building component, its properties and material. Hence, a building information model is necessary to obtain meaningful conclusions from damage information. This paper analyzes the requirements, which arise from practice, and the research that has been done in modeling damage and related information for bridges. With a look at damage categories and use cases related to inspection and maintenance, scientific literature is discussed and synthesized. Finally, research gaps and needs are identified and discussed.
The goal of the research is the development of a computer system to plan, simulate and visualize erection processes in construction. In the research construction cranes are treated as robots with predefined degrees of freedom and crane-specific motion planning techniques are developed to generate time-optimized and collision-free paths for each piece to be erected in the project. Using inverse kinematics and structural dynamics simulation, the computer system then computes the crane motions and velocities necessary to achieve the previously calculated paths. The main benefits of the research are the accurate planning and scheduling of crane operations leading to optimization of crane usage and project schedules, as well as improving overall crane safety in the project. This research is aimed at the development of systems that will allow computer-assisted erection of civil infrastructure and ultimately to achieve fully-automated erection processes using robotic cranes...