600 Technik, Medizin, angewandte Wissenschaften
Refine
Document Type
- Doctoral Thesis (57)
- Article (31)
- Master's Thesis (11)
- Preprint (7)
- Bachelor Thesis (4)
- Book (2)
- Report (2)
Institute
- Institut für Strukturmechanik (ISM) (35)
- Professur Baubetrieb und Bauverfahren (9)
- Bauhaus-Institut für zukunftsweisende Infrastruktursysteme (b.is) (8)
- Professur Informatik in der Architektur (7)
- Professur Modellierung und Simulation - Konstruktion (7)
- F. A. Finger-Institut für Baustoffkunde (FIB) (5)
- Professur Bauchemie und Polymere Werkstoffe (4)
- Professur Stahl- und Hybridbau (4)
- Professur Bauphysik (3)
- Professur Betriebswirtschaftslehre im Bauwesen (3)
- Institut für Konstruktiven Ingenieurbau (IKI) (2)
- Professur Baumanagement und Bauwirtschaft (2)
- Professur Baustatik und Bauteilfestigkeit (2)
- Professur Bodenmechanik (2)
- Professur Modellierung und Simulation - Mechanik (2)
- Professur Siedlungswasserwirtschaft (2)
- Professur Stochastik und Optimierung (2)
- Professur Werkstoffe des Bauens (2)
- Bauhaus-Institut für Geschichte und Theorie der Architektur und Planung (1)
- Fakultät Kunst und Gestaltung (1)
- Graduiertenkolleg 1462 (1)
- Materialforschungs- und -prüfanstalt an der Bauhaus-Universität (1)
- Professur Angewandte Mathematik (1)
- Professur Aufbereitung von Baustoffen und Wiederverwertung (1)
- Professur Biotechnologie in der Ressourcenwirtschaft (1)
- Professur Grundbau (1)
- Professur Human-Computer Interaction (1)
- Professur Informatik im Bauwesen (1)
- Professur Konstruktives Entwerfen und Tragwerkslehre (1)
- Professur Medieninformatik (1)
- Professur Sozialwissenschaftliche Stadtforschung (1)
- Professur Stahlbau (1)
- Professur Tragwerkslehre (1)
- Promotionsstudiengang Kunst und Design-Freie Kunst-Medienkunst (Ph.D) (1)
- bauhaus.institut für experimentelle Architektur (1)
Keywords
- Maschinelles Lernen (8)
- OA-Publikationsfonds2022 (8)
- Finite-Elemente-Methode (6)
- Isogeometric Analysis (5)
- Aerodynamik (4)
- Beton (4)
- Erdbeben (4)
- Simulation (4)
- machine learning (4)
- Abwasser (3)
The imperative to transform current energy provisions is widely acknowledged. However, scant attention has hitherto been directed toward rural municipalities and their innate resources, notably biogenic resources. In this paper, a methodological framework is developed to interconnect resources from waste, wastewater, and agricultural domains for energy utilization. This entails cataloging existing resources, delineating their potential via quantitative assessments utilizing diverse technologies, and encapsulating them in a conceptual model. The formulated models underwent iterative evaluation with engagement from diverse stakeholders. Consequently, 3 main concepts, complemented by 72 sub-concepts, were delineated, all fostering positive contributions to climate protection and providing heat supply in the rural study area. The outcomes’ replicability is underscored by the study area’s generic structure and the employed methodology. Through these inquiries, a framework for the requisite energy transition, with a pronounced emphasis on the coupling of waste, wastewater, and agriculture sectors in rural environments, is robustly analyzed.
The study presents a Machine Learning (ML)-based framework designed to forecast the stress-strain relationship of arc-direct energy deposited mild steel. Based on microstructural characteristics previously extracted using microscopy and X-ray diffraction, approximately 1000 new parameter sets are generated by applying the Latin Hypercube Sampling Method (LHSM). For each parameter set, a Representative Volume Element (RVE) is synthetically created via Voronoi Tessellation. Input raw data for ML-based algorithms comprises these parameter sets or RVE-images, while output raw data includes their corresponding stress-strain relationships calculated after a Finite Element (FE) procedure. Input data undergoes preprocessing involving standardization, feature selection, and image resizing. Similarly, the stress-strain curves, initially unsuitable for training traditional ML algorithms, are preprocessed using cubic splines and occasionally Principal Component Analysis (PCA). The later part of the study focuses on employing multiple ML algorithms, utilizing two main models. The first model predicts stress-strain curves based on microstructural parameters, while the second model does so solely from RVE images. The most accurate prediction yields a Root Mean Squared Error of around 5 MPa, approximately 1% of the yield stress. This outcome suggests that ML models offer precise and efficient methods for characterizing dual-phase steels, establishing a framework for accurate results in material analysis.
Diese Arbeit soll einen Beitrag zum Neuen Steuerungsmodell der Öffentlichen Verwaltung auf staatlicher Ebene in Deutschland leisten. Sie dient der Untersuchung zum Aufbau eines ökonomischen Modells zur Koordination von Dienstliegenschaften auf staatlicher Ebene.
Die Untersuchung der Dienstliegenschaften zeigt, dass diese eine interne Dienstbarkeit des Staates als Wirtschaftssubjekt an den Staat als Hoheitsträger fingieren.
Die Untersuchung der Liegenschaftsverwaltung belegt, dass sie vor allem ein Controlling-Instrument für den Informationsfluss zu Entscheidungen über Dienstliegenschaften zwischen dem Hoheitsträger und dem Wirtschaftssubjekt darstellt.
Die Untersuchung der Transaktionskosten beweist, dass eine Koordination am effizientesten über die eigene Organisation mittels dezentraler Aufgabenkonzentration in Form eines SSC erreicht werden kann.
Die Untersuchung der Handlungs- und Verfügungsrechte ergab, dass die Verfügungsrechte an den Dienstliegenschaften weiter den Ressorts bzw. Nutzern obliegen. Allein das Handlungsrecht der Aufgabenwahrnehmung ist an die Liegenschaftsverwaltung übergegangen.
Die Untersuchung der Prinzipal-Agent-Verhältnisse teilt der Liegenschaftsverwaltung die Rolle eines Erfüllungsgehilfen des Wirtschaftssubjektes zu. Die monetäre Abwicklung der Ge-schäfte zwischen dem Wirtschaftssubjekt und dem Hoheitsträger obliegt als zuarbeitende Organisationseinheit ohne eigene Entscheidungsgewalt der Liegenschaftsverwaltung.
Aus diesen Thesen ergeben sich monetäre Handlungsmöglichkeiten, die in den Aufbau des Modells einfließen. Es rückt das Wirtschaftssubjekt als Entscheidungsträger über den Ressourceneinsatz in den Mittelpunkt der Betrachtung. Der Hoheitsträger ist hierbei nur der Forderungsberechtigte an das Wirtschaftssubjekt und die Liegenschaftsverwaltung dessen immobilienwirtschaftlicher Vertreter.
Diese Konstellation berücksichtigend erfolgt dem Credo des NPM gemäß die Beschreibung des Modells und des institutionellen Arrangements nach den gängigen Modellen aus der Privatwirtschaft. Anhand der Untersuchungsergebnisse wird eine Liegenschaftsverwaltung skizziert, welche einerseits den tatsächlichen betriebswirtschaftlichen Anforderungen der Immobilienwirtschaft und andererseits den Gegebenheiten des haushaltswirtschaftlichen Umfeldes der öffentlichen Verwaltung gerecht werden kann.
Das aufgebaute Realmodell dient gleichzeitig dem Vergleich mit dem bestehenden und in der Praxis angewandten Mieter-Vermieter-Modell als Idealmodell. Der Vergleich zeigt, dass sich aus der Untersuchung zivilrechtlicher Institutionen zwar kein schuldrechtliches, wohl aber ein dingliches Recht ableiten lässt. Damit einher geht die Vermutung, dass es sich bei dem Mieter-Vermieter-Modell im staatlichen Bereich um einen Fall von Modellplatonismus handelt.
Structures under wind action can exhibit various aeroelastic interaction phenomena, which can lead to destructive and catastrophic events. Such unstable interaction can be beneficially used for small-scale aeroelastic energy harvesting. Proper understanding and prediction of fluid−structure interactions (FSI) phenomena are therefore crucial in many engineering fields. This research intends to develop coupled FSI models to extend the applicability of Vortex Particle Methods (VPM) for numerically analysing the complex FSI of thin-walled flexible structures under steady and fluctuating incoming flows. In this context, the flow around deforming thin bodies is analysed using the two-dimensional and pseudo-three-dimensional implementations of VPM. The structural behaviour is modelled and analysed using the Finite Element Method. The partitioned coupling approach is considered because of the flexibility of using different mathematical procedures for solving fluid and solid mechanics. The developed coupled models are validated with several benchmark FSI problems in the literature. Finally, the models are applied to several fundamental and application field of FSI problems of different thin-walled flexible structures irrespective of their size.
Biomembranes are selectively permeable barriers that separate the internal components of the cell from its surroundings. They have remarkable mechanical behavior which is characterized by many phenomena, but most noticeably their fluid-like in-plane behavior and solid-like out-of-plane behavior. Vesicles have been studied in the context of discrete models, such as Molecular Dynamics, Monte Carlo methods, Dissipative Particle Dynamics, and Brownian Dynamics. Those methods, however, tend to have high computational costs, which limited their uses for studying atomistic details. In order to broaden the scope of this research, we resort to the continuum models, where the atomistic details of the vesicles are neglected, and the focus shifts to the overall morphological evolution. Under the umbrella of continuum models, vesicles morphology has been studied extensively. However, most of those studies were limited to the mechanical response of vesicles by considering only the bending energy and aiming for the solution by minimizing the total energy of the system. Most of the literature is divided between two geometrical representation methods; the sharp interface methods and the diffusive interface methods. Both of those methods track the boundaries and interfaces implicitly. In this research, we focus our attention on solving two non-trivial problems. In the first one, we study a constrained Willmore problem coupled with an electrical field, and in the second one, we investigate the hydrodynamics of a vesicle doublet suspended in an external viscous fluid flow.
For the first problem, we solve a constrained Willmore problem coupled with an electrical field using isogeometric analysis to study the morphological evolution of vesicles subjected to static electrical fields. The model comprises two phases, the lipid bilayer, and the electrolyte. This two-phase problem is modeled using the phase-field method, which is a subclass of the diffusive interface methods mentioned earlier. The bending, flexoelectric, and dielectric energies of the model are reformulated using the phase-field parameter. A modified Augmented-Lagrangian (ALM) approach was used to satisfy the constraints while maintaining numerical stability and a relatively large time step. This approach guarantees the satisfaction of the constraints at each time step over the entire temporal domain.
In the second problem, we study the hydrodynamics of vesicle doublet suspended in an external viscous fluid flow. Vesicles in this part of the research are also modeled using the phase-field model. The bending energy and energies associated with enforcing the global volume and area are considered. In addition, the local inextensibility condition is ensured by introducing an additional equation to the system. To prevent the vesicles from numerically overlapping, we deploy an interaction energy definition to maintain a short-range repulsion between the vesicles. The fluid flow is modeled using the incompressible Navier-Stokes equations and the vesicle evolution in time is modeled using two advection equations describing the process of advecting each vesicle by the fluid flow. To overcome the velocity-pressure saddle point system, we apply the Residual-Based Variational MultiScale (RBVMS) method to the Navier-Stokes equations and solve the coupled systems using isogeometric analysis. We study vesicle doublet hydrodynamics in shear flow, planar extensional flow, and parabolic flow under various configurations and boundary conditions.
The results reveal several interesting points about the electrodynamics and hydrodynamics responses of single vesicles and vesicle doublets. But first, it can be seen that isogeometric analysis as a numerical tool has the ability to model and solve 4th-order PDEs in a primal variational framework at extreme efficiency and accuracy due to the abilities embedded within the NURBS functions without the need to reduce the order of the PDE by creating an intermediate environment. Refinement whether by knot insertion, order increasing or both is far easier to obtain than traditional mesh-based methods. Given the wide variety of phenomena in natural sciences and engineering that are mathematically modeled by high-order PDEs, the isogeometric analysis is among the most robust methods to address such problems as the basis functions can easily attain high global continuity.
On the applicational side, we study the vesicle morphological evolution based on the electromechanical liquid-crystal model in 3D settings. This model describing the evolution of vesicles is composed of time-dependent, highly nonlinear, high-order PDEs, which are nontrivial to solve. Solving this problem requires robust numerical methods, such as isogeometric analysis. We concluded that the vesicle tends to deform under increasing magnitudes of electric fields from the original sphere shape to an oblate-like shape. This evolution is affected by many factors and requires fine-tuning of several parameters, mainly the regularization parameter which controls the thickness of the diffusive interface width. But it is most affected by the method used for enforcing the constraints. The penalty method in presence of an electrical field tends to lock on the initial phase-field and prevent any evolution while a modified version of the ALM has proven to be sufficiently stable and accurate to let the phase-field evolve while satisfying the constraints over time at each time step. We show additionally the effect of including the flexoelectric nature of the Biomembranes in the computation and how it affects the shape evolution as well as the effect of having different conductivity ratios. All the examples were solved based on a staggered scheme, which reduces the computational cost significantly.
For the second part of the research, we consider vesicle doublet suspended in a shear flow, in a planar extensional flow, and in a parabolic flow. When the vesicle doublet is suspended in a shear flow, it can either slip past each other or slide on top of each other based on the value of the vertical displacement, that is the vertical distance between the center of masses between the two vesicles, and the velocity profile applied. When the vesicle doublet is suspended in a planar extensional flow in a configuration that resembles a junction, the time in which both vesicles separate depends largely on the value of the vertical displacement after displacing as much fluid from between the two vesicles. However, when the vesicles are suspended in a tubular channel with a parabolic fluid flow, they develop a parachute-like shape upon converging towards each other before exiting the computational domain from the predetermined outlets. This shape however is affected largely by the height of the tubular channel in which the vesicle is suspended. The velocity essential boundary conditions are imposed weakly and strongly. The weak implementation of the boundary conditions was used when the velocity profile was defined on the entire boundary, while the strong implementation was used when the velocity profile was defined on a part of the boundary. The strong implementation of the essential boundary conditions was done by selectively applying it to the predetermined set of elements in a parallel-based code. This allowed us to simulate vesicle hydrodynamics in a computational domain with multiple inlets and outlets. We also investigate the hydrodynamics of oblate-like shape vesicles in a parabolic flow. This work has been done in 2D configuration because of the immense computational load resulting from a large number of degrees of freedom, but we are actively seeking to expand it to 3D settings and test a broader set of parameters and geometrical configurations.
Care of ageing adults has become a dominant field of application for assistive robot technologies, promising support for ageing adults residing in care homes and staff, in dealing with practical routine tasks and providing social and emotional relieve. A time consuming and human intensive necessity is the maintenance of high hygiene quality in care homes. Robotic vacuum cleaners have been proven effective for doing the job elsewhere, but—in the context of care homes—are counterproductive for residents’ well-being and do not get accepted. This is because people with dementia manifest their agency in more implicit and emotional ways, while making sense of the world around them. Starting from these premises, we explored how a zoomorphic designed vacuum cleaner could better accommodate the sensemaking of people with dementia. Our design reconceptualises robotic vacuum cleaners as a cat-like robot, referring to a playful behaviour and appearance to communicate a non-threatening and familiar role model. Data from an observational study shows that residents responded positively to our prototype, as most of them engaged playfully with it as if it was a pet or a cat-like toy, for example luring it with gestures. Some residents simply ignored the robot, indicating that it was not perceived as frightening or annoying. The level of activity influenced reactions; residents ignored our prototype if busy with other occupations, which proves that it did not cause significant disturbance. We further report results from focus group sessions with formal and informal caregivers who discussed a video prototype of our robot. Caregivers encouraged us to enhance the animal like characteristics (in behaviour and materiality) even further to result in richer interactions and provoke haptic pleasure but also pointed out that residents should not mistake the robot for a real cat.
Antimicrobial resistances (AMR) are ranked among the top ten threats to public health and societal development worldwide. Toilet wastewater contained in domestic wastewater is a significant source of AMR entering the aquatic environment. The current commonly implemented combined sewer systems at times cause overflows during rain events, resulting in the discharge of untreated wastewater into the aquatic environment, thus promoting AMR. In this short research article, we describe an approach to transform combined sewer systems into source separation-modified combined sewer systems that separately treat toilet wastewater. We employ simulations for demonstrating that source separation-modified combined sewer systems reduce the emission of AMR- causing substances by up to 11.5 logarithm levels. Thus, source separation- modified combined sewer systems are amongst the most effective means of combating AMR.
KEYWORDS
Vertical green system for gray water treatment: Analysis of the VertiKKA-module in a field test
(2022)
This work presents a modular Vertical Green System (VGS) for gray water treatment, developed at the Bauhaus-Universität Weimar. The concept was transformed into a field study with four modules built and tested with synthetic gray water. Each module set contains a small and larger module with the same treatment substrate and was fed hourly. A combination of lightweight structural material and biochar of agricultural residues and wood chips was used as the treatment substrate. In this article, we present the first 18 weeks of operation. Regarding the treatment efficiency, the parameters chemical oxygen demand (COD), total phosphorous (TP), ortho-phosphate (ortho-P), total bound nitrogen (TNb), ammonium nitrogen (NH4-N), and nitrate nitrogen (NO3-N) were analyzed and are presented in this work. The results of the modules with agricultural residues are promising. Up to 92% COD reduction is stated in the data. The phosphate and nitrogen fractions are reduced significantly in these modules. By contrast, the modules with wood chips reduce only 67% of the incoming COD and respectively less regarding phosphates and the nitrogen fraction.
Aktuell findet aufgrund gesellschaftspolitischer Forderungen in vielen Industriezweigen ein Umdenken in Richtung Effizienz und Ökologie aber auch Digitalisierung und Industrie 4.0 statt. In dieser Hinsicht steht die Bauindustrie, im Vergleich zu Industrien wie IT, Automobil- oder Maschinenbau, noch am Anfang.
Dabei sind die Potentiale zur Einsparung und Optimierung gerade in der Bauindustrie aufgrund der großen Mengen an zu verarbeiteten Materialien besonders hoch. Die internationale Ressourcen- und Klimadebatte führt verstärkt dazu, dass auch in der Zement- und Betonherstellung neue Konzepte erstellt und geprüft werden. Einerseits erfolgt intensive Forschung und Entwicklung im Bereich alternativer, klimafreundlicher Zemente. Andererseits werden auch auf Seiten der Betonherstellung innovative materialsparende Konzepte geprüft, wie die aktuelle Entwicklung von 3D-Druck mit Beton zeigt.
Aufgrund der hohen Anforderungen an Konstruktion, Qualität und Langlebigkeit von Bauwerken, besitzen Betonfertigteile oftmals Vorteile gegenüber Ortbeton. Die hohe Oberflächenqualität und Dauerhaftigkeit aber auch die Gleichmäßigkeit und witterungsunabhängige Herstellung sind Merkmale, die im Zusammenhang mit Betonfertigteilen immer wieder erwähnt werden. Dabei ist es essenziell, dass auch der Betonherstellungsprozess im Fertigteilwerk kritisch hinterfragt wird, damit eine effizientere und nachhaltigere Produktion von Betonfertigteilen möglich wird.
Bei der Herstellung von Betonteilen im Fertigteilwerk liegt ein besonderer Fokus auf der Optimierung der Frühfestigkeitsentwicklung. Hohe Frühfestigkeiten sind Voraussetzung für einen hochfrequenten Schalungszyklus, was Arbeiten im 2- bzw. 3-Schichtbetrieb ermöglicht. Oft werden zur Sicherstellung hoher Frühfestigkeiten hochreaktive Zemente in Kombination mit hohen Zementgehalten im Beton und/oder einer Wärmebehandlung eingesetzt. Unter dieser Prämisse ist eine ökologisch nachhaltige Betonproduktion mit verminderter CO2 Bilanz nicht möglich.
In der vorliegenden Arbeit wird ein neues Verfahren zur Beschleunigung von Beton eingeführt. Hierbei werden die Bestandteile Zement und Wasser (Zementsuspension) mit Ultraschall vorbehandelt. Ausgangspunkt der Arbeit sind vorangegangene Untersuchungen zum Einfluss von Ultraschall auf die Hydration von Zement bzw. dessen Hauptbestandteil Tricalciumsilikat (C3S), die im Rahmen dieser Arbeit weiter vertieft werden. Darüber hinaus wird die Produktion von Beton mit Ultraschall im Technikumsmaßstab betrachtet. Die so erlangten Erfahrungen dienten dazu, das Ultraschall-Betonmischsystem weiterzuentwickeln und erstmalig zur industriellen Betonproduktion zu nutzen.
In der vorliegenden Arbeit werden die Auswirkungen von Ultraschall auf die Hydratation von C3S zunächst weitergehend und grundlegend untersucht. Dies erfolgte mittels Messung der elektrischen Leitfähigkeit, Analyse der Ionenkonzentration (ICP-OES), Thermoanalyse, Messung der BET-Oberfläche sowie einer optischen Auswertung mittels Rasterelektronenmikroskopie (REM). Der Fokus liegt auf den ersten Stunden der Hydratation, also der Zeit, die durch die Ultraschallbehandlung am stärksten beeinflusst wird.
In den Untersuchungen zeigt sich, dass die Beschleunigungswirkung von Ultraschall in verdünnten C3S Suspensionen (w/f-Wert = 50) stark von der Portlanditkonzentration der Lösung abhängt. Je niedriger die Portlanditkonzentration, desto größer ist die Beschleunigung. Ergänzende Untersuchungen der Ionenkonzentration der Lösung sowie Untersuchungen am hydratisierten C3S zeigen, dass unmittelbar nach der Beschallung (nach ca. 15 Minuten Hydratation) erste Hydratphasen vorliegen. Die durch Ultraschall initiiere Beschleunigung ist in den ersten 24 Stunden am stärksten und klingt dann sukzessive ab. Die Untersuchungen schließen mit Experimenten an C3S-Pasten (w/f-Wert = 0,50), die die Beobachtungen an den verdünnten Suspensionen bestätigen und infolge der Beschallung ein früheres Auftreten und einen größeren Anteil an C-S-H Phasen zeigen. Es wird gefolgert, dass die unmittelbar infolge von Ultraschall erzeugten C-S-H Phasen als Kristallisationskeim während der folgenden Reaktion dienen und daher Ultraschall als in-situ Keimbildungstechnik angesehen werden kann. Optisch zeigt sich, dass die C-S-H Phasen der beschallten Pasten nicht nur viel früher auftreten, sondern kleiner sind und fein verteilt über die Oberfläche des C3S vorliegen. Auch dieser Effekt wird als vorteilhaft für den sich anschließenden regulären Strukturaufbau angesehen.
Im nächsten Schritt wird daher der Untersuchungsfokus vom Modellsystem mit C3S auf Portlandzement erweitert. Hierbei wird der Frage nachgegangen, wie sich eine Änderung der Zusammensetzung der Zementsuspension (w/z-Wert, Fließmittelmenge) beziehungsweise eine Änderung des Ultraschallenergieeintrag auf die Fließeigenschaften und das Erhärtungsverhalten auswirken.
Um den Einfluss verschiedener Faktoren gleichzeitig zu betrachten, werden mit Hilfe von statistischen Versuchsplänen Modelle erstellt, die das Verhalten der einzelnen Faktoren beschreiben. Zur Beschreibung der Fließeigenschaften wurde das Setzfließ- und Ausbreitmaß von Zementsuspensionen herangezogen. Die Beschleunigung der Erhärtung wurde mit Hilfe der Ermittlung des Zeitpunkts des normalen Erstarrens der Zementsuspension bestimmt.
Die Ergebnisse dieser Untersuchungen zeigen deutlich, dass die Fließeigenschaften und der Erstarrungsbeginn nicht linear mit steigendem Ultraschall-Energieeintrag verändert werden. Es zeigt sich, dass es besonders bei den Verarbeitungseigenschaften der Portlandzementsuspensionen zur Ausbildung eines spezifischen Energieeintrages kommt, bis zu welchem das Setzfließ- und das Ausbreitmaß erhöht werden. Bei Überschreiten dieses Punktes, der als kritischer Energieeintrag definiert wurde, nimmt das Setzfließ- und Ausbreitmaß wieder ab. Das Auftreten dieses Punktes ist im besonderen Maße abhängig vom w/z-Wert. Mit sinkendem w/z-Wert wird der Energieeintrag, der eine Verbesserung der Fließeigenschaften hervorruft, reduziert. Bei sehr niedrigen w/z-Werten (< 0,35), kann keine Verbesserung mehr beobachtet werden.
Wird Fließmittel vor der Beschallung zur Zementsuspension zugegeben, können die Eigenschaften der Zementsuspension maßgeblich beeinflusst werden. In beschallten Suspensionen mit Fließmittel, konnte in Abhängigkeit des Energieeintrages die fließmittelbedingte Verzögerung des Erstarrungsbeginns deutlich reduziert werden. Weiterhin zeigt sich, dass der Energieeintrag, der notwendig ist um den Erstarrungsbeginn um einen festen Betrag zu reduzieren, bei Suspensionen mit Fließmittel deutlich reduziert ist.
Auf Grundlage der Beobachtungen an Zementsuspensionen wird der Einfluss von Ultraschall in einen dispergierenden und einen beschleunigenden Effekt unterteilt. Bei hohen w/z-Werten dominiert der dispergierende Einfluss von Ultraschall und der Erstarrungsbeginn wird moderat verkürzt. Bei niedrigeren w/z-Werten der Zementsuspension, dominiert der beschleunigende Effekt wobei kein oder sogar ein negativer Einfluss auf die Verarbeitungseigenschaften beobachtet werden kann.
Im nächsten Schritt werden die Untersuchungen auf den Betonmaßstab mit Hilfe einer Technikumsanlage erweitert und der Einfluss eines zweistufigen Mischens (also dem Herstellen einer Zementsuspension im ersten Schritt und dem darauffolgenden Vermischen mit der Gesteinskörnung im zweiten Schritt) mit Ultraschall auf die Frisch- und Festbetoneigenschaften betrachtet. Durch die Anlagentechnik, die mit der Beschallung größerer Mengen Zementsuspension einhergeht, kommen weitere Einflussfaktoren auf die Zementsuspension hinzu (z. B. Pumpgeschwindigkeit, Temperatur, Druck). Im Rahmen der Untersuchungen wurde eine Betonrezeptur mit und ohne Ultraschall hergestellt und die Frisch- und Festbetoneigenschaften verglichen. Darüber hinaus wurde ein umfangreiches Untersuchungsprogramm zur Ermittlung wesentlicher Dauerhaftigkeitsparameter durchgeführt. Aufbauend auf den Erfahrungen mit der Technikumsanlage wurde das Ultraschall-Vormischsystem in mehreren Stufen weiterentwickelt und abschließend in einem Betonwerk zur Betonproduktion verwendet.
Die Untersuchungen am Beton zeigen eine deutliche Steigerung der Frühdruckfestigkeiten des Portlandzementbetons. Hierbei kann die zum Entschalen von Betonbauteilen notwendige Druckfestigkeit von 15 MPa deutlich früher erreicht werden. Das Ausbreitmaß der Betone (w/z-Wert = 0,47) wird infolge der Beschallung leicht reduziert, was sich mit den Ergebnissen aus den Untersuchungen an reinen Zementsuspensionen deckt. Bei Applikation eines Überdruckes in der Beschallkammer oder einer Kühlung der Suspension während der Beschallung, kann das Ausbreitmaß leicht gesteigert werden. Allerdings werden die hohen Frühdruckfestigkeiten der ungekühlten beziehungsweise drucklosen Variante nicht mehr erreicht.
In den Untersuchungen kann gezeigt werden, dass das Potential durch die Ultraschall-Beschleunigung genutzt werden kann, um entweder die Festigkeitsklasse des Zementes leitungsneutral zu reduzieren (von CEM I 52,5 R auf CEM I 42,5 R) oder eine 4-stündige Wärmebehandlung vollständig zu substituieren. Die Dauerhaftigkeit der Betone wird dabei nicht negativ beeinflusst. In den Untersuchungen zum Sulfat-, Karbonatisierung-, Chlorideindring- oder Frost/Tauwiderstand kann weder ein positiver noch ein negativer Einfluss durch die Beschallung abgeleitet werden. Ebenso kann in einer Untersuchung zur Alkali-Kieselsäure-Reaktion kein negativer Einfluss durch die Ultraschallbehandlung beobachtet werden.
In den darauf aufbauenden Untersuchungen wird die Anlagentechnik weiterentwickelt, um die Ultraschallbehandlung stärker an eine reale Betonproduktion anzupassen. In der ersten Iterationsstufe wird das in den Betonuntersuchungen verwendete Anlagenkonzept 1 modifiziert (von der In-line-Beschallung zur Batch-Beschallung) und als Analgenkonzept 2 für weitere Untersuchungen genutzt. Hierbei wird eine neue Betonrezeptur mit höherem w/z-Wert (0,52) verwendet, wobei die Druckfestigkeiten ebenfalls deutlich gesteigert werden können. Im Gegensatz zum ersten Beton, wird das Ausbreitmaß dieser Betonzusammensetzung gesteigert, was zur Reduktion von Fließmittel genutzt wird. Dies deckt sich ebenfalls mit den Beobachtungen an reinen Portlandzementsuspensionen, wo eine deutliche Verbesserung der Fließfähigkeit bei höheren w/z-Werten beschrieben wird.
Für diese Betonrezeptur wird ein Vergleich mit einem kommerziell erhältlichen Erhärtungsbeschleuniger (synthetische C-S-H-Keime) angestellt. Hierbei zeigt sich, dass die Beschleunigungswirkung beider Technologien vergleichbar ist. Eine Kombination beider Technologien führt zu einer weiteren deutlichen Steigerung der Frühfestigkeiten, so dass hier von einem synergistischen Effekt ausgegangen werden kann.
In der letzten Iterationsstufe, dem Anlagenkonzept 3, wird beschrieben, wie das Mischsystem im Rahmen einer universitären Ausgründung signifikant weiterentwickelt wird und erstmals in einem Betonwerk zur Betonproduktion verwendet wird. Bei den Überlegungen zur Weiterentwicklung des Ultraschall-Mischsystems wird der Fokus auf die Praktikabilität gelegt und gezeigt, dass das ultraschallgestütze Mischsystem die Druckfestigkeitsentwicklung auch im Werksmaßstab deutlich beschleunigen kann. Damit ist die Voraussetzung für eine ökologisch nachhaltige Optimierung eines Fertigteilbetons unter realen Produktionsbedingungen geschaffen worden.
Inhaltlich beschäftigt sich die Arbeit, die im Rahmen des Promotionsstudiengangs Kunst und Gestaltung an der Bauhaus-Universität entstand, mit der Erforschung sozio-interaktiver Potentiale der Videotelefonie im Kontext von Nähe und Verbundenheit mit Fokus auf Eigenbild, Embodiment sowie den Rederechtswechsel.
Die Videotelefonie als Kommunikationsform hat sich – und darauf deuten die Erfahrungen der Co- vid-19-Pandemie hin – im lebensweltlichen Alltag der Menschen etabliert und wird dort in naher Zukunft nicht mehr wegzudenken sein. Auf Basis ihrer Möglichkeiten und Errungenschaften ist es inzwischen Realität und Lebenswirklichkeit, dass die Kommunikation sowohl im privaten als auch im geschäftlichen Kontext mittels verschiedenster Kanäle stattfindet. Der Videotelefonie kommt hierbei als solche nicht nur eine tragende Funktion, sondern auch eine herausragende Rolle bei der vermeintlichen Reproduktion der Face-to-Face-Kommunikation im digitalen Raum zu und wird wie selbstverständlich zum zwischenmenschlichen Austausch genutzt. Just an diesem Punkt knüpft die Forschungsarbeit an. Zentral stand dabei das Vorhaben einer dezidierte Untersuchung des Forschungsgegenstandes Videotelefonie, sowohl aus Kultur- als auch Technikhistorischer, aber auch Medien-, Wahrnehmungs- wie Kommunikations- theoretischer Perspektive, indem analytische und phänosemiotische Perspektiven miteinander in Beziehung gesetzt werden (z.B. Wahrnehmungsbedingungen, Interaktionsmerkmale, realisierte Kommunikationsprozesse etc.). Damit verbundenes, wünschenswertes Ziel war es, eine möglichst zeitgemäße wie relevante Forschungsfrage zu adressieren, die neben den kulturellen Technisierungs- und Mediatisierungstendenzen in institutionellen und privaten Milieus ebenfalls eine conditio sine qua non der pandemischen (Massen-)Kommunikation entwirft.
Die Arbeit ist damit vor allem im Bereich des Produkt- und Interactiondesigns zu verorten. Darüber hinaus hatte sie das Ziel der Darlegung und Begründung der Videotelefonie als eigenständige Kommunikationsform, welche durch eigene, kommunikative Besonderheiten, die sich in ihrer jeweiligen Ingebrauchnahme sowie durch spezielle Wahrnehmungsbedingungen äußern, und die die Videotelefonie als »Rederechtswechselmedium« avant la lettre konsolidieren, gekennzeichnet ist. Dabei sollte der Beweis erbracht werden, dass die Videotelefonie nicht als Schwundstufe einer Kommunikation Face-to-Face, sondern als ein eigenständiges Mediatisierungs- und Kommunikationsereignis zu verstehen sei. Und eben nicht als eine beliebige – sich linear vom Telefon ausgehende – entwickelte Form der audio-visuellen Fernkommunikation darstellt, sondern die gestalterische (Bewegtbild-)Technizität ein eigenständiges Funktionsmaß offeriert, welches wiederum ein innovatives Kommunikationsmilieu im Kontext einer Rederechtswechsel-Medialität stabilisiert.
The reduction of the cement clinker content is an important prerequisite for the improvement of the CO2-footprint of concrete. Nevertheless, the durability of such concretes must be sufficient to guarantee a satisfactory service life of structures. Salt frost scaling resistance is a critical factor in this regard, as it is often diminished at increased clinker substitution rates. Furthermore, only insufficient long-term experience for such concretes exists. A high salt frost scaling resistance thus cannot be achieved by applying only descriptive criteria, such as the concrete composition. It is therefore to be expected, that in the long term a performance based service life prediction will replace the descriptive concept.
To achieve the important goal of clinker reduction for concretes also in cold and temperate climates it is important to understand the underlying mechanisms for salt frost scaling. However, conflicting damage theories dominate the current State of the Art. It was consequently derived as the goal of this thesis to evaluate existing damage theories and to examine them experimentally. It was found that only two theories have the potential to describe the salt frost attack satisfactorily – the glue spall theory and the cryogenic suction theory.
The glue spall theory attributes the surface scaling to the interaction of an external ice layer with the concrete surface. Only when moderate amounts of deicing salt are present in the test solution the resulting mechanical properties of the ice can cause scaling. However, the results in this thesis indicate that severe scaling also occurs at deicing salt levels, at which the ice is much too soft to damage concrete. Thus, the inability of the glue spall theory to account for all aspects of salt frost scaling was shown.
The cryogenic suction theory is based on the eutectic behavior of salt solutions, which consist of two phases – water ice and liquid brine – between the freezing point and the eutectic temperature. The liquid brine acts as an additional moisture reservoir, which facilitates the growth of ice lenses in the surface layer of the concrete. The experiments in this thesis confirmed, that the ice formation in hardened cement paste increases due to the suction of brine at sub-zero temperatures. The extent of additional ice formation was influenced mainly by the porosity and by the chloride binding capacity of the hardened cement paste.
Consequently, the cryogenic suction theory plausibly describes the actual generation of scaling, but it has to be expanded by some crucial aspects to represent the salt frost scaling attack completely. The most important aspect is the intensive saturation process, which is ascribed to the so-called micro ice lens pump. Therefore a combined damage theory was proposed, which considers multiple saturation processes. Important aspects of this combined theory were confirmed experimentally.
As a result, the combined damage theory constitutes a good basis to understand the salt frost scaling attack on concrete on a fundamental level. Furthermore, a new approach was identified, to account for the reduced salt frost scaling resistance of concretes with reduced clinker content.
For the safe and efficient operation of dams, frequent monitoring and maintenance are required. These are usually expensive, time consuming, and cumbersome. To alleviate these issues, we propose applying a wave-based scheme for the location and quantification of damages in dams.
To obtain high-resolution “interpretable” images of the damaged regions, we drew inspiration from non-linear full-multigrid methods for inverse problems and applied a new cyclic multi-stage full-waveform inversion (FWI) scheme. Our approach is less susceptible to the stability issues faced by the standard FWI scheme when dealing with ill-posed problems. In this paper, we first selected an optimal acquisition setup and then applied synthetic data to demonstrate the capability of our approach in identifying a series of anomalies in dams by a mixture of reflection and transmission tomography. The results had sufficient robustness, showing the prospects of application in the field of non-destructive testing of dams.
A safe and economic structural design based on the semi-probabilistic concept requires statistically representative safety elements, such as characteristic values, design values, and partial safety factors. Regarding climate loads, the safety levels of current design codes strongly reflect experiences based on former measurements and investigations assuming stationary conditions, i.e. involving constant frequencies and intensities. However, due to climate change, occurrence of corresponding extreme weather events is expected to alter in the future influencing the reliability and safety of structures and their components. Based on established approaches, a systematically refined data-driven methodology for the determination of design parameters considering nonstationarity as well as standardized targets of structural reliability or safety, respectively, is therefore proposed. The presented procedure picks up fundamentals of European standardization and extends them with respect to nonstationarity by applying a shifting time window method. Taking projected snow loads into account, the application of the method is exemplarily demonstrated and various influencing parameters are discussed.
Finite Element Simulations of dynamically excited structures are mainly influenced by the mass, stiffness, and damping properties of the system, as well as external loads. The prediction quality of dynamic simulations of vibration-sensitive components depends significantly on the use of appropriate damping models. Damping phenomena have a decisive influence on the vibration amplitude and the frequencies of the vibrating structure. However, developing realistic damping models is challenging due to the multiple sources that cause energy dissipation, such as material damping, different types of friction, or various interactions with the environment.
This thesis focuses on thermoelastic damping, which is the main cause of material damping in homogeneous materials. The effect is caused by temperature changes due to mechanical strains. In vibrating structures, temperature gradients arise in adjacent tension and compression areas. Depending on the vibration frequency, they result in heat flows, leading to increased entropy and the irreversible transformation of mechanical energy into thermal energy.
The central objective of this thesis is the development of efficient simulation methods to incorporate thermoelastic damping in finite element analyses based on modal superposition. The thermoelastic loss factor is derived from the structure's mechanical mode shapes and eigenfrequencies. In subsequent analyses that are performed in the time and frequency domain, it is applied as modal damping.
Two approaches are developed to determine the thermoelastic loss in thin-walled plate structures, as well as three-dimensional solid structures. The realistic representation of the dissipation effects is verified by comparing the simulation results with experimentally determined data. Therefore, an experimental setup is developed to measure material damping, excluding other sources of energy dissipation.
The three-dimensional solid approach is based on the determination of the generated entropy and therefore the generated heat per vibration cycle, which is a measure for thermoelastic loss in relation to the total strain energy. For thin plate structures, the amount of bending energy in a modal deformation is calculated and summarized in the so-called Modal Bending Factor (MBF). The highest amount of thermoelastic loss occurs in the state of pure bending. Therefore, the MBF enables a quantitative classification of the mode shapes concerning the thermoelastic damping potential.
The results of the developed simulations are in good agreement with the experimental results and are appropriate to predict thermoelastic loss factors. Both approaches are based on modal superposition with the advantage of a high computational efficiency. Overall, the modeling of thermoelastic damping represents an important component in a comprehensive damping model, which is necessary to perform realistic simulations of vibration processes.
Design-related reassessment of structures integrating Bayesian updating of model safety factors
(2022)
In the semi-probabilistic approach of structural design, the partial safety factors are defined by considering some degree of uncertainties to actions and resistance, associated with the parameters’ stochastic nature. However, uncertainties for individual structures can be better examined by incorporating measurement data provided by sensors from an installed health monitoring scheme. In this context, the current study proposes an approach to revise the partial safety factor for existing structures on the action side, γE by integrating Bayesian model updating. A simple numerical example of a beam-like structure with artificially generated measurement data is used such that the influence of different sensor setups and data uncertainties on revising the safety factors can be investigated. It is revealed that the health monitoring system can reassess the current capacity reserve of the structure by updating the design safety factors, resulting in a better life cycle assessment of structures. The outcome is furthermore verified by analysing a real life small railway steel bridge ensuring the applicability of the proposed method to practical applications.
Bolted connections are widely employed in structures like transmission poles, wind turbines, and television (TV) towers. The behaviour of bolted connections is often complex and plays a significant role in the overall dynamic characteristics of the structure. The goal of this work is to conduct a fatigue lifecycle assessment of such a bolted connection block of a 193 m tall TV tower, for which 205 days of real measurement data have been obtained from the installed monitoring devices. Based on the recorded data, the best-fit stochastic wind distribution for 50 years, the decisive wind action, and the locations to carry out the fatigue analysis have been decided. A 3D beam model of the entire tower is developed to extract the nodal forces corresponding to the connection block location under various mean wind speeds, which is later coupled with a detailed complex finite element model of the connection block, with over three million degrees of freedom, for acquiring stress histories on some pre-selected bolts. The random stress histories are analysed using the rainflow counting algorithm (RCA) and the damage is estimated using Palmgren-Miner's damage accumulation law. A modification is proposed to integrate the loading sequence effect into the RCA, which otherwise is ignored, and the differences between the two RCAs are investigated in terms of the accumulated damage.
Material failure can be tackled by so-called nonlocal models, which introduce an intrinsic length scale into the formulation and, in the case of material failure, restore the well-posedness of the underlying boundary value problem or initial boundary value problem. Among nonlocal models, peridynamics (PD) has attracted a lot of attention as it allows the natural transition from continuum to discontinue and thus allows modeling of discrete cracks without the need to describe and track the crack topology, which has been a major obstacle in traditional discrete crack approaches. This is achieved by replacing the divergence of the Cauchy stress tensor through an integral over so-called bond forces, which account for the interaction of particles. A quasi-continuum approach is then used to calibrate the material parameters of the bond forces, i.e., equating the PD energy with the energy of a continuum. One major issue for the application of PD to general complex problems is that they are limited to fairly simple material behavior and pure mechanical problems based on explicit time integration. PD has been extended to other applications but losing simultaneously its simplicity and ease in modeling material failure. Furthermore, conventional PD suffers from instability and hourglass modes that require stabilization. It also requires the use of constant horizon sizes, which drastically reduces its computational efficiency. The latter issue was resolved by the so-called dual-horizon peridynamics (DH-PD) formulation and the introduction of the duality of horizons.
Within the nonlocal operator method (NOM), the concept of nonlocality is further extended and can be considered a generalization of DH-PD. Combined with the energy functionals of various physical models, the nonlocal forms based on the dual-support concept can be derived. In addition, the variation of the energy functional allows implicit formulations of the nonlocal theory. While traditional integral equations are formulated in an integral domain, the dual-support approaches are based on dual integral domains. One prominent feature of NOM is its compatibility with variational and weighted residual methods. The NOM yields a direct numerical implementation based on the weighted residual method for many physical problems without the need for shape functions. Only the definition of the energy or boundary value problem is needed to drastically facilitate the implementation. The nonlocal operator plays an equivalent role to the derivatives of the shape functions in meshless methods and finite element methods (FEM). Based on the variational principle, the residual and the tangent stiffness matrix can be obtained with ease by a series of matrix multiplications. In addition, NOM can be used to derive many nonlocal models in strong form.
The principal contributions of this dissertation are the implementation and application of NOM, and also the development of approaches for dealing with fractures within the NOM, mostly for dynamic fractures. The primary coverage and results of the dissertation are as follows:
-The first/higher-order implicit NOM and explicit NOM, including a detailed description of the implementation, are presented. The NOM is based on so-called support, dual-support, nonlocal operators, and an operate energy functional ensuring stability. The nonlocal operator is a generalization of the conventional differential operators. Combining with the method of weighted residuals and variational principles, NOM establishes the residual and tangent stiffness matrix of operate energy functional through some simple matrix without the need of shape functions as in other classical computational methods such as FEM. NOM only requires the definition of the energy drastically simplifying its implementation. For the sake of conciseness, the implementation in this chapter is focused on linear elastic solids only, though the NOM can handle more complex nonlinear problems. An explicit nonlocal operator method for the dynamic analysis of elasticity solid problems is also presented. The explicit NOM avoids the calculation of the tangent stiffness matrix as in the implicit NOM model. The explicit scheme comprises the Verlet-velocity algorithm. The NOM can be very flexible and efficient for solving partial differential equations (PDEs). It's also quite easy for readers to use the NOM and extend it to solve other complicated physical phenomena described by one or a set of PDEs. Several numerical examples are presented to show the capabilities of this method.
-A nonlocal operator method for the dynamic analysis of (thin) Kirchhoff plates is proposed. The nonlocal Hessian operator is derived from a second-order Taylor series expansion. NOM is higher-order continuous, which is exploited for thin plate analysis that requires $C^1$ continuity. The nonlocal dynamic governing formulation and operator energy functional for Kirchhoff plates are derived from a variational principle. The Verlet-velocity algorithm is used for time discretization. After confirming the accuracy of the nonlocal Hessian operator, several numerical examples are simulated by the nonlocal dynamic Kirchhoff plate formulation.
-A nonlocal fracture modeling is developed and applied to the simulation of quasi-static and dynamic fractures using the NOM. The phase field's nonlocal weak and associated strong forms are derived from a variational principle. The NOM requires only the definition of energy. We present both a nonlocal implicit phase field model and a nonlocal explicit phase field model for fracture; the first approach is better suited for quasi-static fracture problems, while the key application of the latter one is dynamic fracture. To demonstrate the performance of the underlying approach, several benchmark examples for quasi-static and dynamic fracture are solved.
The aim of this study is controlling of spurious oscillations developing around discontinuous solutions of both linear and non-linear wave equations or hyperbolic partial differential equations (PDEs). The equations include both first-order and second-order (wave) hyperbolic systems. In these systems even smooth initial conditions, or smoothly varying source (load) terms could lead to discontinuous propagating solutions (fronts). For the first order hyperbolic PDEs, the concept of central high resolution schemes is integrated with the multiresolution-based adaptation to capture properly both discontinuous propagating fronts and effects of fine-scale responses on those of larger scales in the multiscale manner. This integration leads to using central high resolution schemes on non-uniform grids; however, such simulation is unstable, as the central schemes are originally developed to work properly on uniform cells/grids. Hence, the main concern is stable collaboration of central schemes and multiresoltion-based cell adapters. Regarding central schemes, the considered approaches are: 1) Second order central and central-upwind schemes; 2) Third order central schemes; 3) Third and fourth order central weighted non-oscillatory schemes (central-WENO or CWENO); 4) Piece-wise parabolic methods (PPMs) obtained with two different local stencils. For these methods, corresponding (nonlinear) stability conditions are studied and modified, as well. Based on these stability conditions several limiters are modified/developed as follows: 1) Several second-order limiters with total variation diminishing (TVD) feature, 2) Second-order uniformly high order accurate non-oscillatory (UNO) limiters, 3) Two third-order nonlinear scaling limiters, 4) Two new limiters for PPMs. Numerical results show that adaptive solvers lead to cost-effective computations (e.g., in some 1-D problems, number of adapted grid points are less than 200 points during simulations, while in the uniform-grid case, to have the same accuracy, using of 2049 points is essential). Also, in some cases, it is confirmed that fine scale responses have considerable effects on higher scales.
In numerical simulation of nonlinear first order hyperbolic systems, the two main concerns are: convergence and uniqueness. The former is important due to developing of the spurious oscillations, the numerical dispersion and the numerical dissipation. Convergence in a numerical solution does not guarantee that it is the physical/real one (the uniqueness feature). Indeed, a nonlinear systems can converge to several numerical results (which mathematically all of them are true). In this work, the convergence and uniqueness are directly studied on non-uniform grids/cells by the concepts of local numerical truncation error and numerical entropy production, respectively. Also, both of these concepts have been used for cell/grid adaptations. So, the performance of these concepts is also compared by the multiresolution-based method. Several 1-D and 2-D numerical examples are examined to confirm the efficiency of the adaptive solver. Examples involve problems with convex and non-convex fluxes. In the latter case, due to developing of complex waves, proper capturing of real answers needs more attention. For this purpose, using of method-adaptation seems to be essential (in parallel to the cell/grid adaptation). This new type of adaptation is also performed in the framework of the multiresolution analysis.
Regarding second order hyperbolic PDEs (mechanical waves), the regularization concept is used to cure artificial (numerical) oscillation effects, especially for high-gradient or discontinuous solutions. There, oscillations are removed by the regularization concept acting as a post-processor. Simulations will be performed directly on the second-order form of wave equations. It should be mentioned that it is possible to rewrite second order wave equations as a system of first-order waves, and then simulated the new system by high resolution schemes. However, this approach ends to increasing of variable numbers (especially for 3D problems).
The numerical discretization is performed by the compact finite difference (FD) formulation with desire feature; e.g., methods with spectral-like or optimized-error properties. These FD methods are developed to handle high frequency waves (such as waves near earthquake sources). The performance of several regularization approaches is studied (both theoretically and numerically); at last, a proper regularization approach controlling the Gibbs phenomenon is recommended.
At the end, some numerical results are provided to confirm efficiency of numerical solvers enhanced by the regularization concept. In this part, shock-like responses due to local and abrupt changing of physical properties, and also stress wave propagation in stochastic-like domains are studied.
In this work, the degradation performance for the photocatalytic oxidation of eight micropollutants (amisulpride, benzotriazole, candesartan, carbamazepine, diclofenac, gabapentin, methlybenzotriazole, and metoprolol) within real secondary effluent was investigated using three different reactor designs. For all reactor types, the influence of irradiation power on its reaction rate and energetic efficiency was investigated. Flat cell and batch reactor showed almost similar substance specific degradation behavior. Within the immersion rotary body reactor, benzotriazole and methylbenzotriazole showed a significantly lower degradation affinity. The flat cell reactor achieved the highest mean degradation rate, with half time values ranging from 5 to 64 min with a mean of 18 min, due to its high catalysts surface to hydraulic volume ratio. The EE/O values were calculated for all micro-pollutants as well as the mean degradation rate constant of each experimental step. The lowest substance specific energy per order (EE/O) values of 5 kWh/m3 were measured for benzotriazole within the batch reactor. The batch reactor also reached the lowest mean values (11.8–15.9 kWh/m3) followed by the flat cell reactor (21.0–37.0 kWh/m3) and immersion rotary body reactor (23.9–41.0 kWh/m3). Catalyst arrangement and irradiation power were identified as major influences on the energetic performance of the reactors. Low radiation intensities as well as the use of submerged catalyst arrangement allowed a reduction in energy demand by a factor of 3–4. A treatment according to existing treatment goals of wastewater treatment plants (80% total degradation) was achieved using the batch reactor with a calculated energy demand of 7000 Wh/m3.
The Finite Element Method (FEM) is widely used in engineering for solving Partial Differential Equations (PDEs) over complex geometries. To this end, it is required to provide the FEM software with a geometric model that is typically constructed in a Computer-Aided Design (CAD) software. However, FEM and CAD use different approaches for the mathematical description of the geometry. Thus, it is required to generate a mesh, which is suitable for FEM, based on the CAD model. Nonetheless, this procedure is not a trivial task and it can be time consuming. This issue becomes more significant for solving shape and topology optimization problems, which consist in evolving the geometry iteratively. Therefore, the computational cost associated to the mesh generation process is increased exponentially for this type of applications.
The main goal of this work is to investigate the integration of CAD and CAE in shape and topology optimization. To this end, numerical tools that close the gap between design and analysis are presented. The specific objectives of this work are listed below:
• Automatize the sensitivity analysis in an isogeometric framework for applications in shape optimization. Applications for linear elasticity are considered.
• A methodology is developed for providing a direct link between the CAD model and the analysis mesh. In consequence, the sensitivity analysis can be performed in terms of the design variables located in the design model.
• The last objective is to develop an isogeometric method for shape and topological optimization. This method should take advantage of using Non-Uniform Rational B-Splines (NURBS) with higher continuity as basis functions.
Isogeometric Analysis (IGA) is a framework designed to integrate the design and analysis in engineering problems. The fundamental idea of IGA is to use the same basis functions for modeling the geometry, usually NURBS, for the approximation of the solution fields. The advantage of integrating design and analysis is two-fold. First, the analysis stage is more accurate since the system of PDEs is not solved using an approximated geometry, but the exact CAD model. Moreover, providing a direct link between the design and analysis discretizations makes possible the implementation of efficient sensitivity analysis methods. Second, the computational time is significantly reduced because the mesh generation process can be avoided.
Sensitivity analysis is essential for solving optimization problems when gradient-based optimization algorithms are employed. Automatic differentiation can compute exact gradients, automatically by tracking the algebraic operations performed on the design variables. For the automation of the sensitivity analysis, an isogeometric framework is used. Here, the analysis mesh is obtained after carrying out successive refinements, while retaining the coarse geometry for the domain design. An automatic differentiation (AD) toolbox is used to perform the sensitivity analysis. The AD toolbox takes the code for computing the objective and constraint functions as input. Then, using a source code transformation approach, it outputs a code for computing the objective and constraint functions, and their sensitivities as well. The sensitivities obtained from the sensitivity propagation method are compared with analytical sensitivities, which are computed using a full isogeometric approach.
The computational efficiency of AD is comparable to that of analytical sensitivities. However, the memory requirements are larger for AD. Therefore, AD is preferable if the memory requirements are satisfied. Automatic sensitivity analysis demonstrates its practicality since it simplifies the work of engineers and designers.
Complex geometries with sharp edges and/or holes cannot easily be described with NURBS. One solution is the use of unstructured meshes. Simplex-elements (triangles and tetrahedra for two and three dimensions respectively) are particularly useful since they can automatically parameterize a wide variety of domains. In this regard, unstructured Bézier elements, commonly used in CAD, can be employed for the exact modelling of CAD boundary representations. In two dimensions, the domain enclosed by NURBS curves is parameterized with Bézier triangles. To describe exactly the boundary of a two-dimensional CAD model, the continuity of a NURBS boundary representation is reduced to C^0. Then, the control points are used to generate a triangulation such that the boundary of the domain is identical to the initial CAD boundary representation. Thus, a direct link between the design and analysis discretizations is provided and the sensitivities can be propagated to the design domain.
In three dimensions, the initial CAD boundary representation is given as a collection of NURBS surfaces that enclose a volume. Using a mesh generator (Gmsh), a tetrahedral mesh is obtained. The original surface is reconstructed by modifying the location of the control points of the tetrahedral mesh using Bézier tetrahedral elements and a point inversion algorithm. This method offers the possibility of computing the sensitivity analysis using the analysis mesh. Then, the sensitivities can be propagated into the design discretization. To reuse the mesh originally generated, a moving Bézier tetrahedral mesh approach was implemented.
A gradient-based optimization algorithm is employed together with a sensitivity propagation procedure for the shape optimization cases. The proposed shape optimization approaches are used to solve some standard benchmark problems in structural mechanics. The results obtained show that the proposed approach can compute accurate gradients and evolve the geometry towards optimal solutions. In three dimensions, the moving mesh approach results in faster convergence in terms of computational time and avoids remeshing at each optimization step.
For considering topological changes in a CAD-based framework, an isogeometric phase-field based shape and topology optimization is developed. In this case, the diffuse interface of a phase-field variable over a design domain implicitly describes the boundaries of the geometry. The design variables are the local values of the phase-field variable. The descent direction to minimize the objective function is found by using the sensitivities of the objective function with respect to the design variables. The evolution of the phase-field is determined by solving the time dependent Allen-Cahn equation.
Especially for topology optimization problems that require C^1 continuity, such as for flexoelectric structures, the isogeometric phase field method is of great advantage. NURBS can achieve the desired continuity more efficiently than the traditional employed functions. The robustness of the method is demonstrated when applied to different geometries, boundary conditions, and material configurations. The applications illustrate that compared to piezoelectricity, the electrical performance of flexoelectric microbeams is larger under bending. In contrast, the electrical power for a structure under compression becomes larger with piezoelectricity.
Quantification of cracks in concrete thin sections considering current methods of image analysis
(2022)
Image analysis is used in this work to quantify cracks in concrete thin sections via modern image processing. Thin sections were impregnated with a yellow epoxy resin, to increase the contrast between voids and other phases of the concrete. By the means of different steps of pre-processing, machine learning and python scripts, cracks can be quantified in an area of up to 40 cm2. As a result, the crack area, lengths and widths were estimated automatically within a single workflow. Crack patterns caused by freeze-thaw damages were investigated. To compare the inner degradation of the investigated thin sections, the crack density was used. Cracks in the thin sections were measured manually in two different ways for validation of the automatic determined results. On the one hand, the presented work shows that the width of cracks can be determined pixelwise, thus providing the plot of a width distribution. On the other hand, the automatically measured crack length differs in comparison to the manually measured ones.
In this work, we present a deep collocation method (DCM) for three-dimensional potential problems in non-homogeneous media. This approach utilizes a physics-informed neural network with material transfer learning reducing the solution of the non-homogeneous partial differential equations to an optimization problem. We tested different configurations of the physics-informed neural network including smooth activation functions, sampling methods for collocation points generation and combined optimizers. A material transfer learning technique is utilized for non-homogeneous media with different material gradations and parameters, which enhance the generality and robustness of the proposed method. In order to identify the most influential parameters of the network configuration, we carried out a global sensitivity analysis. Finally, we provide a convergence proof of our DCM. The approach is validated through several benchmark problems, also testing different material variations.
In machine learning, if the training data is independently and identically distributed as the test data then a trained model can make an accurate predictions for new samples of data. Conventional machine learning has a strong dependence on massive amounts of training data which are domain specific to understand their latent patterns. In contrast, Domain adaptation and Transfer learning methods are sub-fields within machine learning that are concerned with solving the inescapable problem of insufficient training data by relaxing the domain dependence hypothesis. In this contribution, this issue has been addressed and by making a novel combination of both the methods we develop a computationally efficient and practical algorithm to solve boundary value problems based on nonlinear partial differential equations. We adopt a meshfree analysis framework to integrate the prevailing geometric modelling techniques based on NURBS and present an enhanced deep collocation approach that also plays an important role in the accuracy of solutions. We start with a brief introduction on how these methods expand upon this framework. We observe an excellent agreement between these methods and have shown that how fine-tuning a pre-trained network to a specialized domain may lead to an outstanding performance compare to the existing ones. As proof of concept, we illustrate the performance of our proposed model on several benchmark problems.
BIM-basierte Digitalisierung von Bestandsgebäuden aus Sicht des FM am Beispiel von Heizungsanlagen
(2022)
Das Ziel der Arbeit ist, für das Facility Management relevante Informationen für die mit Building Information Modeling basierende Erstellung von Bestandsgebäuden am Beispiel einer Hei- zungsanlage zu definieren. Darauf basierend sind die notwendigen Arbeitsschritte der Objek- taufnahme abgeleitet. Für die Definition der Arbeitsschritte wurden das grundlegende Vorge- hen bei einer Objektaufnahme sowie die gesetzlichen Gegebenheiten für den Betrieb einer Heizungsanlage dargelegt. Darüber hinaus sind in der vorliegenden Ausarbeitung die Vorteile und Herausforderungen hinsichtlich des Zusammenspiels von Building Information Modeling und Facility Management analysiert. Die definierten Arbeitsschritte sind anhand eines Beispiel- projektes angewendet worden. Im Rahmen des Beispielprojekts sind die entscheidenden Be- triebsdaten je Anlagenteil in Form von Informationsanforderungen nach DIN 17412 definiert. Das Gebäudemodell ist durch Parameter mit den für das Facility Management relevanten In- formationen ergänzt. Die Resultate des Beispielprojektes sind mit aussagekräftigen Schnitten, Plänen sowie 3-D-Visualisierungen dargestellt. Abschließend sind die Ergebnisse in Bezug auf das FM validiert. Aus den Arbeitsschritten und Ergebnissen ist eine Leitlinie erstellt worden für den Digitalisierungsprozess von Bestandsgebäuden für das Facility Management.
Multi-criteria decision analysis (MCDA) is an established methodology to support the decision-making of multi-objective problems. For conducting an MCDA, in most cases, a set of objectives (SOO) is required, which consists of a hierarchical structure comprised of objectives, criteria, and indicators. The development of an SOO is usually based on moderated development processes requiring high organizational and cognitive effort from all stakeholders involved. This article proposes elementary interactions as a key paradigm of an algorithm-driven development process for an SOO that requires little moderation efforts. Elementary interactions are self-contained information requests that may be answered with little cognitive effort. The pairwise comparison of elements in the well-known analytical hierarchical process (AHP) is an example of an elementary interaction. Each elementary interaction in the development process presented contributes to the stepwise development of an SOO. Based on the hypothesis that an SOO may be developed exclusively using elementary interactions (EIs), a concept for a multi-user platform is proposed. Essential components of the platform are a Model Aggregator, an Elementary Interaction Stream Generator, a Participant Manager, and a Discussion Forum. While the latter component serves the professional exchange of the participants, the first three components are intended to be automatable by algorithms. The platform concept proposed has been evaluated partly in an explorative validation study demonstrating the general functionality of the algorithms outlined. In summary, the platform concept suggested demonstrates the potential to ease SOO development processes as the platform concept does not restrict the application domain; it is intended to work with little administration moderation efforts, and it supports the further development of an existing SOO in the event of changes in external conditions. The algorithm-driven development of SOOs proposed in this article may ease the development of MCDA applications and, thus, may have a positive effect on the spread of MCDA applications.
The seismic vulnerability assessment of existing reinforced concrete (RC) buildings is a significant source of disaster mitigation plans and rescue services. Different countries evolved various Rapid Visual Screening (RVS) techniques and methodologies to deal with the devastating consequences of earthquakes on the structural characteristics of buildings and human casualties. Artificial intelligence (AI) methods, such as machine learning (ML) algorithm-based methods, are increasingly used in various scientific and technical applications. The investigation toward using these techniques in civil engineering applications has shown encouraging results and reduced human intervention, including uncertainties and biased judgment. In this study, several known non-parametric algorithms are investigated toward RVS using a dataset employing different earthquakes. Moreover, the methodology encourages the possibility of examining the buildings’ vulnerability based on the factors related to the buildings’ importance and exposure. In addition, a web-based application built on Django is introduced. The interface is designed with the idea to ease the seismic vulnerability investigation in real-time. The concept was validated using two case studies, and the achieved results showed the proposed approach’s potential efficiency
In recent years, lightweight materials, such as polymer composite materials (PNCs) have been studied and developed due to their excellent physical and chemical properties. Structures composed of these composite materials are widely used in aerospace engineering structures, automotive components, and electrical devices. The excellent and outstanding mechanical, thermal, and electrical properties of Carbon nanotube (CNT) make it an ideal filler to strengthen polymer materials’ comparable properties. The heat transfer of composite materials has very promising engineering applications in many fields, especially in electronic devices and energy storage equipment. It is essential in high-energy density systems since electronic components need heat dissipation functionality. Or in other words, in electronic devices the generated heat should ideally be dissipated by light and small heat sinks.
Polymeric composites consist of fillers embedded in a polymer matrix, the first ones will significantly affect the overall (macroscopic) performance of the material. There are many common carbon-based fillers such as single-walled carbon nanotubes (SWCNT), multi-walled carbon nanotubes (MWCNT), carbon nanobuds (CNB), fullerene, and graphene. Additives inside the matrix have become a popular subject for researchers. Some extraordinary characters, such as high-performance load, lightweight design, excellent chemical resistance, easy processing, and heat transfer, make the design of polymeric nanotube composites (PNCs) flexible. Due to the reinforcing effects with different fillers on composite materials, it has a higher degree of freedom and can be designed for the structure according to specific applications’ needs. As already stated, our research focus will be on SWCNT enhanced PNCs. Since experiments are timeconsuming, sometimes expensive and cannot shed light into phenomena taking place for instance at the interfaces/interphases of composites, they are often complemented through theoretical and computational analysis.
While most studies are based on deterministic approaches, there is a comparatively lower number of stochastic methods accounting for uncertainties in the input parameters. In deterministic models, the output of the model is fully determined by the parameter values and the initial conditions. However, uncertainties in the input parameters such as aspect ratio, volume fraction, thermal properties of fiber and matrix need to be taken into account for reliable predictions. In this research, a stochastic multiscale method is provided to study the influence of numerous uncertain input parameters on the thermal conductivity of the composite. Therefore, a hierarchical multi-scale method based on computational homogenization is presented in to predict the macroscopic thermal conductivity based on the fine-scale structure. In order to study the inner mechanism, we use the finite element method and employ surrogate models to conduct a Global Sensitivity Analysis (GSA). The SA is performed in order to quantify the influence of the conductivity of the fiber, matrix, Kapitza resistance, volume fraction and aspect ratio on the macroscopic conductivity. Therefore, we compute first-order and total-effect sensitivity indices with different surrogate models.
As stochastic multiscale models are computational expensive, surrogate approaches are commonly exploited. With the emergence of high performance computing and artificial intelligence, machine learning has become a popular modeling tool for numerous applications. Machine learning (ML) is commonly used in regression and maps data through specific rules with algorithms to build input and output models. They are particularly useful for nonlinear input-output relationships when sufficient data is available. ML has also been used in the design of new materials and multiscale analysis. For instance, Artificial neural networks and integrated learning seem to be ideally for such a task. They can theoretically simulate any non-linear relationship through the connection of neurons. Mapping relationships are employed to carry out data-driven simulations of inputs and outputs in stochastic modeling.
This research aims to develop a stochastic multi-scale computational models of PNCs in heat transfer. Multi-scale stochastic modeling with uncertainty analysis and machine learning methods consist of the following components:
-Uncertainty Analysis. A surrogate based global sensitivity analysis is coupled with a hierarchical multi-scale method employing computational homogenization. The effect of the conductivity of the fibers and the matrix, the Kapitza resistance, volume fraction and aspect ratio on the ’macroscopic’ conductivity of the composite is systematically studied. All selected surrogate models yield consistently the conclusions that the most influential input parameters are the aspect ratio followed by the volume fraction. The Kapitza Resistance has no significant effect on the thermal conductivity of the PNCs. The most accurate surrogate model in terms of the R2 value is the moving least square (MLS).
-Hybrid Machine Learning Algorithms. A combination of artificial neural network (ANN) and particle swarm optimization (PSO) is applied to estimate the relationship between variable input and output parameters. The ANN is used for modeling the composite while PSO improves the prediction performance through an optimized global minimum search. The thermal conductivity of the fibers and the matrix, the kapitza resistance, volume fraction and aspect ratio are selected as input parameters. The output is the macroscopic (homogenized) thermal conductivity of the composite. The results show that the PSO significantly improves the predictive ability of this hybrid intelligent algorithm, which outperforms traditional neural networks.
-Stochastic Integrated Machine Learning. A stochastic integrated machine learning based multiscale approach for the prediction of the macroscopic thermal conductivity in PNCs is developed. Seven types of machine learning models are exploited in this research, namely Multivariate Adaptive Regression Splines (MARS), Support Vector Machine (SVM), Regression Tree (RT), Bagging Tree (Bag), Random Forest (RF), Gradient Boosting Machine (GBM) and Cubist. They are used as components of stochastic modeling to construct the relationship between the variable of the inputs’ uncertainty and the macroscopic thermal conductivity of PNCs. Particle Swarm Optimization (PSO) is used for hyper-parameter tuning to find the global optimal values leading to a significant reduction in the computational cost. The advantages and disadvantages of various methods are also analyzed in terms of computing time and model complexity to finally give a recommendation for the applicability of different models.
Hitze in der Stadt Jena
(2022)
Die vorliegende Arbeit befasst sich mit den spezifischen Faktoren und Wechselwirkungen des städtischen Klimas und Strategien zur Prävention und Kompensation lokaler Klimaveränderungen. Problematische Merkmale des Stadtklimas werden sich infolge des Klimawandels stärker ausprägen. Insbesondere die Hitzebelastung wird zunehmen und die Lebensbedingungen in der Stadt negativ beeinflussen. Infolge höherer Temperaturen in Städten und einer höheren Temperaturdifferenz zum Umland verändern sich Windströme und die Wasserbilanz. Es sind Strategien notwendig, um den Schadstoffausstoß, die Flächeninanspruchnahme, die Abfallproduktion und den Wasser-, Energie- und Ressourcenverbrauch zu verringern, um sowohl langfristig den Klimawandel als auch dessen bereits unvermeidbaren Auswirkungen auf Städte zu begrenzen.
Beispielhaft untersucht die Arbeit das Stadtklima, dessen zukünftige Veränderungen infolge des Klimawandels, bauliche Maßnahmen und Anpassungsstrategien der Stadt Jena. Jena ist die zweitgrößte Stadt im Bundesland Thüringen und gehört heute zu den wärmsten und trockensten Großstädten Deutschlands.
Die Ergebnisse der Arbeit werden anschließend anhand eines städtebaulichen Konzepts und Entwurfs angewendet. Das Bachstraßenareal liegt in der Innenstadt, dem am stärksten von Hitze betroffenen Stadtteil. Als ehemaliger Hauptstandort des Jenaer Universitätsklinikums, soll es zu einem nachhaltigen Wissenschaftscampus der Lebenswissenschaften umgebaut werden, wobei ein Großteil der denkmalgeschützten, ehemaligen Klinikgebäude erhalten bleibt. Der Fokus liegt dabei auf der Umsetzung der zuvor formulierten, nachhaltigen Strategien zur Verbesserung des lokalen Stadtklimas und einer Abschwächung der Auswirkungen des Klimawandels auf den besonders stark betroffenen Innenstadtbereich Jenas.
Atlas der Datenkörper. Körperbilder in Kunst, Design und Wissenschaft im Zeitalter digitaler Medien
(2022)
Digitale Technologien und soziale Medien verändern die Selbst- und Körperwahrnehmung und verzerren, verstärken oder produzieren dabei spezifische Körperbilder. Die Beiträger*innen kartographieren diese Phänomene, fragen nach ihrer medialen Existenzweise sowie nach den Möglichkeiten ihrer Kritik. Dabei begegnen sie ihrer Neuartigkeit mit einer transdisziplinären Herangehensweise. Aus sowohl der Perspektive künstlerischer und gestalterischer Forschung als auch der Kunst-, Kultur- und Medienwissenschaft sowie der Psychologie und Neurowissenschaft wird die Landschaft rezenter Körperbilder und Techniken einer digitalen Körperlichkeit untersucht.
Tropical coral reefs, one of the world’s oldest ecosystems which support some of the highest levels of biodiversity on the planet, are currently facing an unprecedented ecological crisis during this massive human-activity-induced period of extinction. Hence, tropical reefs symbolically stand for the destructive effects of human activities on nature [4], [5]. Artificial reefs are excellent examples of how architectural design can be combined with ecosystem regeneration [6], [7], [8]. However, to work at the interface between the artificial and the complex and temporal nature of natural systems presents a challenge, i.a. in respect to the B-rep modelling legacy of computational modelling.
The presented doctorate investigates strategies on how to apply digital practice to realise what is an essential bulwark to retain reefs in impossibly challenging times. Beyond the main question of integrating computational modelling and high precision monitoring strategies in artificial coral reef design, this doctorate explores techniques, methods, and linking frameworks to support future research and practice in ecology led design contexts.
Considering the many existing approaches for artificial coral reefs design, one finds they often fall short in precisely understanding the relationships between architectural and ecological aspects (e.g. how a surface design and material composition can foster coral larvae settlement, or structural three-dimensionality enhance biodiversity) and lack an integrated underwater (UW) monitoring process. Such a process is necessary in order to gather knowledge about the ecosystem and make it available for design, and to learn whether artificial structures contribute to reef regeneration or rather harm the coral reef ecosystem.
For the research, empirical experimental methods were applied: Algorithmic coral reef design, high precision UW monitoring, computational modelling and simulation, and validated through parallel real-world physical experimentation – two Artificial Reef Prototypes (ARPs) in Gili Trawangan, Indonesia (2012–today). Multiple discrete methods and sub techniques were developed in seventeen computational experiments and applied in a way in which many are cross valid and integrated in an overall framework that is offered as a significant contribution to the field. Other main contributions include the Ecosystem-aware design approach, Key Performance Indicators (KPIs) for coral reef design, algorithmic design and fabrication of Biorock cathodes, new high precision UW monitoring strategies, long-term real-world constructed experiments, new digital analysis methods and two new front-end web-based tools for reef design and monitoring reefs. The methodological framework is a finding of the research that has many technical components that were tested and combined in this way for the very first time.
In summary, the thesis responds to the urgency and relevance in preserving marine species in tropical reefs during this massive extinction period by offering a differentiated approach towards artificial coral reefs – demonstrating the feasibility of digitally designing such ‘living architecture’ according to multiple context and performance parameters. It also provides an in-depth critical discussion of computational design and architecture in the context of ecosystem regeneration and Planetary Thinking. In that respect, the thesis functions as both theoretical and practical background for computational design, ecology and marine conservation – not only to foster the design of artificial coral reefs technically but also to provide essential criteria and techniques for conceiving them.
Keywords: Artificial coral reefs, computational modelling, high precision underwater monitoring, ecology in design.
In recent years, the discussion of digitalization has arrived in the media, at conferences, and in committees of the construction and real estate industry. While some areas are producing innovations and some contributors can be described as pioneers, other topics still show deficits with regard to digital transformation. The building permit process can also be counted in this category. Regardless of how architects and engineers in planning offices rely on innovative methods, building documents have so far remained in paper form in too many cases, or are printed out after electronic submission to the authority. Existing resources – for example in the form of a building information model, which could provide support in the building permit process – are not being taken advantage of. In order to use digital tools to support decision-making by the building permit authorities, it is necessary to understand the current situation and to question conditions before pursuing the overall automation of internal authority processes as the sole solution.
With a substantive-organizational consideration of the relevant areas that influence building permit determination, an improvement of the building permit procedure within authorities is proposed. Complex areas – such as legal situations, the use of technology, as well as the subjective alternative action – are determined and structured. With the development of a model for the determination of building permitability, both an understanding of influencing factors is conveyed and an increase in transparency for all parties involved is created.
In addition to an international literature review, an empirical study served as the research method. The empirical study was conducted in the form of qualitative expert interviews in order to determine the current state in the field of building permit procedures. The collected data material was processed and subsequently subjected to a software-supported content analysis. The results were processed, in combination with findings from the literature review, in various analyses to form the basis for a proposed model.
The result of the study is a decision model that closes the gap between the current processes within the building authorities and an overall automation of the building permit review process. The model offers support to examiners and applicants in determining building permit eligibility, through its process-oriented structuring of decision-relevant facts. The theoretical model could be transferred into practice in the form of a web application.
The detailed structural analysis of thin-walled circular pipe members often requires the use of a shell or solid-based finite element method. Although these methods provide a very good approximation of the deformations, they require a higher degree of discretization which causes high computational costs. On the other hand, the analysis of thin-walled circular pipe members based on classical beam theories is easy to implement and needs much less computation time, however, they are limited in their ability to approximate the deformations as they cannot consider the deformation of the cross-section.
This dissertation focuses on the study of the Generalized Beam Theory (GBT) which is both accurate and efficient in analyzing thin-walled members. This theory is based on the separation of variables in which the displacement field is expressed as a combination of predetermined deformation modes related to the cross-section, and unknown amplitude functions defined on the beam's longitudinal axis. Although the GBT was initially developed for long straight members, through the consideration of complementary deformation modes, which amend the null transverse and shear membrane strain assumptions of the classical GBT, problems involving short members, pipe bends, and geometrical nonlinearity can also be analyzed using GBT. In this dissertation, the GBT formulation for the analysis of these problems is developed and the application and capabilities of the method are illustrated using several numerical examples. Furthermore, the displacement and stress field results of these examples are verified using an equivalent refined shell-based finite element model.
The developed static and dynamic GBT formulations for curved thin-walled circular pipes are based on the linear kinematic description of the curved shell theory. In these formulations, the complex problem in pipe bends due to the strong coupling effect of the longitudinal bending, warping and the cross-sectional ovalization is handled precisely through the derivation of the coupling tensors between the considered GBT deformation modes. Similarly, the geometrically nonlinear GBT analysis is formulated for thin-walled circular pipes based on the nonlinear membrane kinematic equations. Here, the initial linear and quadratic stress and displacement tangent stiffness matrices are built using the third and fourth-order GBT deformation mode coupling tensors.
Longitudinally, the formulation of the coupled GBT element stiffness and mass matrices are presented using a beam-based finite element formulation. Furthermore, the formulated GBT elements are tested for shear and membrane locking problems and the limitations of the formulations regarding the membrane locking problem are discussed.
Isogeometric analysis (IGA) is a numerical method for solving partial differential equations (PDEs), which was introduced with the aim of integrating finite element analysis with computer-aided design systems. The main idea of the method is to use the same spline basis functions which describe the geometry in CAD systems for the approximation of solution fields in the finite element method (FEM). Originally, NURBS which is a standard technology employed in CAD systems was adopted as basis functions in IGA but there were several variants of IGA using other technologies such as T-splines, PHT splines, and subdivision surfaces as basis functions. In general, IGA offers two key advantages over classical FEM: (i) by describing the CAD geometry exactly using smooth, high-order spline functions, the mesh generation process is simplified and the interoperability between CAD and FEM is improved, (ii) IGA can be viewed as a high-order finite element method which offers basis functions with high inter-element continuity and therefore can provide a primal variational formulation of high-order PDEs in a straightforward fashion. The main goal of this thesis is to further advance isogeometric analysis by exploiting these major advantages, namely precise geometric modeling and the use of smooth high-order splines as basis functions, and develop robust computational methods for problems with complex geometry and/or complex multi-physics.
As the first contribution of this thesis, we leverage the precise geometric modeling of isogeometric analysis and propose a new method for its coupling with meshfree discretizations. We exploit the strengths of both methods by using IGA to provide a smooth, geometrically-exact surface discretization of the problem domain boundary, while the Reproducing Kernel Particle Method (RKPM) discretization is used to provide the volumetric discretization of the domain interior. The coupling strategy is based upon the higher-order consistency or reproducing conditions that are directly imposed in the physical domain. The resulting coupled method enjoys several favorable features: (i) it preserves the geometric exactness of IGA, (ii) it circumvents the need for global volumetric parameterization of the problem domain, (iii) it achieves arbitrary-order approximation accuracy while preserving higher-order smoothness of the discretization. Several numerical examples are solved to show the optimal convergence properties of the coupled IGA–RKPM formulation, and to demonstrate its effectiveness in constructing volumetric discretizations for complex-geometry objects.
As for the next contribution, we exploit the use of smooth, high-order spline basis functions in IGA to solve high-order surface PDEs governing the morphological evolution of vesicles. These governing equations are often consisted of geometric PDEs, high-order PDEs on stationary or evolving surfaces, or a combination of them. We propose an isogeometric formulation for solving these PDEs. In the context of geometric PDEs, we consider phase-field approximations of mean curvature flow and Willmore flow problems and numerically study the convergence behavior of isogeometric analysis for these problems. As a model problem for high-order PDEs on stationary surfaces, we consider the Cahn–Hilliard equation on a sphere, where the surface is modeled using a phase-field approach. As for the high-order PDEs on evolving surfaces, a phase-field model of a deforming multi-component vesicle, which consists of two fourth-order nonlinear PDEs, is solved using the isogeometric analysis in a primal variational framework. Through several numerical examples in 2D, 3D and axisymmetric 3D settings, we show the robustness of IGA for solving the considered phase-field models.
Finally, we present a monolithic, implicit formulation based on isogeometric analysis and generalized-alpha time integration for simulating hydrodynamics of vesicles according to a phase-field model. Compared to earlier works, the number of equations of the phase-field model which need to be solved is reduced by leveraging high continuity of NURBS functions, and the algorithm is extended to 3D settings. We use residual-based variational multi-scale method (RBVMS) for solving Navier–Stokes equations, while the rest of PDEs in the phase-field model are treated using a standard Galerkin-based IGA. We introduce the resistive immersed surface (RIS) method into the formulation which can be employed for an implicit description of complex geometries using a diffuse-interface approach. The implementation highlights the robustness of the RBVMS method for Navier–Stokes equations of incompressible flows with non-trivial localized forcing terms including bending and tension forces of the vesicle. The potential of the phase-field model and isogeometric analysis for accurate simulation of a variety of fluid-vesicle interaction problems in 2D and 3D is demonstrated.
Das Ziel der Arbeit ist, eine mögliche Verbesserung der Güte der Lebensdauervorhersage für Gusseisenwerkstoffe mit Kugelgraphit zu erreichen, wobei die Gießprozesse verschiedener Hersteller berücksichtigt werden.
Im ersten Schritt wurden Probenkörper aus GJS500 und GJS600 von mehreren Gusslieferanten gegossen und daraus Schwingproben erstellt.
Insgesamt wurden Schwingfestigkeitswerte der einzelnen gegossenen Proben sowie der Proben des Bauteils von verschiedenen Gussherstellern weltweit entweder durch direkte Schwingversuche oder durch eine Sammlung von Betriebsfestigkeitsversuchen bestimmt.
Dank der metallografischen Arbeit und Korrelationsanalyse konnten drei wesentliche Parameter zur Bestimmung der lokalen Dauerfestigkeit festgestellt werden: 1. statische Festigkeit, 2. Ferrit- und Perlitanteil der Mikrostrukturen und 3. Kugelgraphitanzahl pro Flächeneinheit.
Basierend auf diesen Erkenntnissen wurde ein neues Festigkeitsverhältnisdiagramm (sogenanntes Sd/Rm-SG-Diagramm) entwickelt.
Diese neue Methodik sollte vor allem ermöglichen, die Bauteildauerfestigkeit auf der Grundlage der gemessenen oder durch eine Gießsimulation vorhersagten lokalen Zugfestigkeitswerte sowie Mikrogefügenstrukturen besser zu prognostizieren.
Mithilfe der Versuche sowie der Gießsimulation ist es gelungen, unterschiedliche Methoden der Lebensdauervorhersage unter Berücksichtigung der Herstellungsprozesse weiterzuentwickeln.
Antimicrobial resistance (AMR) is identified by the World Health Organization (WHO) as one of the top ten threats to public health worldwide. In addition to public health, AMR also poses a major threat to food security and economic development. Current sanitation systems contribute to the emergence and spread of AMR and lack effective AMR mitigation measures. This study assesses source separation of blackwater as a mitigation measure against AMR. A source-separation-modified combined sanitation system with separate collection of blackwater and graywater is conceptually described. Measures taken at the source, such as the separate collection and discharge of material flows, were not considered so far on a load balance basis, i.e., they have not yet been evaluated for their effectiveness. The sanitation system described is compared with a combined system and a separate system regarding AMR emissions by means of simulation. AMR is represented in the simulation model by one proxy parameter each for antibiotics (sulfamethoxa-zole), antibiotic-resistant bacteria (extended-spectrum beta-lactamase E. Coli), and antibiotic re-sistance genes (blaTEM). The simulation results suggest that the source-separation-based sanitation system reduces emissions of antibiotic-resistant bacteria and antibiotic resistance genes into the aquatic environment by more than six logarithm steps compared to combined systems. Sulfa-methoxazole emissions can be reduced by 75.5% by keeping blackwater separate from graywater and treating it sufficiently. In summary, sanitation systems incorporating source separation are, to date, among the most effective means of preventing the emission of AMR into the aquatic envi-ronment.
Die vorliegende Arbeit richtet sich an Ingenieur*innen und Wissenschaftler*innen der technischen Gebäudeausrüstung. Sie greift einen sich abzeichnenden Änderungsbedarf in der Umwelt- und Nachhaltigkeitsbewertung von Gebäuden und wärmetechnischen Anlagen auf. Der aktuell genutzte nicht erneuerbare Primärenergiebedarf wird insbesondere hinsichtlich künftiger politischer Klima- und Umweltschutzziele als alleinige Bewertungsgröße nicht ausreichend sein. Die mit dieser Arbeit vorgestellte Ökoeffizienzbewertungsmethode kann als geeignetes Instrument zur Lösung der Probleme beitragen. Sie ermöglicht systematische, ganzheitliche Bewertungen und reproduzierbare Vergleiche wärmetechnischer Anlagen bezüglich ihrer ökologischen und ökonomischen Nachhaltigkeit. Die wesentlichsten Neuentwicklungen sind die spezifische Umweltleistung, in Erweiterung zum genutzten Primärenergiefaktor, und der Ökoeffizienzindikator UWI.
Accurate prediction of stable alluvial hydraulic geometry, in which erosion and sedimentation are in equilibrium, is one of the most difficult but critical topics in the field of river engineering. Data mining algorithms have been gaining more attention in this field due to their high performance and flexibility. However, an understanding of
the potential for these algorithms to provide fast, cheap, and accurate predictions of hydraulic geometry is lacking. This study provides the first quantification of this potential. Using at-a-station field data, predictions of flow depth, water-surface width and longitudinal water surface slope are made using three standalone data mining techniques -, Instance-based Learning (IBK), KStar, Locally Weighted Learning (LWL) - along with four types of novel hybrid algorithms in which the standalone models are trained with Vote, Attribute Selected
Classifier (ASC), Regression by Discretization (RBD), and Cross-validation Parameter Selection (CVPS) algorithms (Vote-IBK, Vote-Kstar, Vote-LWL, ASC-IBK, ASC-Kstar, ASC-LWL, RBD-IBK, RBD-Kstar, RBD-LWL, CVPSIBK, CVPS-Kstar, CVPS-LWL). Through a comparison of their predictive performance and a sensitivity analysis of the driving variables, the results reveal: (1) Shield stress was the most effective parameter in the prediction of all geometry dimensions; (2) hybrid models had a higher prediction power than standalone data mining models,
empirical equations and traditional machine learning algorithms; (3) Vote-Kstar model had the highest performance in predicting depth and width, and ASC-Kstar in estimating slope, each providing very good prediction performance. Through these algorithms, the hydraulic geometry of any river can potentially be predicted accurately and with ease using just a few, readily available flow and channel parameters. Thus, the results reveal that these models have great potential for use in stable channel design in data poor catchments, especially in developing nations where technical modelling skills and understanding of the hydraulic and sediment processes occurring in the river system may be lacking.
Die Auseinandersetzung mit der Digitalisierung ist in den letzten Jahren in den Medien, auf Konferenzen und in Ausschüssen der Bau- und Immobilienbranche angekommen. Während manche Bereiche Neuerungen hervorbringen und einige Akteure als Pioniere zu bezeichnen sind, weisen andere Themen noch Defizite hinsichtlich der digitalen Transformation auf. Zu dieser Kategorie kann auch das Baugenehmigungsverfahren gezählt werden. Unabhängig davon, wie Architekten und Ingenieure in den Planungsbüros auf innovative Methoden setzen, bleiben die Bauvorlagen bisher zuhauf in Papierform oder werden nach der elektronischen Einreichung in der Behörde ausgedruckt. Vorhandene Ressourcen, beispielsweise in Form eines Bauwerksinformationsmodells, die Unterstützung bei der Baugenehmigungsfeststellung bieten können, werden nicht ausgeschöpft. Um mit digitalen Werkzeugen eine Entscheidungshilfe für die Baugenehmigungsbehörden zu erarbeiten, ist es notwendig, den Ist-Zustand zu verstehen und Gegebenheiten zu hinterfragen, bevor eine Gesamtautomatisierung der innerbehördlichen Vorgänge als alleinige Lösung zu verfolgen ist.
Mit einer inhaltlich-organisatorischen Betrachtung der relevanten Bereiche, die Einfluss auf die Baugenehmigungsfeststellung nehmen, wird eine Optimierung des Baugenehmigungsverfahrens in den
Behörden angestrebt. Es werden die komplexen Bereiche, wie die Gesetzeslage, der Einsatz von Technologie aber auch die subjektiven Handlungsalternativen, ermittelt und strukturiert. Mit der Entwicklung eines Modells zur Feststellung der Baugenehmigungsfähigkeit wird sowohl ein Verständnis für Einflussfaktoren vermittelt als auch eine Transparenzsteigerung für alle Beteiligten geschaffen.
Neben einer internationalen Literaturrecherche diente eine empirische Studie als Untersuchungsmethode. Die empirische Studie wurde in Form von qualitativen Experteninterviews durchgeführt, um den Ist-Zustand im Bereich der Baugenehmigungsverfahren festzustellen. Das erhobene Datenmaterial wurde aufbereitet und anschließend einer softwaregestützten Inhaltsanalyse unterzogen. Die Ergebnisse wurden in Kombination mit den Erkenntnissen der Literaturrecherche in verschiedenen Analysen als Modellgrundlage aufgearbeitet.
Ergebnis der Untersuchung stellt ein Entscheidungsmodell dar, welches eine Lücke zwischen den gegenwärtigen
Abläufen in den Baubehörden und einer Gesamtautomatisierung der Baugenehmigungsprüfung schließt. Die prozessorientierte Strukturierung entscheidungsrelevanter Sachverhalte im Modell ermöglicht eine Unterstützung bei der Baugenehmigungsfeststellung für Prüfer und Antragsteller. Das theoretische Modell konnte in Form einer Webanwendung in die Praxis übertragen werden.
Für die Verminderung der betonspezifischen CO2-Emissionen wird ein verstärkter Einsatz klinkerreduzierter Zemente bzw. Betone angestrebt. Die Reduzierung des Klinkergehaltes darf jedoch nicht zu einer lebensdauerrelevanten Beeinträchtigung der Betondauerhaftigkeit führen. In diesem Zusammenhang stellt der Frost-Tausalz-Widerstand eine kritische Größe dar, da er bei höheren Klinkersubstitutionsraten häufig negativ beeinflusst wird. Erschwerend kommt hinzu, dass für klinkerreduzierte Betone nur ein unzureichender Erfahrungsschatz vorliegt. Ein hoher Frost-Tausalz-Widerstand kann daher nicht ausschließlich anhand deskriptiver Vorgaben gewährleistet werden. Demgemäß sollte perspektivisch auch für frost-tausalzbeanspruchte Bauteile eine performancebasierte Lebensdauerbetrachtung erfolgen.
Eine unverzichtbare Grundlage für das Erreichen dieser Ziele ist ein Verständnis für die Schadensvorgänge beim Frost-Tausalz-Angriff. Der Forschungsstand ist jedoch geprägt von widersprüchlichen Schadenstheorien. Somit wurde als Zielstellung für diese Arbeit abgeleitet, die existierenden Schadenstheorien unter Berücksichtigung des aktuellen Wissensstandes zu bewerten und mit eigenen Untersuchungen zu prüfen und einzuordnen. Die Sichtung des Forschungsstandes zeigte, dass nur zwei Theorien das Potential haben, den Frost-Tausalz-Angriff umfassend abzubilden – die Glue Spall Theorie und die Cryogenic Suction Theorie.
Die Glue Spall Theorie führt die Entstehung von Abwitterungen auf die mechanische Schädigung der Betonoberfläche durch eine anhaftende Eisschicht zurück. Dabei sollen nur bei moderaten Tausalzkonzentrationen in der einwirkenden Lösung kritische Spannungszustände in der Eisschicht auftreten, die eine Schädigung der Betonoberfläche hervorrufen können. In dieser Arbeit konnte jedoch nachgewiesen werden, dass starke Abwitterungen auch bei Tausalz¬konzentrationen auftreten, bei denen eine mechanische Schädigung des Betons durch das Eis auszuschließen ist. Damit wurde die fehlende Eignung der Glue Spall Theorie aufgezeigt.
Die Cryogenic Suction Theorie fußt auf den eutektischen Eigenschaften von Tausalz-lösungen, die im gefrorenen Zustand immer als Mischung auf festem Wassereis und flüssiger, hochkonzentrierter Salzlösung bestehen, solange ihre Eutektikumstemperatur nicht unter¬schritten wird. Die flüssige Phase im salzhaltigen Eis stellt für gefrorenen Beton ein bisher nicht berücksichtigtes Flüssigkeitsreservoir dar, welches trotz der hohen Salzkonzentration die Eisbildung in der Betonrandzone verstärken und so die Entstehung von Abwitterungen verursachen soll. In dieser Arbeit wurde bestätigt, dass die Eisbildung im Zementstein beim Gefrieren in hochkonzentrierter Tausalzlösung tatsächlich verstärkt wird. Das Ausmaß der zusätzlichen Eisbildung wurde dabei auch von der Fähigkeit des Zementsteins zur Bindung von Chloridionen aus der Tausalzlösung beeinflusst.
Zusammenfassend wurde festgestellt, dass die Cryogenic Suction Theorie eine gute Beschreibung des Frost-Tausalz-Angriffes darstellt, aber um weitere Aspekte ergänzt werden muss. Die Berücksichtigung der intensiven Sättigung von Beton durch den Prozess der Mikroeislinsenpumpe stellt hier die wichtigste Erweiterung dar. Basierend auf dieser Überlegung wurde eine kombinierte Schadenstheorie aufgestellt. Wichtige Annahmen dieser Theorie konnten experimentell bestätigt werden. Im Ergebnis wurde so die Grundlage für ein tiefergehendes Verständnis des Frost-Tausalz-Angriffes geschaffen. Zudem wurde ein neuer Ansatz identifiziert, um die (potentielle) Verringerung des Frost-Tausalz-Widerstandes klinkerreduzierter Betone zu erklären.
In den letzten Jahrzehnten unterlag der Straßenbetriebsdienst tiefgreifenden Veränderungen. Diese Veränderungen schließt auch die betriebliche Steuerungsphilosophie ein, um eine planungsrationale und ökonomische Gestaltung des Straßenbetriebsdienstes zu unterstützen. Dabei erfolgt eine verbindliche Vorgabe der Leistungsinhalte und -umfänge und ermöglicht eine Budgetierung für das vorgesehene Jahresarbeitsprogramm.
Ziel der Untersuchung ist die Entwicklung eines Modells für die Ermittlung von leistungsbezogenen Musterjahresganglinien zur Unterstützung der Jahresarbeitsplanung. Dafür lagen für jede Leistung des Leistungsbereiches „Grünpflege“ jeweils 260 einzelne Jahresganglinien vor.
Im Ergebnis der Untersuchung wird die leistungsbezogene Musterjahresganglinie in vier Schritten ermittelt. Im ersten Schritt erfolgt die Prüfung der Datenqualität; im zweiten Schritt eine Korrelationsanalyse; im dritten Schritt die fachliche Überprüfung der Leistungsausprägung und im vierten Schritt die Ermittlung der leistungsbezogenen Musterjahresganglinie aus den verbliebenen leistungsbezogenen Jahresganglinien.
“How to understand the interaction between urban space and social processes” is a significant question in urban studies. To answer that, the city needs to be recognized as both a physical and a social entity and urban theory and practice need to connect these (Hillier 2007). The present research aims to re-examine the complex correlation between spatial and social inequality manifestations in the city of Tehran regarding the concept of segregation.
It observes the causes and consequences of segregation in Tehran and provides an insight into both concepts of socio-spatial segregation and neighborhood effects and creates a link between them. First, I argue when, where, and for whom spatial locations affect the chances of social networks in Tehran. Then, I discuss how neighborhood effects can emerge via social network mechanisms and thus affect the perceptions of residents in the neighborhoods.
This work presents a robust status monitoring approach for detecting damage in cantilever structures based on logistic functions. Also, a stochastic damage identification approach based on changes of eigenfrequencies is proposed. The proposed algorithms are verified using catenary poles of electrified railways track. The proposed damage features overcome the limitation of frequency-based damage identification methods available in the literature, which are valid to detect damage in structures to Level 1 only. Changes in eigenfrequencies of cantilever structures are enough to identify possible local damage at Level 3, i.e., to cover damage detection, localization, and quantification. The proposed algorithms identified the damage with relatively small errors, even at a high noise level.
Die Arbeit leistet einen wissenschaftlichen Beitrag zur Erforschung der Einsatzmöglichkeiten eines Immobilienportfoliomanagements für öffentliche museale Schlösserverwaltungen in Deutschland. Insbesondere wird ein für deren Organisation spezifisches Modell zur Investitionssteuerung herausgearbeitet und dessen Anwendbarkeit in der Praxis mit Experten diskutiert.
One of the most important subjects of hydraulic engineering is the reliable estimation of the transverse distribution in the rectangular channel of bed and wall shear stresses. This study makes use of the Tsallis entropy, genetic programming (GP) and adaptive neuro-fuzzy inference system (ANFIS) methods to assess the shear stress distribution (SSD) in the rectangular channel.
To evaluate the results of the Tsallis entropy, GP and ANFIS models, laboratory observations were used in which shear stress was measured using an optimized Preston tube. This is then used to measure the SSD in various aspect ratios in the rectangular channel. To investigate the shear stress percentage, 10 data series with a total of 112 different data for were used. The results of the sensitivity analysis show that the most influential parameter for the SSD in smooth rectangular channel is the dimensionless parameter B/H, Where the transverse coordinate is B, and the flow depth is H. With the parameters (b/B), (B/H) for the bed and (z/H), (B/H) for the wall as inputs, the modeling of the GP was better than the other one. Based on the analysis, it can be concluded that the use of GP and ANFIS algorithms is more effective in estimating shear stress in smooth rectangular channels than the Tsallis entropy-based equations.
In the last two decades, Peridynamics (PD) attracts much attention in the field of fracture mechanics. One key feature of PD is the nonlocality, which is quite different from the ideas in conventional methods such as FEM and meshless method. However, conventional PD suffers from problems such as constant horizon, explicit algorithm, hourglass mode. In this thesis, by examining the nonlocality with scrutiny, we proposed several new concepts such as dual-horizon (DH) in PD, dual-support (DS) in smoothed particle hydrodynamics (SPH), nonlocal operators and operator energy functional. The conventional PD (SPH) is incorporated in the DH-PD (DS-SPH), which can adopt an inhomogeneous discretization and inhomogeneous support domains. The DH-PD (DS-SPH) can be viewed as some fundamental improvement on the conventional PD (SPH). Dual formulation of PD and SPH allows h-adaptivity while satisfying the conservations of linear momentum, angular momentum and energy. By developing the concept of nonlocality further, we introduced the nonlocal operator method as a generalization of DH-PD. Combined with energy functional of various physical models, the nonlocal forms based on dual-support concept are derived. In addition, the variation of the energy functional allows implicit formulation of the nonlocal theory. At last, we developed the higher order nonlocal operator method which is capable of solving higher order partial differential equations on arbitrary domain in higher dimensional space. Since the concepts are developed gradually, we described our findings chronologically.
In chapter 2, we developed a DH-PD formulation that includes varying horizon sizes and solves the "ghost force" issue. The concept of dual-horizon considers the unbalanced interactions between the particles with different horizon sizes. The present formulation fulfills both the balances of linear momentum and angular momentum exactly with arbitrary particle discretization. All three peridynamic formulations, namely bond based, ordinary state based and non-ordinary state based peridynamics can be implemented within the DH-PD framework. A simple adaptive refinement procedure (h-adaptivity) is proposed reducing the computational cost. Both two- and three- dimensional examples including the Kalthoff-Winkler experiment and plate with branching cracks are tested to demonstrate the capability of the method.
In chapter 3, a nonlocal operator method (NOM) based on the variational principle is proposed for the solution of waveguide problem in computational electromagnetic field. Common differential operators as well as the variational forms are defined within the context of nonlocal operators. The present nonlocal formulation allows the assembling of the tangent stiffness matrix with ease, which is necessary for the eigenvalue analysis of the waveguide problem. The present formulation is applied to solve 1D Schrodinger equation, 2D electrostatic problem and the differential electromagnetic vector wave equations based on electric fields.
In chapter 4, a general nonlocal operator method is proposed which is applicable for solving partial differential equations (PDEs) of mechanical problems. The nonlocal operator can be regarded as the integral form, ``equivalent'' to the differential form in the sense of a nonlocal interaction model. The variation of a nonlocal operator plays an equivalent role as the derivatives of the shape functions in the meshless methods or those of the finite element method. Based on the variational principle, the residual and the tangent stiffness matrix can be obtained with ease. The nonlocal operator method is enhanced here also with an operator energy functional to satisfy the linear consistency of the field. A highlight of the present method is the functional derived based on the nonlocal operator can convert the construction of residual and stiffness matrix into a series of matrix multiplications using the predefined nonlocal operators. The nonlocal strong forms of different functionals can be obtained easily via the concept of support and dual-support. Several numerical examples of different types of PDEs are presented.
In chapter 5, we extended the NOM to higher order scheme by using a higher order Taylor series expansion of the unknown field. Such a higher order scheme improves the original NOM in chapter 3 and chapter 4, which can only achieve one-order convergence. The higher order NOM obtains all partial derivatives with specified maximal order simultaneously without resorting to shape functions. The functional based on the nonlocal operators converts the construction of residual and stiffness matrix into a series of matrix multiplication on the nonlocal operator matrix. Several numerical examples solved by strong form or weak form are presented to show the capabilities of this method.
In chapter 6, the NOM proposed as a particle-based method in chapter 3,4,5, has difficulty in imposing accurately the boundary conditions of various orders. In this paper, we converted the particle-based NOM into a scheme with interpolation property. The new scheme describes partial derivatives of various orders at a point by the nodes in the support and takes advantage of the background mesh for numerical integration. The boundary conditions are enforced via the modified variational principle. The particle-based NOM can be viewed a special case of NOM with interpolation property when nodal integration is used. The scheme based on numerical integration greatly improves the stability of the method, as a consequence, the operator energy functional in particle-based NOM is not required. We demonstrated the capabilities of current method by solving the gradient solid problems and comparing the numerical results with the available exact solutions.
In chapter 7, we derived the DS-SPH in solid within the framework of variational principle. The tangent stiffness matrix of SPH can be obtained with ease, and can be served as the basis for the present implicit SPH. We proposed an hourglass energy functional, which allows the direct derivation of hourglass force and hourglass tangent stiffness matrix. The dual-support is {involved} in all derivations based on variational principles and is automatically satisfied in the assembling of stiffness matrix. The implementation of stiffness matrix comprises with two steps, the nodal assembly based on deformation gradient and global assembly on all nodes. Several numerical examples are presented to validate the method.
This dissertation investigates the interactions between urban form, allocation of activities, and pedestrian movement in the context of urban planning. The ability to assess the long-term impact of urban planning decisions on what people do and how they get there is of central importance, with various disciplines addressing this topic. This study focuses on approaches proposed by urban morphologists, urban economists, and transportation planners, each aiming the attention at a different part of the form-activity-movement interaction. Even though there is no doubt about the advantages of these highly focused approaches, it remains unclear what is the cost of ignoring the effect of some interactions while considering others. The general aim of this dissertation is to empirically test the validity of the individual models and quantify the impact of this isolationist approach on their precision and bias.
For this purpose, we propose a joined form-activity-movement interaction model and conduct an empirical study in Weimar, Germany. We estimate how the urban form and activities affect movement as well as how movement and urban form affect activities. By estimating these effects in isolation and simultaneously, we assess the bias of the individual models.
On the one hand, the empirical study results confirm the significance of all interactions suggested by the individual models. On the other hand, we were able to show that when these interactions are estimated in isolation, the resulting predictions are biased. To conclude, we do not question the knowledge brought by transportation planners, urban morphologists, and urban economists. However, we argue that it might be of little use on its own.
We see the relevance of this study as being twofold. On the one hand, we proposed a novel methodological framework for the simultaneous estimation of the form-activity-movement interactions. On the other hand, we provide empirical evidence about the strengths and limitations of current approaches.
In the last decades, Finite Element Method has become the main method in statics and dynamics analysis in engineering practice. For current problems, this method provides a faster, more flexible solution than the analytic approach. Prognoses of complex engineer problems that used to be almost impossible to solve are now feasible.
Although the finite element method is a robust tool, it leads to new questions about engineering solutions. Among these new problems, it is possible to divide into two major groups: the first group is regarding computer performance; the second one is related to understanding the digital solution.
Simultaneously with the development of the finite element method for numerical solutions, a theory between beam theory and shell theory was developed: Generalized Beam Theory, GBT. This theory has not only a systematic and analytical clear presentation of complicated structural problems, but also a compact and elegant calculation approach that can improve computer performance.
Regrettably, GBT was not internationally known since the most publications of this theory were written in German, especially in the first years. Only in recent years, GBT has gradually become a fertile research topic, with developments from linear to non-linear analysis.
Another reason for the misuse of GBT is the isolated application of the theory. Although recently researches apply finite element method to solve the GBT's problems numerically, the coupling between finite elements of GBT and other theories (shell, solid, etc) is not the subject of previous research. Thus, the main goal of this dissertation is the coupling between GBT and shell/membrane elements. Consequently, one achieves the benefits of both sides: the versatility of shell elements with the high performance of GBT elements.
Based on the assumptions of GBT, this dissertation presents how the separation of variables leads to two calculation's domains of a beam structure: a cross-section modal analysis and the longitudinal amplification axis. Therefore, there is the possibility of applying the finite element method not only in the cross-section analysis, but also the development for an exact GBT's finite element in the longitudinal direction.
For the cross-section analysis, this dissertation presents the solution of the quadratic eigenvalue problem with an original separation between plate and membrane mechanism. Subsequently, one obtains a clearer representation of the deformation mode, as well as a reduced quadratic eigenvalue problem.
Concerning the longitudinal direction, this dissertation develops the novel exact elements, based on hyperbolic and trigonometric shape functions. Although these functions do not have trivial expressions, they provide a recursive procedure that allows periodic derivatives to systematise the development of stiffness matrices. Also, these shape functions enable a single-element discretisation of the beam structure and ensure a smooth stress field.
From these developments, this dissertation achieves the formulation of its primary objective: the connection of GBT and shell elements in a mixed model. Based on the displacement field, it is possible to define the coupling equations applied in the master-slave method. Therefore, one can model the structural connections and joints with finite shell elements and the structural beams and columns with GBT finite element.
As a side effect, the coupling equations limit the displacement field of the shell elements under the assumptions of GBT, in particular in the neighbourhood of the coupling cross-section.
Although these side effects are almost unnoticeable in linear analysis, they lead to cumulative errors in non-linear analysis. Therefore, this thesis finishes with the evaluation of the mixed GBT-shell models in non-linear analysis.
Complex vortex flow patterns around bridge piers, especially during floods, cause scour process that can result in the failure of foundations. Abutment scour is a complex three-dimensional phenomenon that is difficult to predict especially with traditional formulas obtained using empirical approaches such as regressions. This paper presents a test of a standalone Kstar model with five novel hybrid algorithm of bagging (BA-Kstar), dagging (DA-Kstar), random committee (RC-Kstar), random subspace (RS-Kstar), and weighted instance handler wrapper (WIHWKstar) to predict scour depth (ds) for clear water condition. The dataset consists of 99 scour depth data from flume experiments (Dey and Barbhuiya, 2005) using abutment shapes such as vertical, semicircular and 45◦ wing. Four dimensionless parameter of relative flow depth (h/l), excess abutment Froude number (Fe), relative sediment size (d50/l) and relative submergence (d50/h) were considered for the prediction of relative scour depth (ds/l). A portion of the dataset was used for the calibration (70%), and the remaining used for model validation. Pearson correlation coefficients helped deciding relevance of the input parameters combination and finally four different combinations of input parameters were used. The performance of the models was assessed visually and with quantitative metrics. Overall, the best input combination for vertical abutment shape is the combination of Fe, d50/l and h/l, while for semicircular and 45◦ wing the combination of the Fe and d50/l is the most effective input parameter combination. Our results show that incorporating Fe, d50/l and h/l lead to higher performance while involving d50/h reduced the models prediction power for vertical abutment shape and for semicircular and 45◦ wing involving h/l and d50/h lead to more error. The WIHW-Kstar provided the highest performance in scour depth prediction around vertical abutment shape while RC-Kstar model outperform of other models for scour depth prediction around semicircular and 45◦ wing.
This thesis presents the advances and applications of phase field modeling in fracture analysis. In this approach, the sharp crack surface topology in a solid is approximated by a diffusive crack zone governed by a scalar auxiliary variable. The uniqueness of phase field modeling is that the crack paths are automatically determined as part of the solution and no interface tracking is required. The damage parameter varies continuously over the domain. But this flexibility comes with associated difficulties: (1) a very fine spatial discretization is required to represent sharp local gradients correctly; (2) fine discretization results in high computational cost; (3) computation of higher-order derivatives for improved convergence rates and (4) curse of dimensionality in conventional numerical integration techniques. As a consequence, the practical applicability of phase field models is severely limited.
The research presented in this thesis addresses the difficulties of the conventional numerical integration techniques for phase field modeling in quasi-static brittle fracture analysis. The first method relies on polynomial splines over hierarchical T-meshes (PHT-splines) in the framework of isogeometric analysis (IGA). An adaptive h-refinement scheme is developed based on the variational energy formulation of phase field modeling. The fourth-order phase field model provides increased regularity in the exact solution of the phase field equation and improved convergence rates for numerical solutions on a coarser discretization, compared to the second-order model. However, second-order derivatives of the phase field are required in the fourth-order model. Hence, at least a minimum of C1 continuous basis functions are essential, which is achieved using hierarchical cubic B-splines in IGA. PHT-splines enable the refinement to remain local at singularities and high gradients, consequently reducing the computational cost greatly. Unfortunately, when modeling complex geometries, multiple parameter spaces (patches) are joined together to describe the physical domain and there is typically a loss of continuity at the patch boundaries. This decrease of smoothness is dictated by the geometry description, where C0 parameterizations are normally used to deal with kinks and corners in the domain. Hence, the application of the fourth-order model is severely restricted. To overcome the high computational cost for the second-order model, we develop a dual-mesh adaptive h-refinement approach. This approach uses a coarser discretization for the elastic field and a finer discretization for the phase field. Independent refinement strategies have been used for each field.
The next contribution is based on physics informed deep neural networks. The network is trained based on the minimization of the variational energy of the system described by general non-linear partial differential equations while respecting any given law of physics, hence the name physics informed neural network (PINN). The developed approach needs only a set of points to define the geometry, contrary to the conventional mesh-based discretization techniques. The concept of `transfer learning' is integrated with the developed PINN approach to improve the computational efficiency of the network at each displacement step. This approach allows a numerically stable crack growth even with larger displacement steps. An adaptive h-refinement scheme based on the generation of more quadrature points in the damage zone is developed in this framework. For all the developed methods, displacement-controlled loading is considered. The accuracy and the efficiency of both methods are studied numerically showing that the developed methods are powerful and computationally efficient tools for accurately predicting fractures.
The concept of information entropy together with the principle of maximum entropy to open channel flow is essentially based on some physical consideration of the problem under consideration. This paper is a discussion on Yeganeh and Heidari (2020)’s paper, who proposed a new approach for measuring vertical distribution of streamwise velocity in open channels. The discussers argue that their approach is conceptually incorrect and thus leads to a physically unrealistic situation. In addition, the discussers found some wrong mathematical expressions (which are assumed to be typos) written in the paper, and also point out that the authors did not cite some of the original papers on the topic.