Refine
Document Type
- Article (38)
- Doctoral Thesis (37)
- Conference Proceeding (4)
- Book (3)
- Master's Thesis (3)
- Other (3)
- Periodical (2)
- Working Paper (2)
- Bachelor Thesis (1)
- Review (1)
Institute
- Institut für Strukturmechanik (ISM) (21)
- Institut für Europäische Urbanistik (11)
- Professur Bauphysik (7)
- Bauhaus-Institut für zukunftsweisende Infrastruktursysteme (b.is) (4)
- Professur Werkstoffe des Bauens (4)
- Professur Modellierung und Simulation - Mechanik (3)
- Professur Stahl- und Hybridbau (3)
- Promotionsstudiengang Kunst und Design-Freie Kunst-Medienkunst (Ph.D) (3)
- DFG-Graduiertenkolleg 2227 "Identität und Erbe" (2)
- Junior-Professur Organisation und vernetzte Medien (2)
Keywords
- OA-Publikationsfonds2022 (26)
- Beton (5)
- Maschinelles Lernen (5)
- Architektur (4)
- Denkmalpflege (4)
- Finite-Elemente-Methode (4)
- Isogeometric Analysis (4)
- Isogeometrische Analyse (4)
- Peridynamics (3)
- Peridynamik (3)
Year of publication
- 2022 (94) (remove)
Open Innovation in kleinen und mittleren Unternehmen (KMU) hat sich stark ausdifferenziert. Dabei zeigt die Empirie, dass KMU unterschiedliche Wege in der offenen Entwicklung von Innovationen begehen. Um die bestehende Literatur zu erweitern, wurden mit dieser Dissertation die Ziele verfolgt 1) offene Innovationsaktivitäten in KMU aus einer Prozessperspektive aufzudecken und genau zu beschreiben und 2) zu erklären, warum sich die Öffnung von Innovationsprozessen in KMU unterscheidet. Dafür wurde auf eine multiple Fallstudienanalyse zurückgegriffen. Untersuchungsobjekte waren kleine etablierte High-Tech Unternehmen aus den neuen Bundesländern. Die Ergebnisse zeigen sechs Prozessmodelle der offenen Innovationsentwicklung, beschrieben als Open Innovation Muster. Deskriptionen dieser Muster unter Berücksichtigung von formenden Innovationsaktivitäten, ausgetauschtem Wissen, beteiligten externen Akteuren und Gründen für und gegen Open Innovation vermitteln ein über den bisherigen Forschungsstand hinausgehendes Verständnis von Open Innovation in KMU. Zudem zeigen die Ergebnisse, dass die Entrepreneurial Orientation erklärt, warum KMU bei der Ausgestaltung von offenen Innovationsprozessen unterschiedlich vorgehen. In der Dissertation wird detailliert dargelegt, welche Open Innovation Muster sich anhand der Entrepreneurial Orientation von KMU (nicht-entrepreneurial bis entrepreneurial) zeigen. Die Ergebnisse liefern sowohl wissenschaftliche Implikationen, als auch Handlungsempfehlungen für die Unternehmenspraxis.
Plastic structural analysis may be applied without any difficulty and with little effort for structural member verifications with regard to lateral torsional buckling of doubly symmetric rolled I sections. Suchlike analyses can be performed based on the plastic zone theory, specifically using finite beam elements with seven degrees of freedom and 2nd order theory considering material nonlinearity. The existing Eurocode enables these approaches and the coming-up generation will provide corresponding regulations in EN 1993-1-14. The investigations allow the determination of computationally accurate limit loads, which are determined in the present paper for selected structural systems with different sets of parameters, such as length, steel grade and cross section types. The results are compared to approximations gained by more sophisticated FEM analyses (commercial software Ansys Workbench applying solid elements) for reasons of verification/validation. In this course, differences in the results of the numerical models are addressed and discussed. In addition, results are compared to resistances obtained by common design regulations based on reduction factors χlt including regulations of EN 1993-1-1 (including German National Annex) as well as prEN 1993-1-1: 2020-08 (proposed new Eurocode generation). Concluding, correlations of results and their advantages as well as disadvantages are discussed.
A safe and economic structural design based on the semi-probabilistic concept requires statistically representative safety elements, such as characteristic values, design values, and partial safety factors. Regarding climate loads, the safety levels of current design codes strongly reflect experiences based on former measurements and investigations assuming stationary conditions, i.e. involving constant frequencies and intensities. However, due to climate change, occurrence of corresponding extreme weather events is expected to alter in the future influencing the reliability and safety of structures and their components. Based on established approaches, a systematically refined data-driven methodology for the determination of design parameters considering nonstationarity as well as standardized targets of structural reliability or safety, respectively, is therefore proposed. The presented procedure picks up fundamentals of European standardization and extends them with respect to nonstationarity by applying a shifting time window method. Taking projected snow loads into account, the application of the method is exemplarily demonstrated and various influencing parameters are discussed.
Object-Oriented Damage Information Modeling Concepts and Implementation for Bridge Inspection
(2022)
Bridges are designed to last for more than 50 years and consume up to 50% of their life-cycle costs during their operation phase. Several inspections and assessment actions are executed during this period. Bridge and damage information must be gathered, digitized, and exchanged between different stakeholders. Currently, the inspection and assessment practices rely on paper-based data collection and exchange, which is time-consuming and error-prone, and leads to loss of information. Storing and exchanging damage and building information in a digital format may lower costs and errors during inspection and assessment and support future needs, for example, immediate simulations regarding performance assessment, automated maintenance planning, and mixed reality inspections. This study focused on the concept for modeling damage information to support bridge reviews and structural analysis. Starting from the definition of multiple use cases and related requirements, the data model for damage information is defined independently from the subsequent implementation. In the next step, the implementation via an established standard is explained. Functional tests aim to identify problems in the concept and implementation. To show the capability of the final model, two example use cases are illustrated: the inspection review of the entire bridge and a finite-element analysis of a single component. Main results are the definition of necessary damage data, an object-oriented damage model, which supports multiple use cases, and the implementation of the model in a standard. Furthermore, the tests have shown that the standard is suitable to deliver damage information; however, several software programs lack proper implementation of the standard.
Finite Element Simulations of dynamically excited structures are mainly influenced by the mass, stiffness, and damping properties of the system, as well as external loads. The prediction quality of dynamic simulations of vibration-sensitive components depends significantly on the use of appropriate damping models. Damping phenomena have a decisive influence on the vibration amplitude and the frequencies of the vibrating structure. However, developing realistic damping models is challenging due to the multiple sources that cause energy dissipation, such as material damping, different types of friction, or various interactions with the environment.
This thesis focuses on thermoelastic damping, which is the main cause of material damping in homogeneous materials. The effect is caused by temperature changes due to mechanical strains. In vibrating structures, temperature gradients arise in adjacent tension and compression areas. Depending on the vibration frequency, they result in heat flows, leading to increased entropy and the irreversible transformation of mechanical energy into thermal energy.
The central objective of this thesis is the development of efficient simulation methods to incorporate thermoelastic damping in finite element analyses based on modal superposition. The thermoelastic loss factor is derived from the structure's mechanical mode shapes and eigenfrequencies. In subsequent analyses that are performed in the time and frequency domain, it is applied as modal damping.
Two approaches are developed to determine the thermoelastic loss in thin-walled plate structures, as well as three-dimensional solid structures. The realistic representation of the dissipation effects is verified by comparing the simulation results with experimentally determined data. Therefore, an experimental setup is developed to measure material damping, excluding other sources of energy dissipation.
The three-dimensional solid approach is based on the determination of the generated entropy and therefore the generated heat per vibration cycle, which is a measure for thermoelastic loss in relation to the total strain energy. For thin plate structures, the amount of bending energy in a modal deformation is calculated and summarized in the so-called Modal Bending Factor (MBF). The highest amount of thermoelastic loss occurs in the state of pure bending. Therefore, the MBF enables a quantitative classification of the mode shapes concerning the thermoelastic damping potential.
The results of the developed simulations are in good agreement with the experimental results and are appropriate to predict thermoelastic loss factors. Both approaches are based on modal superposition with the advantage of a high computational efficiency. Overall, the modeling of thermoelastic damping represents an important component in a comprehensive damping model, which is necessary to perform realistic simulations of vibration processes.
In this study, we propose a nonlocal operator method (NOM) for the dynamic analysis of (thin) Kirchhoff plates. The nonlocal Hessian operator is derived based on a second-order Taylor series expansion. The NOM does not require any shape functions and associated derivatives as ’classical’ approaches such as FEM, drastically facilitating the implementation. Furthermore, NOM is higher order continuous, which is exploited for thin plate analysis that requires C1 continuity. The nonlocal dynamic governing formulation and operator energy functional for Kirchhoff plates are derived from a variational principle. The Verlet-velocity algorithm is used for the time discretization. After confirming the accuracy of the nonlocal Hessian operator, several numerical examples are simulated by the nonlocal dynamic Kirchhoff plate formulation.
The current thesis presents research about new methods of citizen participation based on digital technologies. The focus on the research lies on decentralized methods of participation where citizens take the role of co-creators. The research project first conducted a review of the literature on citizen participation, its origins and the different paradigms that have emerged over the years. The literature review also looked at the influence of technologies on participation processes and the theoretical frameworks that have emerged to understand the introduction of technologies in the context of urban development. The literature review generated the conceptual basis for the further development of the thesis.
The research begins with a survey of technology enabled participation applications that examined the roles and structures emerging due to the introduction of technology. The results showed that cities use technology mostly to control and monitor urban infrastructure and are rather reluctant to give citizens the role of co-creators. Based on these findings, three case studies were developed. Digital tools for citizen participation were conceived and introduced for each case study. The adoption and reaction of the citizens were observed using three data collection methods.
The results of the case studies showed consistently that previous participation and engagement with informal citizen participation are a determinining factor in the potential adoption of digital tools for decentralized engagement. Based on these results, the case studies proposed methods and frameworks that can be used for the conception and introduction of technologies for decentralized citizen participation.
Bauablaufplänen kommt bei der Realisierung von Bauprojekten eine zentrale Rolle zu. Sie dienen der Koordination von Schnittstellen und bilden für die am Projekt Beteiligten die Grundlage für ihre individuelle Planung. Eine verlässliche Terminplanung ist daher von großer Bedeutung, tatsächlich sind aber gerade Bauablaufpläne für ihre Unzuverlässigkeit bekannt.
Aufgrund der langen Vorlaufzeiten bei der Planung von Bauprojekten sind zum Zeitpunkt der Planung viele Informationen nur als Schätzwerte bekannt. Auf der Grundlage dieser geschätzten und damit mit Unsicherheiten behafteten Daten werden im Bauwesen deterministische Terminpläne erstellt. Kommt es während der Realisierung zu Diskrepanzen zwischen Schätzungen und Realität, erfordert dies die Anpassung der Pläne. Aufgrund zahlreicher Abhängigkeiten zwischen den geplanten Aktivitäten können einzelne Planänderungen vielfältige weitere Änderungen und Anpassungen nach sich ziehen und damit einen reibungslosen Projektablauf gefährden.
In dieser Arbeit wird ein Vorgehen entwickelt, welches Bauablaufpläne erzeugt, die im Rahmen der durch das Projekt definierten Abhängigkeiten und Randbedingungen in der Lage sind, Änderungen möglichst gut zu absorbieren. Solche Pläne, die bei auftretenden Änderungen vergleichsweise geringe Anpassungen des Terminplans erfordern, werden hier als robust bezeichnet.
Ausgehend von Verfahren der Projektplanung und Methoden zur Berücksichtigung von Unsicherheiten werden deterministische Terminpläne bezüglich ihres Verhaltens bei eintretenden Änderungen betrachtet. Hierfür werden zunächst mögliche Unsicherheiten als Ursachen für Änderungen benannt und mathematisch abgebildet. Damit kann das Verhalten von Abläufen für mögliche Änderungen betrachtet werden, indem die durch Änderungen erzwungenen angepassten Terminpläne simuliert werden. Für diese Monte-Carlo-Simulationen der angepassten Terminpläne wird sichergestellt, dass die angepassten Terminpläne logische Weiterentwicklungen des deterministischen Terminplans darstellen. Auf der Grundlage dieser Untersuchungen wird ein stochastisches Maß zur Quantifizierung der Robustheit erarbeitet, welches die Fähigkeit eines Planes, Änderungen zu absorbieren, beschreibt. Damit ist es möglich, Terminpläne bezüglich ihrer Robustheit zu vergleichen.
Das entwickelte Verfahren zur Quantifizierung der Robustheit wird in einem Optimierungsverfahren auf Basis Genetischer Algorithmen angewendet, um gezielt robuste Terminpläne zu erzeugen. An Beispielen werden die Methoden demonstriert und ihre Wirksamkeit nachgewiesen.
In the wake of the news industry’s digitization, novel organizations that differ considerably from traditional media firms in terms of their functional roles and organizational practices of media work are emerging. One new type is the field repair organization, which is characterized by supporting high‐quality media work to compensate for the deficits (such as those which come from cost savings and layoffs) which have become apparent in legacy media today. From a practice‐theoretical research perspective and based on semi‐structured interviews, virtual field observations, and document analysis, we have conducted a single case study on Science Media Center Germany (SMC), a unique non‐profit news start‐up launched in 2016 in Cologne, Germany. Our findings show that, in addition to field repair activities, SMC aims to facilitate progress and innovation in the field, which we refer to as field advancement. This helps to uncover emerging needs and anticipates problems before they intensify or even occur, proactively providing products and tools for future journalism. This article contributes to our understanding of novel media organizations with distinct functions in the news industry, allowing for advancements in theory on media work and the organization of journalism in times of digital upheaval.
Compactly, this thesis encompasses two major parts to examine mechanical responses of polymer compounds and two dimensional materials:
1- Molecular dynamics approach is investigated to study transverse impact behavior of polymers, polymer compounds and two dimensional materials.
2- Large deflection of circular and rectangular membranes is examined by employing continuum mechanics approach.
Two dimensional materials (2D), including, Graphene and molybdenum disulfide (MoS2), exhibited new and promising physical and chemical properties, opening new opportunities to be utilized alone or to enhance the performance of conventional materials. These 2D materials have attracted tremendous attention owing to their outstanding physical properties, especially concerning transverse impact loading.
Polymers, with the backbone of carbon (organic polymers) or do not include carbon atoms in the backbone (inorganic polymers) like polydimethylsiloxane (PDMS), have extraordinary characteristics particularly their flexibility leads to various easy ways of forming and casting. These simple shape processing label polymers as an excellent material often used as a matrix in composites (polymer compounds).
In this PhD work, Classical Molecular Dynamics (MD) is implemented to calculate transverse impact loading of 2D materials as well as polymer compounds reinforced with graphene sheets. In particular, MD was adopted to investigate perforation of the target and impact resistance force . By employing MD approach, the minimum velocity of the projectile that could create perforation and passes through the target is obtained. The largest investigation was focused on how graphene could enhance the impact properties of the compound. Also the purpose of this work was to discover the effect of the atomic arrangement of 2D materials on the impact problem. To this aim, the impact properties of two different 2D materials, graphene and MoS2, are studied. The simulation of chemical functionalization was carried out systematically, either with covalently bonded molecules or with non-bonded ones, focusing the following efforts on the covalently bounded species, revealed as the most efficient linkers.
To study transverse impact behavior by using classical MD approach , Large-scale Atomic/Molecular Massively Parallel Simulator (LAMMPS) software, that is well-known among most researchers, is employed. The simulation is done through predefined commands in LAMMPS. Generally these commands (atom style, pair style, angle style, dihedral style, improper style, kspace style, read data, fix, run, compute and so on) are used to simulate and run the model for the desired outputs. Depends on the particles and model types, suitable inter-atomic potentials (force fields) are considered. The ensembles, constraints and boundary conditions are applied depends upon the problem definition. To do so, atomic creation is needed. Python codes are developed to generate particles which explain atomic arrangement of each model. Each atomic arrangement introduced separately to LAMMPS for simulation. After applying constraints and boundary conditions, LAMMPS also include integrators like velocity-Verlet integrator or Brownian dynamics or other types of integrator to run the simulation and finally the outputs are emerged. The outputs are inspected carefully to appreciate the natural behavior of the problem. Appreciation of natural properties of the materials assist us to design new applicable materials.
In investigation on the large deflection of circular and rectangular membranes, which is related to the second part of this thesis, continuum mechanics approach is implemented. Nonlinear Föppl membrane theory, which carefully release nonlinear governing equations of motion, is considered to establish the non-linear partial differential equilibrium equations of the membranes under distributed and centric point loads. The Galerkin and energy methods are utilized to solve non-linear partial differential equilibrium equations of circular and rectangular plates respectively. Maximum deflection as well as stress through the film region, which are kinds of issue in many industrial applications, are obtained.
Zugang zu gesunder und nachhaltiger Ernährung ist in Berlin nicht für alle Menschen eine Selbstverständlichkeit. Um Ernährung für alle gewährleisten zu können, braucht es einen Wandel des Ernährungssystems in Berlin, der eine ökologische, klima- und sozialgerechte Nahrungsproduktion und Verteilung für alle Menschen in der Stadt ermöglicht.
Einen Beitrag um die Ernährung in der Stadt gerechter und nachhaltiger zu gestalten kann ein sogenannter LebensMittelPunkt (LMP) leisten.
LebensMittelPunkte entstehen meist aus ehrenamtlichen Initiativen, können aber auch in Zusammenarbeit mit städtischen Verwaltungen etabliert werden. Eine Zusammenarbeit zwischen zivilgesellschaftlichen Organisationen und Verwaltungen kann dabei Potenziale und Ressourcen freisetzen.
Dieser Leitfaden soll ernährungspolitischen Initiativen und Vereinen aus der Zivilgesellschaft sowie kommunalen oder bezirklichen Verwaltungen in Berlin – und darüber hinaus – Empfehlungen geben, wie ein LebensMittelPunkt in einer gemeinsamen Kooperation aufgebaut werden kann.
Das Ziel der Arbeit ist, eine mögliche Verbesserung der Güte der Lebensdauervorhersage für Gusseisenwerkstoffe mit Kugelgraphit zu erreichen, wobei die Gießprozesse verschiedener Hersteller berücksichtigt werden.
Im ersten Schritt wurden Probenkörper aus GJS500 und GJS600 von mehreren Gusslieferanten gegossen und daraus Schwingproben erstellt.
Insgesamt wurden Schwingfestigkeitswerte der einzelnen gegossenen Proben sowie der Proben des Bauteils von verschiedenen Gussherstellern weltweit entweder durch direkte Schwingversuche oder durch eine Sammlung von Betriebsfestigkeitsversuchen bestimmt.
Dank der metallografischen Arbeit und Korrelationsanalyse konnten drei wesentliche Parameter zur Bestimmung der lokalen Dauerfestigkeit festgestellt werden: 1. statische Festigkeit, 2. Ferrit- und Perlitanteil der Mikrostrukturen und 3. Kugelgraphitanzahl pro Flächeneinheit.
Basierend auf diesen Erkenntnissen wurde ein neues Festigkeitsverhältnisdiagramm (sogenanntes Sd/Rm-SG-Diagramm) entwickelt.
Diese neue Methodik sollte vor allem ermöglichen, die Bauteildauerfestigkeit auf der Grundlage der gemessenen oder durch eine Gießsimulation vorhersagten lokalen Zugfestigkeitswerte sowie Mikrogefügenstrukturen besser zu prognostizieren.
Mithilfe der Versuche sowie der Gießsimulation ist es gelungen, unterschiedliche Methoden der Lebensdauervorhersage unter Berücksichtigung der Herstellungsprozesse weiterzuentwickeln.
Immanuel Kant’s thought is a central historical and theoretical reference in Hans Blumenberg’s metaphorological project. This is demonstrated by the fact that in the Paradigms the author outlines the concept of absolute metaphor by explicitly referring to §59 of the Critique of the Power of Judgment and recognizing in the Kantian symbol a model for his own metaphorics. However, Kant’s name also appears in the chapter on the metaphor of the “terra incognita” that not only did he theorize the presence of symbolic hypotyposis in our language [...] but also made extensive use of metaphors linked to “determinate historical experiences”. In particular: geographical metaphors. In my essay, I would like to start from the analysis of Kant’s geographical metaphors in order to try to rethink Blumenberg’s archaeological method as an archaeology of media that grounds the study of metaphors in the materiality of communication and the combination of tools, agents and media.
In this thesis, a new approach is developed for applications of shape optimization on the time harmonic wave propagation (Helmholtz equation) for acoustic problems. This approach is introduced for different dimensional problems: 2D, 3D axi-symmetric and fully 3D problems. The boundary element method (BEM) is coupled with the isogeometric analysis (IGA) forming the so-called (IGABEM) which speeds up meshing and gives higher accuracy in comparison with standard BEM. BEM is superior for handling unbounded domains by modeling only the inner boundaries and avoiding the truncation error, present in the finite element method (FEM) since BEM solutions satisfy the Sommerfeld radiation condition automatically. Moreover, BEM reduces the space dimension by one from a volumetric three-dimensional problem to a surface two-dimensional problem, or from a surface two-dimensional problem to a perimeter one-dimensional problem. Non-uniform rational B-splines basis functions (NURBS) are used in an isogeometric setting to describe both the CAD geometries and the physical fields.
IGABEM is coupled with one of the gradient-free optimization methods, the Particle Swarm Optimization (PSO) for structural shape optimization problems. PSO is a straightforward method since it does not require any sensitivity analysis but it has some trade-offs with regard to the computational cost. Coupling IGA with optimization problems enables the NURBS basis functions to represent the three models: shape design, analysis and optimization models, by a definition of a set of control points to be the control variables and the optimization parameters as well which enables an easy transition between the three models.
Acoustic shape optimization for various frequencies in different mediums is performed with PSO and the results are compared with the benchmark solutions from the literature for different dimensional problems proving the efficiency of the proposed approach with the following remarks:
- In 2D problems, two BEM methods are used: the conventional isogeometric boundary element method (IGABEM) and the eXtended IGABEM (XIBEM) enriched with the partition-of-unity expansion using a set of plane waves, where the results are generally in good agreement with the linterature with some computation advantage to XIBEM which allows coarser meshes.
-In 3D axi-symmetric problems, the three-dimensional problem is simplified in BEM from a surface integral to a combination of two 1D integrals. The first is the line integral similar to a two-dimensional BEM problem. The second integral is performed over the angle of revolution. The discretization is applied only to the former integration. This leads to significant computational savings and, consequently, better treatment for higher frequencies over the full three-dimensional models.
- In fully 3D problems, a detailed comparison between two BEM methods: the conventional boundary integral equation (CBIE) and Burton-Miller (BM) is provided including the computational cost. The proposed models are enhanced with a modified collocation scheme with offsets to Greville abscissae to avoid placing collocation points at the corners. Placing collocation points on smooth surface enables accurate evaluation of normals for BM formulation in addition to straightforward prediction of jump-terms and avoids singularities in $\mathcal{O} (1/r)$ integrals eliminating the need for polar integration. Furthermore, no additional special treatment is required for the hyper-singular integral while collocating on highly distorted elements, such as those containing sphere poles. The obtained results indicate that, CBIE with PSO is a feasible alternative (except for a small number of fictitious frequencies) which is easier to implement. Furthermore, BM presents an outstanding treatment of the complicated geometry of mufflers with internal extended inlet/outlet tube as an interior 3D Helmholtz acoustic problem instead of using mixed or dual BEM.
The Finite Element Method (FEM) is widely used in engineering for solving Partial Differential Equations (PDEs) over complex geometries. To this end, it is required to provide the FEM software with a geometric model that is typically constructed in a Computer-Aided Design (CAD) software. However, FEM and CAD use different approaches for the mathematical description of the geometry. Thus, it is required to generate a mesh, which is suitable for FEM, based on the CAD model. Nonetheless, this procedure is not a trivial task and it can be time consuming. This issue becomes more significant for solving shape and topology optimization problems, which consist in evolving the geometry iteratively. Therefore, the computational cost associated to the mesh generation process is increased exponentially for this type of applications.
The main goal of this work is to investigate the integration of CAD and CAE in shape and topology optimization. To this end, numerical tools that close the gap between design and analysis are presented. The specific objectives of this work are listed below:
• Automatize the sensitivity analysis in an isogeometric framework for applications in shape optimization. Applications for linear elasticity are considered.
• A methodology is developed for providing a direct link between the CAD model and the analysis mesh. In consequence, the sensitivity analysis can be performed in terms of the design variables located in the design model.
• The last objective is to develop an isogeometric method for shape and topological optimization. This method should take advantage of using Non-Uniform Rational B-Splines (NURBS) with higher continuity as basis functions.
Isogeometric Analysis (IGA) is a framework designed to integrate the design and analysis in engineering problems. The fundamental idea of IGA is to use the same basis functions for modeling the geometry, usually NURBS, for the approximation of the solution fields. The advantage of integrating design and analysis is two-fold. First, the analysis stage is more accurate since the system of PDEs is not solved using an approximated geometry, but the exact CAD model. Moreover, providing a direct link between the design and analysis discretizations makes possible the implementation of efficient sensitivity analysis methods. Second, the computational time is significantly reduced because the mesh generation process can be avoided.
Sensitivity analysis is essential for solving optimization problems when gradient-based optimization algorithms are employed. Automatic differentiation can compute exact gradients, automatically by tracking the algebraic operations performed on the design variables. For the automation of the sensitivity analysis, an isogeometric framework is used. Here, the analysis mesh is obtained after carrying out successive refinements, while retaining the coarse geometry for the domain design. An automatic differentiation (AD) toolbox is used to perform the sensitivity analysis. The AD toolbox takes the code for computing the objective and constraint functions as input. Then, using a source code transformation approach, it outputs a code for computing the objective and constraint functions, and their sensitivities as well. The sensitivities obtained from the sensitivity propagation method are compared with analytical sensitivities, which are computed using a full isogeometric approach.
The computational efficiency of AD is comparable to that of analytical sensitivities. However, the memory requirements are larger for AD. Therefore, AD is preferable if the memory requirements are satisfied. Automatic sensitivity analysis demonstrates its practicality since it simplifies the work of engineers and designers.
Complex geometries with sharp edges and/or holes cannot easily be described with NURBS. One solution is the use of unstructured meshes. Simplex-elements (triangles and tetrahedra for two and three dimensions respectively) are particularly useful since they can automatically parameterize a wide variety of domains. In this regard, unstructured Bézier elements, commonly used in CAD, can be employed for the exact modelling of CAD boundary representations. In two dimensions, the domain enclosed by NURBS curves is parameterized with Bézier triangles. To describe exactly the boundary of a two-dimensional CAD model, the continuity of a NURBS boundary representation is reduced to C^0. Then, the control points are used to generate a triangulation such that the boundary of the domain is identical to the initial CAD boundary representation. Thus, a direct link between the design and analysis discretizations is provided and the sensitivities can be propagated to the design domain.
In three dimensions, the initial CAD boundary representation is given as a collection of NURBS surfaces that enclose a volume. Using a mesh generator (Gmsh), a tetrahedral mesh is obtained. The original surface is reconstructed by modifying the location of the control points of the tetrahedral mesh using Bézier tetrahedral elements and a point inversion algorithm. This method offers the possibility of computing the sensitivity analysis using the analysis mesh. Then, the sensitivities can be propagated into the design discretization. To reuse the mesh originally generated, a moving Bézier tetrahedral mesh approach was implemented.
A gradient-based optimization algorithm is employed together with a sensitivity propagation procedure for the shape optimization cases. The proposed shape optimization approaches are used to solve some standard benchmark problems in structural mechanics. The results obtained show that the proposed approach can compute accurate gradients and evolve the geometry towards optimal solutions. In three dimensions, the moving mesh approach results in faster convergence in terms of computational time and avoids remeshing at each optimization step.
For considering topological changes in a CAD-based framework, an isogeometric phase-field based shape and topology optimization is developed. In this case, the diffuse interface of a phase-field variable over a design domain implicitly describes the boundaries of the geometry. The design variables are the local values of the phase-field variable. The descent direction to minimize the objective function is found by using the sensitivities of the objective function with respect to the design variables. The evolution of the phase-field is determined by solving the time dependent Allen-Cahn equation.
Especially for topology optimization problems that require C^1 continuity, such as for flexoelectric structures, the isogeometric phase field method is of great advantage. NURBS can achieve the desired continuity more efficiently than the traditional employed functions. The robustness of the method is demonstrated when applied to different geometries, boundary conditions, and material configurations. The applications illustrate that compared to piezoelectricity, the electrical performance of flexoelectric microbeams is larger under bending. In contrast, the electrical power for a structure under compression becomes larger with piezoelectricity.
In this work, the degradation performance for the photocatalytic oxidation of eight micropollutants (amisulpride, benzotriazole, candesartan, carbamazepine, diclofenac, gabapentin, methlybenzotriazole, and metoprolol) within real secondary effluent was investigated using three different reactor designs. For all reactor types, the influence of irradiation power on its reaction rate and energetic efficiency was investigated. Flat cell and batch reactor showed almost similar substance specific degradation behavior. Within the immersion rotary body reactor, benzotriazole and methylbenzotriazole showed a significantly lower degradation affinity. The flat cell reactor achieved the highest mean degradation rate, with half time values ranging from 5 to 64 min with a mean of 18 min, due to its high catalysts surface to hydraulic volume ratio. The EE/O values were calculated for all micro-pollutants as well as the mean degradation rate constant of each experimental step. The lowest substance specific energy per order (EE/O) values of 5 kWh/m3 were measured for benzotriazole within the batch reactor. The batch reactor also reached the lowest mean values (11.8–15.9 kWh/m3) followed by the flat cell reactor (21.0–37.0 kWh/m3) and immersion rotary body reactor (23.9–41.0 kWh/m3). Catalyst arrangement and irradiation power were identified as major influences on the energetic performance of the reactors. Low radiation intensities as well as the use of submerged catalyst arrangement allowed a reduction in energy demand by a factor of 3–4. A treatment according to existing treatment goals of wastewater treatment plants (80% total degradation) was achieved using the batch reactor with a calculated energy demand of 7000 Wh/m3.
Data acquisition systems and methods to capture high-resolution images or reconstruct 3D point clouds of existing structures are an effective way to document their as-is condition. These methods enable a detailed analysis of building surfaces, providing precise 3D representations. However, for the condition assessment and documentation, damages are mainly annotated in 2D representations, such as images, orthophotos, or technical drawings, which do not allow for the application of a 3D workflow or automated comparisons of multitemporal datasets. In the available software for building heritage data management and analysis, a wide range of annotation and evaluation functions are available, but they also lack integrated post-processing methods and systematic workflows. The article presents novel methods developed to facilitate such automated 3D workflows and validates them on a small historic church building in Thuringia, Germany. Post-processing steps using photogrammetric 3D reconstruction data along with imagery were implemented, which show the possibilities of integrating 2D annotations into 3D documentations. Further, the application of voxel-based methods on the dataset enables the evaluation of geometrical changes of multitemporal annotations in different states and the assignment to elements of scans or building models. The proposed workflow also highlights the potential of these methods for condition assessment and planning of restoration work, as well as the possibility to represent the analysis results in standardised building model formats.
Wer von Erbe im Zusammenhang mit Identität spricht, verspricht sich und Anderen »Kontinuität« und »Stabilität«. Das Versprechen hält indes nur so lange, wie sich Menschen auf die damit verbundenen Erzählungen einlassen. Da diese zunehmend hinterfragt werden und der Begriff »Identität« im politischen Raum zu einer umkämpften Kategorie avanciert ist, werden auch die lange gehegten, gewohnten »Konstruktionen« instabil. Dies zeigt sich insbesondere in Momenten des Konflikts, der übergriffigen Inanspruchnahme und des Verlusts. Der Titel »Instabile Konstruktionen« verweist zugleich auf die beiden Kernbereiche des Kollegs: einerseits auf Architektur und Denkmalpflege, in denen der Begriff Konstruktion sich auf bauliche Manifestationen bezieht, von denen eine gewisse Haltbarkeit und Dauerhaftigkeit erwartet wird und andererseits auf die Kultur- und Sozialwissenschaften, wo Konstruktion die soziale Herstellung symbolischer Sinnwelten meint. Ins Zentrum rückt so der Anspruch, die materielle Umwelt im Wechselverhältnis zu ihrer sozialen Gemachtheit zu verstehen.
The floods in 2002 and 2013, as well as the recent flood of 2021, caused billions Euros worth of property damage in Germany. The aim of the project Innovative Vulnerability and Risk Assessment of Urban Areas against Flood Events (INNOVARU) involved the development of a practicable flood damage model that enables realistic damage statements for the residential building stock. In addition to the determination of local flood risks, it also takes into account the vulnerability of individual buildings and allows for the prognosis of structural damage. In this paper, we discuss an improved method for the prognosis of structural damage due to flood impact. Detailed correlations between inundation level and flow velocities depending on the vulnerability of the building types, as well as the number of storeys, are considered. Because reliable damage data from events with high flow velocities were not available, an innovative approach was adopted to cover a wide range of flow velocities. The proposed approach combines comprehensive damage data collected after the 2002 flood in Germany with damage data of the 2011 Tohoku earthquake tsunami in Japan. The application of the developed methods enables a reliable reinterpretation of the structural damage caused by the August flood of 2002 in six study areas in the Free State of Saxony.
In ten chapters, this thesis presents information retrieval technology which is tailored to the research activities that arise in the context of corpus-based digital humanities projects.
The presentation is structured by a conceptual research process that is introduced in Chapter 1. The process distinguishes a set of five research activities: research question generation, corpus acquisition, research question modeling, corpus annotation, and result dissemination. Each of these research activities elicits different information retrieval tasks with special challenges, for which algorithmic approaches are presented after an introduction of the core information retrieval concepts in Chapter 2.
A vital concept in many of the presented approaches is the keyquery paradigm introduced in Chapter 3, which represents an operation that returns relevant search queries in response to a given set of input documents. Keyqueries are proposed in Chapter 4 for the recommendation of related work, and in Chapter 5 for improving access to aspects hidden in the long tail of search result lists.
With pseudo-descriptions, a document expansion approach is presented in Chapter 6. The approach improves the retrieval performance for corpora where only bibliographic meta-data is originally available. In Chapter 7, the keyquery paradigm is employed to generate dynamic taxonomies for corpora in an unsupervised fashion.
Chapter 8 turns to the exploration of annotated corpora, and presents scoped facets as a conceptual extension to faceted search systems, which is particularly useful in exploratory search settings. For the purpose of highlighting the major topical differences in a sequence of sub-corpora, an algorithm called topical sequence profiling is presented in Chapter 9.
The thesis concludes with two pilot studies regarding the visualization of (re)search results for the means of successful result dissemination: a metaphoric interpretation of the information nutrition label, as well as the philosophical bodies, which are 3D-printed search results.
Bolted connections are widely employed in structures like transmission poles, wind turbines, and television (TV) towers. The behaviour of bolted connections is often complex and plays a significant role in the overall dynamic characteristics of the structure. The goal of this work is to conduct a fatigue lifecycle assessment of such a bolted connection block of a 193 m tall TV tower, for which 205 days of real measurement data have been obtained from the installed monitoring devices. Based on the recorded data, the best-fit stochastic wind distribution for 50 years, the decisive wind action, and the locations to carry out the fatigue analysis have been decided. A 3D beam model of the entire tower is developed to extract the nodal forces corresponding to the connection block location under various mean wind speeds, which is later coupled with a detailed complex finite element model of the connection block, with over three million degrees of freedom, for acquiring stress histories on some pre-selected bolts. The random stress histories are analysed using the rainflow counting algorithm (RCA) and the damage is estimated using Palmgren-Miner's damage accumulation law. A modification is proposed to integrate the loading sequence effect into the RCA, which otherwise is ignored, and the differences between the two RCAs are investigated in terms of the accumulated damage.
In this paper, we present an open-source code for the first-order and higher-order nonlocal operator method (NOM) including a detailed description of the implementation. The NOM is based on so-called support, dual-support, nonlocal operators, and an operate energy functional ensuring stability. The nonlocal operator is a generalization of the conventional differential operators. Combined with the method of weighed residuals and variational principles, NOM establishes the residual and tangent stiffness matrix of operate energy functional through some simple matrix without the need of shape functions as in other classical computational methods such as FEM. NOM only requires the definition of the energy drastically simplifying its implementation. The implementation in this paper is focused on linear elastic solids for sake of conciseness through the NOM can handle more complex nonlinear problems. The NOM can be very flexible and efficient to solve partial differential equations (PDEs), it’s also quite easy for readers to use the NOM and extend it to solve other complicated physical phenomena described by one or a set of PDEs. Finally, we present some classical benchmark problems including the classical cantilever beam and plate-with-a-hole problem, and we also make an extension of this method to solve complicated problems including phase-field fracture modeling and gradient elasticity material.
This dataset presents the numerical analysis of the heat and moisture transport through a facade equipped with a living wall system designated for greywater treatment. While such greening systems provide many environmental benefits, they involve pumping large quantities of water onto the wall assembly, which can increase the risk of moisture in the wall as well as impaired energetic performance due to increased thermal conductivity with increased moisture content in the building materials. This dataset was acquired through numerical simulation using the coupling of two simulation tools, namely Envi-Met and Delphin. This coupling was used to include the complex role the plants play in shaping the near-wall environmental parameters in the hygrothermal simulations. Four different wall assemblies were investigated, each assembly was assessed twice: with and without the living wall. The presented data include the input and output parameters of the simulations, which were presented in the co-submitted article [1].
Reine Calciumsulfatbindemittel weisen eine hohe Löslichkeit auf. Feuchteinwirkung führt zudem zu starken Festigkeitsverlusten. Aus diesem Grund werden diese Bindemittel ausschließlich für Baustoffe und -produkte im Innenbereich ohne permanenten Feuchtebeanspruchung eingesetzt. Eine Möglichkeit, die Feuchtebeständigkeit zu erhöhen, ist die Beimischung puzzolanischer und zementärer Komponenten. Diese Mischsysteme werden Gips-Zement-Puzzolan-Bindemittel (kurz: GZPB) genannt.
Mischungen aus Calciumsulfaten und Portlandzementen allein sind aufgrund der treibenden Ettringitbildung nicht raumbeständig. Durch die Zugabe von puzzolanischen Stoffen können aber Bedingungen im hydratisierenden System geschaffen werden, welche eine rissfreie Erhärtung ermöglichen. Hierfür ist eine exakte Rezeptierung der GZPB notwendig, um die GZPB-typischen, ettringitbedingten Dehnungen zeitlich zu begrenzen. Insbesondere bei alumosilikatischen Puzzolanen treten während der Hydratation gegenüber rein silikatischen Puzzolanen deutlich höhere Expansionen auf, wodurch die Gefahr einer potenziellen Rissbildung steigt. Für die Erstellung geeigneter GZPB-Zusammensetzungen bedarf es daher einer Methodik, um raumbeständig erhärtende Systeme sicher von destruktiven Mischungen unterscheiden zu können.
Sowohl für die Rezeptierung als auch für die Anwendung der GZPB existieren in Deutschland keinerlei Normen. Darüber hinaus sind die Hydratationsvorgänge sowie die entstehenden Produkte nicht konsistent beschrieben. Auch auf die Besonderheiten der GZPB mit alumosilikatischen Puzzolanen wird in der Literatur nur unzureichend eingegangen.
Ziel war es daher, ein grundlegendes Verständnis der Hydratation sowie eine sichere Methodik zur Rezeptierung raumbeständig und rissfrei erhärtender GZPB, insbesondere in Hinblick auf die Verwendung alumosilikatischer Puzzolane, zu erarbeiten. Darüber hinaus sollte systematisch der Einfluss der Einzelkomponenten auf Hydratation und Eigenschaften dieser Bindemittelsysteme untersucht werden. Dies soll ermöglichen, die GZPB für ein breites Anwendungsspektrum als Bindemittel zu etablieren, und somit vorteilhafte Eigenschaften der Calciumsulfate (geringe Schwindneigung, geringe CO2-Emission etc.) mit der Leistungs-fähigkeit von Zementen (Wasserbeständigkeit, Festigkeit, Dauerhaftigkeit etc.) zu verbinden.
Als Ausgangsstoffe der Untersuchungen zu den GZPB wurden Stuckgips und Alpha-Halbhydrat als Calciumsulfatbindemittel in unterschiedlichen Anteilen im GZPB verwendet. Die Puzzolan-Zement-Verhältnisse wurden ebenfalls variiert. Als Puzzolan kam für den Großteil der Untersuchungen ein alumosilikatisches Metakaolin zum Einsatz. Als kalkspendende Komponente diente ein reiner Portlandzement.
Das Untersuchungsprogramm gliederte sich in 4 Teile. Zuerst wurde anhand von CaO- und pH-Wert-Messungen in Suspensionen sowie dem Längenänderungsverhalten von Bindemittelleimen verschiedener Zusammensetzungen eine Vorauswahl geeigneter GZPB-Rezepturen ermittelt. Danach erfolgten, ebenfalls an Bindemittelleimen, Untersuchungen zu den Eigenschaften der als geeignet eingeschätzten GZPB-Mischungen. Hierzu zählten Langzeitbetrachtungen zur rissfreien Erhärtung bei unterschiedlichen Umgebungsbedingungen sowie die Festigkeitsentwicklung im trockenen und feuchten Zustand. Im nächsten Schritt wurde anhand zweier exemplarischer GZPB-Zusammensetzungen (mit silikatischen und alumosilikatischen Puzzolan) die prinzipiell mögliche Phasenzusammensetzung unter Variation des Puzzolan-Zement-Verhältnisses (P/Z-Verhältnis) und des Calciumsulfatanteils im thermodynamischen Gleichgewichtszustand berechnet. Hier wurde im Besonderen auf Unterschiede der silikatischen und alumosilikatischen Puzzolane eingegangen. Im letzten Teil der Untersuchungen wurden die Hydratationskinetik der GZPB sowie die Gefügeentwicklung näher betrachtet. Hierfür wurden die Porenlösungen chemisch analysiert und Sättigungsindizes berechnet, sowie elektronenmikropische, porosimetrische und röntgenografische Untersuchungen durchgeführt. Abschließend wurden die Ergebnisse gesamtheitlich interpretiert, da die Ergebnisse der einzelnen Untersuchungsprogramme miteinander in Wechselwirkung stehen.
Als hauptsächliche Hydratationsprodukte wurden Calciumsulfat-Dihydrat, Ettringit und
C-(A)-S-H-Phasen ermittelt, deren Anteile im GZPB neben dem Calciumsulfatanteil und dem Puzzolan-Zement-Verhältnis auch deutlich vom Wasserangebot und der Gefügeentwicklung abhängen. Bei Verwendung von alumosilikatischen Puzzolans kommt es wahrscheinlich zur teilweisen Substitution des Siliciums durch Aluminium in den C-S-H-Phasen. Dies erscheint aufgrund des Nachweises der für diese Phasen typischen, folienartigen Morphologie wahrscheinlich. Portlandit wurde in raumbeständigen GZPB-Systemen nur zu sehr frühen Zeitpunkten in geringen Mengen gefunden.
In den Untersuchungen konnte ein Teil der in der Literatur beschriebenen, prinzipiellen Hydratationsabläufe bestätigt werden. Bei Verwendung von Halbhydrat als Calciumsulfatkomponente entsteht zuerst Dihydrat und bildet die Primärstruktur der GZPB. In dieses existierende Grundgefüge kristallisieren dann das Ettringit und die C-(A)-S-H-Phasen. In den GZPB sorgen entgegen der Beschreibungen in der Literatur nicht ausschließlich die
C-(A)-S-H-Phasen zur Verbesserung der Feuchtebeständigkeit und der Erhöhung des Festigkeitsniveaus, sondern auch das Ettringit. Beide Phasen überwachsen im zeitlichen Verlauf der Hydratation die Dihydratkristalle in der Matrix und hüllen diese – je nach Calciumsulfatanteil im GZPB – teilweise oder vollständig ein. Diese Umhüllung sowie die starke Gefügeverdichtung durch die C-(A)-S-H-Phasen und das Ettringit bedingen, dass ein lösender Angriff durch Wasser erschwert oder gar verhindert wird. Gleichzeitig wird die Gleitfähigkeit an den Kontaktstellen der Dihydratkristalle verringert.
Eine rissfreie und raumbeständige Erhärtung ist für die gefahrlose Anwendung eines GZPB-Systems essentiell. Hierfür ist die Kinetik der Ettringitbildung von elementarer Bedeutung. Die gebildete Ettringitmenge spielt nur eine untergeordnete Rolle. Selbst ausgeprägte, ettringitbedingte Dehnungen und hohe sich bildende Mengen führen zu frühen Zeitpunkten, wenn die Dihydratkristalle noch leicht gegeneinander verschiebbar sind, zu keinen Schäden. Bleibt die Übersättigung bezüglich Ettringit und somit auch der Kristallisationsdruck allerdings über einen langen Zeitraum hoch, genügen bereits geringe Ettringitmengen, um das sich stetig verfestigende Gefüge stark zu schädigen.
Die für die raumbeständige Erhärtung der GZPB notwendige, schnelle Abnahme der Ettringitübersättigung wird hauptsächlich durch die Reaktivität des Puzzolans beeinflusst. Die puzzolanische Reaktion führt zur Bindung des aus dem Zement stammenden Calciumhydroxid durch die Bildung von C-(A)-S-H-Phasen und Ettringit. Hierdurch sinkt die Calcium- und Hydroxidionenkonzentration in der Porenlösung im Verlauf der Hydratation, wodurch auch die Übersättigung bezüglich Ettringit abnimmt. Je höher die Reaktivität des Puzzolans ist, desto schneller sinkt der Sättigungsindex des Ettringits und somit auch der Kristallisationsdruck. Nach Unterschreiten eines noch näher zu klärendem Grenzwert der Übersättigung stagnieren die Dehnungen. Das Ettringit kristallisiert bzw. wächst nun bevorzugt in den Poren ohne eine weitere, äußere Volumenzunahme zu verursachen.
Um eine schadensfreie Erhärtung des GZPB zu gewährleisten, muss gerade in der frühen Phase der Hydratation ein ausreichendes Wasserangebot gewährleistet werden, so dass die Ettringitbildung möglichst vollständig ablaufen kann. Andernfalls kann es bei einer Wiederbefeuchtung zur Reaktivierung der Ettringitbildung kommen, was im eingebauten Zustand Schäden verursachen kann. Die Gewährleistung eines ausreichenden Wasserangebots ist im GZPB-System nicht unproblematisch. In Abhängigkeit der GZPB-Zusammensetzung können sich große Ettringitmengen bilden, die einen sehr hohen Wasserbedarf aufweisen. Deshalb kann es, je nach verwendeten Wasser-Bindemittel-Wert, im Bindemittelleim zu einem Wassermangel kommen, welcher die weitere Hydratation verlangsamt bzw. komplett verhindert. Zudem können GZPB-Systeme teils sehr dichte Gefüge ausbilden, wodurch der Wassertransport zum Reaktionsort des Ettringits zusätzlich behindert wird.
Die Konzeption raumbeständiger GZPB-Systeme muss anhand mehrerer aufeinander aufbauender Untersuchungen erfolgen. Zur Vorauswahl geeigneter Puzzolan-Zementverhältnisse eignen sich die Messungen der CaO-Konzentration und des pH-Wertes in Suspensionen. Als alleinige Beurteilungsgrundlage reicht dies allerdings nicht aus. Zusätzlich muss das Längenänderungs-verhalten beurteilt werden. Raumbeständige Mischungen mit alumosilikatischen Puzzolanen zeigen zu frühen Zeitpunkten starke Dehnungen, welche dann abrupt stagnieren. Stetige – auch geringe – Dehnungen weisen auf eine destruktive Zusammensetzung hin.
Mit diesem mehrstufigen Vorgehen können raumbeständige, stabile GZPB-Systeme konzipiert werden, so dass die Zielstellung der Arbeit erreicht wurde und ein sicherer praktischer Einsatz dieser Bindemittelart gewährleistet werden kann.
This report details the development of Horoskopos, a virtual planetarium for astrology. This project was an attempt to develop a learning tool for studying astrological concepts as connected to observational astronomy. The premise that astrology and observational astronomy were once inseparable from each other in ancient times guided the conceptualization of this tool as an interactive planetarium. The main references were existing software and applications for visualization in astrology and astronomy. Professional astrology teachers were consulted in order to understand better the state of astrological teaching and learning, as well as existing tools and practice. Horoskopos was built using the Unity3D development interface, which is based on the C# programming language. It also relied on the Swiss Ephemeris coding interface from Astrodienst. The development process was experimental and many of the needed skills were developed as needed. Usability tests were performed as new features were added to the interface. The final version of Horoskopos is fully usable, with many interactive visualization features and a defined visual identity. It was validated together with professional astrologers for its effectiveness in concept and visualization.
Hitze in der Stadt Jena
(2022)
Die vorliegende Arbeit befasst sich mit den spezifischen Faktoren und Wechselwirkungen des städtischen Klimas und Strategien zur Prävention und Kompensation lokaler Klimaveränderungen. Problematische Merkmale des Stadtklimas werden sich infolge des Klimawandels stärker ausprägen. Insbesondere die Hitzebelastung wird zunehmen und die Lebensbedingungen in der Stadt negativ beeinflussen. Infolge höherer Temperaturen in Städten und einer höheren Temperaturdifferenz zum Umland verändern sich Windströme und die Wasserbilanz. Es sind Strategien notwendig, um den Schadstoffausstoß, die Flächeninanspruchnahme, die Abfallproduktion und den Wasser-, Energie- und Ressourcenverbrauch zu verringern, um sowohl langfristig den Klimawandel als auch dessen bereits unvermeidbaren Auswirkungen auf Städte zu begrenzen.
Beispielhaft untersucht die Arbeit das Stadtklima, dessen zukünftige Veränderungen infolge des Klimawandels, bauliche Maßnahmen und Anpassungsstrategien der Stadt Jena. Jena ist die zweitgrößte Stadt im Bundesland Thüringen und gehört heute zu den wärmsten und trockensten Großstädten Deutschlands.
Die Ergebnisse der Arbeit werden anschließend anhand eines städtebaulichen Konzepts und Entwurfs angewendet. Das Bachstraßenareal liegt in der Innenstadt, dem am stärksten von Hitze betroffenen Stadtteil. Als ehemaliger Hauptstandort des Jenaer Universitätsklinikums, soll es zu einem nachhaltigen Wissenschaftscampus der Lebenswissenschaften umgebaut werden, wobei ein Großteil der denkmalgeschützten, ehemaligen Klinikgebäude erhalten bleibt. Der Fokus liegt dabei auf der Umsetzung der zuvor formulierten, nachhaltigen Strategien zur Verbesserung des lokalen Stadtklimas und einer Abschwächung der Auswirkungen des Klimawandels auf den besonders stark betroffenen Innenstadtbereich Jenas.
With the growing importance of the World Wide Web, the major challenges our society faces are also increasingly affecting the digital areas of our lives. Some of the associated problems can be addressed by computer science, and some of these specifically by data-driven research. To do so, however, requires to solve open issues related to archive quality and the large volume and variety of the data contained.
This dissertation contributes data, algorithms, and concepts towards leveraging the big data and temporal provenance capabilities of web archives to tackle societal challenges. We selected three such challenges that highlight the central issues of archive quality, data volume, and data variety, respectively:
(1) For the preservation of digital culture, this thesis investigates and improves the automatic quality assurance of the web page archiving process, as well as the further processing of the resulting archive data for automatic analysis.
(2) For the critical assessment of information, this thesis examines large datasets of Wikipedia and news articles and presents new methods for automatically determining quality and bias.
(3) For digital security and privacy, this thesis exploits the variety of content on the web to quantify the security of mnemonic passwords and analyzes the privacy-aware re-finding of the various seen content through private web archives.
Diese Dissertation untersucht Handlungsressourcen von zivilgesellschaftlichen Akteuren in Planungsprozessen um innerstädtische Planungsverfahren. Den theoretischen Rahmen bilden die Kapitalarten von Pierre Bourdieu, die zusammen mit dem Matrixraum von Dieter Läpple zu einem neuen Feldbegriff des ‚Raumfeldes‘ zusammengeführt und operationalisiert wurden. Es handelt sich um eine qualitative Arbeit, die zwischen Stadtsoziologie und Urbanistik zu verorten ist. Als Fallbeispiele wurde die Erweiterung des Berliner Mauerparks sowie das Baugebiet „So! Berlin“ in Berlin gewählt.
Zu den diversen Unternehmungen sozialbewegter „Gegenwissenschaft“, die um 1980 auf der Bildfläche der BRD erschienen, zählte der 1982 gegründete Berliner Wissenschaftsladen e. V., kurz WILAB – eine Art „alternatives“ Spin-off der Technischen Universität Berlin. Der vorliegende Beitrag situiert die Ausgründung des „Ladens“ im Kontext zeitgenössischer Fortschritte der (regionalen) Forschungs- und Technologiepolitik. Gezeigt wird, wie der deindustrialisierenden Inselstadt, qua „innovationspolitischer“ Gegensteuerung, dabei sogar eine gewisse Vorreiterrolle zukam: über die Stadtgrenzen hinaus sichtbare Neuerungen wie die Gründermesse BIG TECH oder das 1983 eröffnete Berliner Innovations- und Gründerzentrum (BIG), der erste „Incubator“ [sic] der BRD, etwa gingen auf das Konto der 1977/78 lancierten Technologie-Transferstelle der TU Berlin, TU-transfer.
Anders gesagt: tendenziell bekam man es hier nun mit Verhältnissen zu tun, die immer weniger mit den Träumen einer „kritischen“, nicht-fremdbestimmten (Gegen‑)Wissenschaft kompatibel waren. Latent konträr zur historiographischen Prominenz des wissenschaftskritischen Zeitgeists fristeten „alternativen“ Zielsetzungen verpflichtete Unternehmungen wie „WILAB“ ein relativ marginalisiertes Nischendasein. Dennoch wirft das am WILAB verfolgte, so gesehen wenig aussichtsreiche Anliegen, eine andere, nämlich „humanere“ Informationstechnologie in die Wege zu leiten, ein instruktives Licht auf die Aufbrüche „unternehmerischer“ Wissenschaft in der BRD um 1980.
The detailed structural analysis of thin-walled circular pipe members often requires the use of a shell or solid-based finite element method. Although these methods provide a very good approximation of the deformations, they require a higher degree of discretization which causes high computational costs. On the other hand, the analysis of thin-walled circular pipe members based on classical beam theories is easy to implement and needs much less computation time, however, they are limited in their ability to approximate the deformations as they cannot consider the deformation of the cross-section.
This dissertation focuses on the study of the Generalized Beam Theory (GBT) which is both accurate and efficient in analyzing thin-walled members. This theory is based on the separation of variables in which the displacement field is expressed as a combination of predetermined deformation modes related to the cross-section, and unknown amplitude functions defined on the beam's longitudinal axis. Although the GBT was initially developed for long straight members, through the consideration of complementary deformation modes, which amend the null transverse and shear membrane strain assumptions of the classical GBT, problems involving short members, pipe bends, and geometrical nonlinearity can also be analyzed using GBT. In this dissertation, the GBT formulation for the analysis of these problems is developed and the application and capabilities of the method are illustrated using several numerical examples. Furthermore, the displacement and stress field results of these examples are verified using an equivalent refined shell-based finite element model.
The developed static and dynamic GBT formulations for curved thin-walled circular pipes are based on the linear kinematic description of the curved shell theory. In these formulations, the complex problem in pipe bends due to the strong coupling effect of the longitudinal bending, warping and the cross-sectional ovalization is handled precisely through the derivation of the coupling tensors between the considered GBT deformation modes. Similarly, the geometrically nonlinear GBT analysis is formulated for thin-walled circular pipes based on the nonlinear membrane kinematic equations. Here, the initial linear and quadratic stress and displacement tangent stiffness matrices are built using the third and fourth-order GBT deformation mode coupling tensors.
Longitudinally, the formulation of the coupled GBT element stiffness and mass matrices are presented using a beam-based finite element formulation. Furthermore, the formulated GBT elements are tested for shear and membrane locking problems and the limitations of the formulations regarding the membrane locking problem are discussed.
Multi-criteria decision analysis (MCDA) is an established methodology to support the decision-making of multi-objective problems. For conducting an MCDA, in most cases, a set of objectives (SOO) is required, which consists of a hierarchical structure comprised of objectives, criteria, and indicators. The development of an SOO is usually based on moderated development processes requiring high organizational and cognitive effort from all stakeholders involved. This article proposes elementary interactions as a key paradigm of an algorithm-driven development process for an SOO that requires little moderation efforts. Elementary interactions are self-contained information requests that may be answered with little cognitive effort. The pairwise comparison of elements in the well-known analytical hierarchical process (AHP) is an example of an elementary interaction. Each elementary interaction in the development process presented contributes to the stepwise development of an SOO. Based on the hypothesis that an SOO may be developed exclusively using elementary interactions (EIs), a concept for a multi-user platform is proposed. Essential components of the platform are a Model Aggregator, an Elementary Interaction Stream Generator, a Participant Manager, and a Discussion Forum. While the latter component serves the professional exchange of the participants, the first three components are intended to be automatable by algorithms. The platform concept proposed has been evaluated partly in an explorative validation study demonstrating the general functionality of the algorithms outlined. In summary, the platform concept suggested demonstrates the potential to ease SOO development processes as the platform concept does not restrict the application domain; it is intended to work with little administration moderation efforts, and it supports the further development of an existing SOO in the event of changes in external conditions. The algorithm-driven development of SOOs proposed in this article may ease the development of MCDA applications and, thus, may have a positive effect on the spread of MCDA applications.
Multi-criteria decision analysis (MCDA) is an established methodology to support the decision-making of multi-objective problems. For conducting an MCDA, in most cases, a set of objectives (SOO) is required, which consists of a hierarchical structure comprised of objectives, criteria, and indicators. The development of an SOO is usually based on moderated development processes requiring high organizational and cognitive effort from all stakeholders involved. This article proposes elementary interactions as a key paradigm of an algorithm-driven development process for an SOO that requires little moderation efforts. Elementary interactions are self-contained information requests that may be answered with little cognitive effort. The pairwise comparison of elements in the well-known analytical hierarchical process (AHP) is an example of an elementary interaction. Each elementary interaction in the development process presented contributes to the stepwise development of an SOO. Based on the hypothesis that an SOO may be developed exclusively using elementary interactions (EIs), a concept for a multi-user platform is proposed. Essential components of the platform are a Model Aggregator, an Elementary Interaction Stream Generator, a Participant Manager, and a Discussion Forum. While the latter component serves the professional exchange of the participants, the first three components are intended to be automatable by algorithms. The platform concept proposed has been evaluated partly in an explorative validation study demonstrating the general functionality of the algorithms outlined. In summary, the platform concept suggested demonstrates the potential to ease SOO development processes as the platform concept does not restrict the application domain; it is intended to work with little administration moderation efforts, and it supports the further development of an existing SOO in the event of changes in external conditions. The algorithm-driven development of SOOs proposed in this article may ease the development of MCDA applications and, thus, may have a positive effect on the spread of MCDA applications.
The theme of this project is the colonial history of the natural rubber industry. It focuses on two species of tropical plants: Ficus elastica and Hevea brasiliensis. Geographically their native habitat is very distant from each other, but they connect by European influence through the exploitation of latex.
The many forms and outcomes from this work manifest the attempt of the artist to create an association between a common household plant, the origin of its name, and the source of rubber. As a ghostly connective tissue, the latex surrounds reconstructed history, old prints, live plants, and drawings, accepting the material's capacity to both erase and preserve the past.
Das Fahrrad ist ein Medium sozialer Veränderung. Seine vielfältigen utopischen Potenziale ergeben sich nicht zuletzt aus seinen ebenso vielfältigen und häufig übersehenen medialen Qualitäten: Es vermittelt, es verbindet, es übersetzt; es modifiziert Wahrnehmung und Organisation von Raum und Zeit, von Körpern und von Sozialität. Umgekehrt kann auch das medienwissenschaftliche Denken fahrradmedial verändert werden. Das Fahrrad ist nicht nur Medium des sozialen und ökologischen Wandels: Radfahren eröffnet Perspektiven, verändert Räume, lässt neue Relationen entstehen und teilt Handlungsmacht neu auf.
Fahrradutopien denkt vom Fahrrad aus und ergänzt dabei bestehende Ansätze zur Mobilitätsforschung um medienkulturwissenschaftliche Perspektiven. Die Beiträge verbinden Medienwissenschaften und Forschungen zu Fahrradaktivismus mit der Liebe zum Radfahren. Fokussiert werden Fahrradfilme und -vlogs, Verkehr und Infrastrukturen, Virtuelle Realität und Fahrrad, Fahrradkollektive und Fahrradfeminismus.
Der Aufruf, die Begriffe Stadt und Kritik in das Zentrum einer Debatte zu stellen, bietet die große Chance, uns weit über begriffliche Klärungen unseres gemeinsamen Arbeitsgegenstands hinaus – die ja auch für sich selbst sehr fruchtbar sein können – über die Funktion zu verständigen, die wir in der Gesellschaft ausüben, wenn wir räumliche Planung praktizieren, erforschen und lehren. Da in der Bundesrepublik nicht nur ein großer Bedarf, sondern auch eine beträchtliche Nachfrage nach öffentlicher Planung besteht und die planungsbezogenen Wissenschaften sich eines insgesamt stabilen institutionellen Standes erfreuen, laufen wir Gefahr, die gesellschaftspolitische Legitimation von Berufsfeld und Wissenschaft zu vernachlässigen, sie als gegeben zu behandeln. Wir müssen uns ja kaum rechtfertigen.
The seismic vulnerability assessment of existing reinforced concrete (RC) buildings is a significant source of disaster mitigation plans and rescue services. Different countries evolved various Rapid Visual Screening (RVS) techniques and methodologies to deal with the devastating consequences of earthquakes on the structural characteristics of buildings and human casualties. Artificial intelligence (AI) methods, such as machine learning (ML) algorithm-based methods, are increasingly used in various scientific and technical applications. The investigation toward using these techniques in civil engineering applications has shown encouraging results and reduced human intervention, including uncertainties and biased judgment. In this study, several known non-parametric algorithms are investigated toward RVS using a dataset employing different earthquakes. Moreover, the methodology encourages the possibility of examining the buildings’ vulnerability based on the factors related to the buildings’ importance and exposure. In addition, a web-based application built on Django is introduced. The interface is designed with the idea to ease the seismic vulnerability investigation in real-time. The concept was validated using two case studies, and the achieved results showed the proposed approach’s potential efficiency
Ernährung bestimmt unser tägliches Leben. Sie erfüllt in erster Linie die physiologische Notwendigkeit unseren Körper am Leben zu halten und ist gleichzeitig Alltagspraxis, durch welche gesamtgesellschaftliche Strukturen sichtbar werden. Innerhalb dieser Alltagspraxen erfüllt Ernährung vor allem eine wichtige Funktion in der Herstellung gesellschaftlicher Teilhabe oder kann strukturelle Ausschlüsse und soziale Ungleichheit bedingen. Dem Wohlfahrtsregime kommt somit eine wichtige Aufgabe in der Grundversorgung der Bevölkerung zu und muss innerhalb der Daseinsvorsorge auf Ausschlüsse von ernährungsbezogener Teilhabe einzelner Bevölkerungsschichten eingehen und sozialer Ungleichheit entgegenwirken. In diesem Working Paper soll der Fragestellung nachgegangen werden, inwiefern Teilhabe bzw. strukturelle Ausschlüsse von Ernährung innerhalb des bundesdeutschen Wohlfahrtsregimes hergestellt werden und durch welche politischen Praktiken und Forderungen aus der Zivilgesellschaft bzw. sozialen Bewegungen ernährungsvermittelte Teilhabe (wieder) hergestellt wird.
This dataset consists mainly of two subsets. The first subset includes measurements and simulation data conducted to validate the simulation tool ENVI-met. The measurements were conducted at the campus of the Bauhaus-University Weimar in Weimar, Germany and consisted of recording exterior air temperature, globe temperature, relative humidity, and wind velocity at 1.5 m at four points on four different days. After the measurements, the geometry of the campus was modelled and meshed; the simulations were conducted using the weather data of the measurements days with the aim of investigating the accuracy of the model.
The second data subset consists of ENVI-met simulation data of the potential of facade greening in improving the outdoor environment and the indoor air temperature during heatwaves in Central European cities. The data consist of the boundary conditions and the simulation output of two simulation models: with and without facade greening. The geometry of the models corresponded to a residential buildings district in Stuttgart, Germany. The simulation output consisted of exterior air temperature, mean radiant temperature, relative humidity, and wind velocity at 12 different probe points in the model in addition to the indoor air temperature of an exemplary building. The dataset presents both vertical profiles of the probed parameters as well as the time series output of the five-day simulation duration. Both data subsets correspond to the investigations presented in the co-submitted article [1].
Kleine Kommunen im ländlichen Raum sind aufgrund ihrer oft eingeschränkten personellen und finanziellen Kapazitäten bisher eher sporadisch in den Themenfeldern Energieeffizienz und Erneuerbare Energien aktiv. Immer wieder stellt sich daher Frage, wie die Klimaschutzstrategien des Bundes und der Länder dort mit dem verfügbaren Personal kostengünstig realisierbar sind. Vor diesem Hintergrund wird ein Werkzeug entwickelt, mit dessen Hilfe der aktive Einstieg in diese Thematik mit geringen Aufwand und überwiegend barrierefrei möglich ist.
Der Aufbau eines prozessorientierten Entwicklungs- und Moderationsmodells zur Erprobung und Umsetzung bezahlbarer Handlungsoptionen für Energieeinsparungen und effizienten Energieeinsatz im überwiegend ländlichen geprägten Raum ist der Schwerpunkt der Softwarelösung.
Kommunen werden mit deren Hilfe in die Lage versetzt, in die notwendigen Prozesse der Energie- und Wärmewende einzusteigen. Dabei soll der modulare Aufbau die regulären Schritte notwendiger (integrierter) Planungsprozesse nicht vollständig ersetzen. Vielmehr können innerhalb der Online-Anwendung - überwiegend automatisiert - konkrete Maßnahmenvorschläge erstellt werden, die ein solides Fundament der künftigen energetischen Entwicklung der Kommunen darstellen.
Für eine gezielte Validierung der Ergebnisse und der Ableitung potentieller Maßnahmen werden für die Erprobung Modellkommunen in Thüringen, Bayern und Hessen als Reallabore einbezogen.
Das Tool steht bisher zunächst nur den beteiligten Modellkommunen zur Verfügung. Die entwickelte Softwarelösung soll künftig Schritt für Schritt allen interessierten Kommunen mit diversen Hilfsmitteln und einer Vielzahl anderer praktischer Bestandteile zur Verfügung gestellt werden.
El paisaje de la Cuenca Lechera Central Argentina: la huella de la producción sobre el territorio
(2022)
In recent times, the study of landscape heritage acquires value by virtue of becoming an alternative to rethink regional development, especially from the point of view of territorial planning. In this sense, the Central Argentine Dairy Basin (CADB) is presented as a space where the traces of different human projects have accumulated over centuries of occupation, which can be read as heritage. The impact of dairy farming and other productive activities has shaped the configuration of its landscape. The main hypothesis assumed that a cultural landscape would have been formed in the CADB, whose configuration would have depended to a great extent on the history of productive activities and their deployment over the territory, and this same history would hide the keys to its alternative.
The thesis approached the object of study from descriptive and cartographic methods that placed the narration of the history of territory and the resources of the landscape as a discursive axis. A series of intentional readings of the territory and its constituent parts pondered the layers of data that have accumulated on it in the form of landscape traces, with the help of an approach from complementary dimensions (natural, sociocultural, productive, planning). Furthermore, the intersection of historical sources was used in order to allow the construction of the territorial story and the detection of the origin of the landscape components. A meticulous cartographic work also helped to spatialise the set of phenomena and elements studied, and was reflected in a multiscalar scanning.
Inhaltlich beschäftigt sich die Arbeit, die im Rahmen des Promotionsstudiengangs Kunst und Gestaltung an der Bauhaus-Universität entstand, mit der Erforschung sozio-interaktiver Potentiale der Videotelefonie im Kontext von Nähe und Verbundenheit mit Fokus auf Eigenbild, Embodiment sowie den Rederechtswechsel.
Die Videotelefonie als Kommunikationsform hat sich – und darauf deuten die Erfahrungen der Co- vid-19-Pandemie hin – im lebensweltlichen Alltag der Menschen etabliert und wird dort in naher Zukunft nicht mehr wegzudenken sein. Auf Basis ihrer Möglichkeiten und Errungenschaften ist es inzwischen Realität und Lebenswirklichkeit, dass die Kommunikation sowohl im privaten als auch im geschäftlichen Kontext mittels verschiedenster Kanäle stattfindet. Der Videotelefonie kommt hierbei als solche nicht nur eine tragende Funktion, sondern auch eine herausragende Rolle bei der vermeintlichen Reproduktion der Face-to-Face-Kommunikation im digitalen Raum zu und wird wie selbstverständlich zum zwischenmenschlichen Austausch genutzt. Just an diesem Punkt knüpft die Forschungsarbeit an. Zentral stand dabei das Vorhaben einer dezidierte Untersuchung des Forschungsgegenstandes Videotelefonie, sowohl aus Kultur- als auch Technikhistorischer, aber auch Medien-, Wahrnehmungs- wie Kommunikations- theoretischer Perspektive, indem analytische und phänosemiotische Perspektiven miteinander in Beziehung gesetzt werden (z.B. Wahrnehmungsbedingungen, Interaktionsmerkmale, realisierte Kommunikationsprozesse etc.). Damit verbundenes, wünschenswertes Ziel war es, eine möglichst zeitgemäße wie relevante Forschungsfrage zu adressieren, die neben den kulturellen Technisierungs- und Mediatisierungstendenzen in institutionellen und privaten Milieus ebenfalls eine conditio sine qua non der pandemischen (Massen-)Kommunikation entwirft.
Die Arbeit ist damit vor allem im Bereich des Produkt- und Interactiondesigns zu verorten. Darüber hinaus hatte sie das Ziel der Darlegung und Begründung der Videotelefonie als eigenständige Kommunikationsform, welche durch eigene, kommunikative Besonderheiten, die sich in ihrer jeweiligen Ingebrauchnahme sowie durch spezielle Wahrnehmungsbedingungen äußern, und die die Videotelefonie als »Rederechtswechselmedium« avant la lettre konsolidieren, gekennzeichnet ist. Dabei sollte der Beweis erbracht werden, dass die Videotelefonie nicht als Schwundstufe einer Kommunikation Face-to-Face, sondern als ein eigenständiges Mediatisierungs- und Kommunikationsereignis zu verstehen sei. Und eben nicht als eine beliebige – sich linear vom Telefon ausgehende – entwickelte Form der audio-visuellen Fernkommunikation darstellt, sondern die gestalterische (Bewegtbild-)Technizität ein eigenständiges Funktionsmaß offeriert, welches wiederum ein innovatives Kommunikationsmilieu im Kontext einer Rederechtswechsel-Medialität stabilisiert.
In machine learning, if the training data is independently and identically distributed as the test data then a trained model can make an accurate predictions for new samples of data. Conventional machine learning has a strong dependence on massive amounts of training data which are domain specific to understand their latent patterns. In contrast, Domain adaptation and Transfer learning methods are sub-fields within machine learning that are concerned with solving the inescapable problem of insufficient training data by relaxing the domain dependence hypothesis. In this contribution, this issue has been addressed and by making a novel combination of both the methods we develop a computationally efficient and practical algorithm to solve boundary value problems based on nonlinear partial differential equations. We adopt a meshfree analysis framework to integrate the prevailing geometric modelling techniques based on NURBS and present an enhanced deep collocation approach that also plays an important role in the accuracy of solutions. We start with a brief introduction on how these methods expand upon this framework. We observe an excellent agreement between these methods and have shown that how fine-tuning a pre-trained network to a specialized domain may lead to an outstanding performance compare to the existing ones. As proof of concept, we illustrate the performance of our proposed model on several benchmark problems.
The growing complexity of modern engineering problems necessitates development of advanced numerical methods. In particular, methods working directly with discrete structures, and thus, representing exactly some important properties of the solution on a lattice and not just approximating the continuous properties, become more and more popular nowadays. Among others, discrete potential theory and discrete function theory provide a variety of methods, which are discrete counterparts of the classical continuous methods for solving boundary value problems. A lot of results related to the discrete potential and function theories have been presented in recent years. However, these results are related to the discrete theories constructed on square lattices, and, thus, limiting their practical applicability and
potentially leading to higher computational costs while discretising realistic domains.
This thesis presents an extension of the discrete potential theory and discrete function theory to rectangular lattices. As usual in the discrete theories, construction of discrete operators is strongly influenced by a definition of discrete geometric setting. For providing consistent constructions throughout the whole thesis, a detailed discussion on the discrete geometric setting is presented in the beginning. After that, the discrete fundamental solution of the discrete Laplace operator on a rectangular lattice, which is the core of the discrete potential theory, its numerical analysis, and practical calculations are presented. By using the discrete fundamental solution of the discrete Laplace operator on a rectangular lattice, the discrete potential theory is then constructed for interior and exterior settings. Several discrete interior and exterior boundary value problems are then solved. Moreover, discrete transmission problems are introduced and several numerical examples of these problems are discussed. Finally, a discrete fundamental solution of the discrete Cauchy-Riemann operator on a rectangular lattice is constructed, and basics of the discrete function theory on a rectangular lattice are provided. This work indicates that the discrete theories provide
solution methods with very good numerical properties to tackle various boundary value problems, as well as transmission problems coupling interior and exterior problems. The results presented in this thesis provide a basis for further development of discrete theories on irregular lattices.
The aim of this study is controlling of spurious oscillations developing around discontinuous solutions of both linear and non-linear wave equations or hyperbolic partial differential equations (PDEs). The equations include both first-order and second-order (wave) hyperbolic systems. In these systems even smooth initial conditions, or smoothly varying source (load) terms could lead to discontinuous propagating solutions (fronts). For the first order hyperbolic PDEs, the concept of central high resolution schemes is integrated with the multiresolution-based adaptation to capture properly both discontinuous propagating fronts and effects of fine-scale responses on those of larger scales in the multiscale manner. This integration leads to using central high resolution schemes on non-uniform grids; however, such simulation is unstable, as the central schemes are originally developed to work properly on uniform cells/grids. Hence, the main concern is stable collaboration of central schemes and multiresoltion-based cell adapters. Regarding central schemes, the considered approaches are: 1) Second order central and central-upwind schemes; 2) Third order central schemes; 3) Third and fourth order central weighted non-oscillatory schemes (central-WENO or CWENO); 4) Piece-wise parabolic methods (PPMs) obtained with two different local stencils. For these methods, corresponding (nonlinear) stability conditions are studied and modified, as well. Based on these stability conditions several limiters are modified/developed as follows: 1) Several second-order limiters with total variation diminishing (TVD) feature, 2) Second-order uniformly high order accurate non-oscillatory (UNO) limiters, 3) Two third-order nonlinear scaling limiters, 4) Two new limiters for PPMs. Numerical results show that adaptive solvers lead to cost-effective computations (e.g., in some 1-D problems, number of adapted grid points are less than 200 points during simulations, while in the uniform-grid case, to have the same accuracy, using of 2049 points is essential). Also, in some cases, it is confirmed that fine scale responses have considerable effects on higher scales.
In numerical simulation of nonlinear first order hyperbolic systems, the two main concerns are: convergence and uniqueness. The former is important due to developing of the spurious oscillations, the numerical dispersion and the numerical dissipation. Convergence in a numerical solution does not guarantee that it is the physical/real one (the uniqueness feature). Indeed, a nonlinear systems can converge to several numerical results (which mathematically all of them are true). In this work, the convergence and uniqueness are directly studied on non-uniform grids/cells by the concepts of local numerical truncation error and numerical entropy production, respectively. Also, both of these concepts have been used for cell/grid adaptations. So, the performance of these concepts is also compared by the multiresolution-based method. Several 1-D and 2-D numerical examples are examined to confirm the efficiency of the adaptive solver. Examples involve problems with convex and non-convex fluxes. In the latter case, due to developing of complex waves, proper capturing of real answers needs more attention. For this purpose, using of method-adaptation seems to be essential (in parallel to the cell/grid adaptation). This new type of adaptation is also performed in the framework of the multiresolution analysis.
Regarding second order hyperbolic PDEs (mechanical waves), the regularization concept is used to cure artificial (numerical) oscillation effects, especially for high-gradient or discontinuous solutions. There, oscillations are removed by the regularization concept acting as a post-processor. Simulations will be performed directly on the second-order form of wave equations. It should be mentioned that it is possible to rewrite second order wave equations as a system of first-order waves, and then simulated the new system by high resolution schemes. However, this approach ends to increasing of variable numbers (especially for 3D problems).
The numerical discretization is performed by the compact finite difference (FD) formulation with desire feature; e.g., methods with spectral-like or optimized-error properties. These FD methods are developed to handle high frequency waves (such as waves near earthquake sources). The performance of several regularization approaches is studied (both theoretically and numerically); at last, a proper regularization approach controlling the Gibbs phenomenon is recommended.
At the end, some numerical results are provided to confirm efficiency of numerical solvers enhanced by the regularization concept. In this part, shock-like responses due to local and abrupt changing of physical properties, and also stress wave propagation in stochastic-like domains are studied.
Die Europatournee des Indischen Menaka-Balletts von 1936-38 ist der Ausgangspunkt dieser archivologischen Navigation entlang der Spuren indischer KünstlerInnen in Europa. In einer breit angelegten Archivrecherche wurden dazu Dokumente, Fundstücke, orale Erinnerungen und ethnografische Beobachtungen aus dem Kontext der Menaka-Tournee durch das nationalsozialistische Deutschland zusammengetragen.
Das Buch beschreibt den Rekonstruktionsprozess eines bedeutsamen Projekts der indischen Tanzmoderne. Es verfolgt dabei eine Methode, mit der sich die fragmentierten Dokumente des Medienereignisses als Spur lesen lassen und nutzt eine künstlerisch-forschende Involvierung in gegenwärtige Erinnerungspolitiken, in welche die verflochtenen Strukturen der künstlerischen Avantgarde zwischen Kolkata, Mumbai und Berlin hineinreichen. Die Spur des Menaka-Ballett erweist sich dabei als Teil weitreichender ideologischer, tänzerischer, musikalischer, filmischer und literarischer Strömungen, die auch in gegenwärtigen kulturellen Bestimmungen fortwirken.
Fotografien, Zeitungsberichte, Film- und Tonaufnahmen, Briefe und persönliche Erinnerungstücke erzählen davon, wie sich, vor dem Hintergrund der im antikolonialen Aufbruch befindlichen Kulturreform in Indien, und der nationsozialistisch-völkischen Kulturpolitik in Deutschland, die Tänzerinnen und Musiker der indischen Ballettgruppe und die deutsche Öffentlichkeit im gegenseitigen Spiegel betrachteten, während die Vorzeichen des kommenden Krieges immer deutlicher wurden.
Isogeometric analysis (IGA) is a numerical method for solving partial differential equations (PDEs), which was introduced with the aim of integrating finite element analysis with computer-aided design systems. The main idea of the method is to use the same spline basis functions which describe the geometry in CAD systems for the approximation of solution fields in the finite element method (FEM). Originally, NURBS which is a standard technology employed in CAD systems was adopted as basis functions in IGA but there were several variants of IGA using other technologies such as T-splines, PHT splines, and subdivision surfaces as basis functions. In general, IGA offers two key advantages over classical FEM: (i) by describing the CAD geometry exactly using smooth, high-order spline functions, the mesh generation process is simplified and the interoperability between CAD and FEM is improved, (ii) IGA can be viewed as a high-order finite element method which offers basis functions with high inter-element continuity and therefore can provide a primal variational formulation of high-order PDEs in a straightforward fashion. The main goal of this thesis is to further advance isogeometric analysis by exploiting these major advantages, namely precise geometric modeling and the use of smooth high-order splines as basis functions, and develop robust computational methods for problems with complex geometry and/or complex multi-physics.
As the first contribution of this thesis, we leverage the precise geometric modeling of isogeometric analysis and propose a new method for its coupling with meshfree discretizations. We exploit the strengths of both methods by using IGA to provide a smooth, geometrically-exact surface discretization of the problem domain boundary, while the Reproducing Kernel Particle Method (RKPM) discretization is used to provide the volumetric discretization of the domain interior. The coupling strategy is based upon the higher-order consistency or reproducing conditions that are directly imposed in the physical domain. The resulting coupled method enjoys several favorable features: (i) it preserves the geometric exactness of IGA, (ii) it circumvents the need for global volumetric parameterization of the problem domain, (iii) it achieves arbitrary-order approximation accuracy while preserving higher-order smoothness of the discretization. Several numerical examples are solved to show the optimal convergence properties of the coupled IGA–RKPM formulation, and to demonstrate its effectiveness in constructing volumetric discretizations for complex-geometry objects.
As for the next contribution, we exploit the use of smooth, high-order spline basis functions in IGA to solve high-order surface PDEs governing the morphological evolution of vesicles. These governing equations are often consisted of geometric PDEs, high-order PDEs on stationary or evolving surfaces, or a combination of them. We propose an isogeometric formulation for solving these PDEs. In the context of geometric PDEs, we consider phase-field approximations of mean curvature flow and Willmore flow problems and numerically study the convergence behavior of isogeometric analysis for these problems. As a model problem for high-order PDEs on stationary surfaces, we consider the Cahn–Hilliard equation on a sphere, where the surface is modeled using a phase-field approach. As for the high-order PDEs on evolving surfaces, a phase-field model of a deforming multi-component vesicle, which consists of two fourth-order nonlinear PDEs, is solved using the isogeometric analysis in a primal variational framework. Through several numerical examples in 2D, 3D and axisymmetric 3D settings, we show the robustness of IGA for solving the considered phase-field models.
Finally, we present a monolithic, implicit formulation based on isogeometric analysis and generalized-alpha time integration for simulating hydrodynamics of vesicles according to a phase-field model. Compared to earlier works, the number of equations of the phase-field model which need to be solved is reduced by leveraging high continuity of NURBS functions, and the algorithm is extended to 3D settings. We use residual-based variational multi-scale method (RBVMS) for solving Navier–Stokes equations, while the rest of PDEs in the phase-field model are treated using a standard Galerkin-based IGA. We introduce the resistive immersed surface (RIS) method into the formulation which can be employed for an implicit description of complex geometries using a diffuse-interface approach. The implementation highlights the robustness of the RBVMS method for Navier–Stokes equations of incompressible flows with non-trivial localized forcing terms including bending and tension forces of the vesicle. The potential of the phase-field model and isogeometric analysis for accurate simulation of a variety of fluid-vesicle interaction problems in 2D and 3D is demonstrated.
Design-related reassessment of structures integrating Bayesian updating of model safety factors
(2022)
In the semi-probabilistic approach of structural design, the partial safety factors are defined by considering some degree of uncertainties to actions and resistance, associated with the parameters’ stochastic nature. However, uncertainties for individual structures can be better examined by incorporating measurement data provided by sensors from an installed health monitoring scheme. In this context, the current study proposes an approach to revise the partial safety factor for existing structures on the action side, γE by integrating Bayesian model updating. A simple numerical example of a beam-like structure with artificially generated measurement data is used such that the influence of different sensor setups and data uncertainties on revising the safety factors can be investigated. It is revealed that the health monitoring system can reassess the current capacity reserve of the structure by updating the design safety factors, resulting in a better life cycle assessment of structures. The outcome is furthermore verified by analysing a real life small railway steel bridge ensuring the applicability of the proposed method to practical applications.
This study deals with design for AI/ML systems, more precisely in the industrial AI context based on case studies from the factory automation field. It therefore touches on core concepts from Human-Centered-Design (HCD), User Experience (UX) and Human Computer Interaction (HCI) on one hand, as well as concepts from Artificial Intelligence (AI), Machine Learning (ML) and the impact of technology on the other. The case studies the research is based on are within the industrial AI domain. However, the final outcomes, the findings, solutions, artifacts and so forth, should be transferable to a wider spectrum of domains. The study’s aim is to examine the role of designers in the age of AI and the factors which are relevant, based on the hypothesis that current AI/ML development lacks the human perspective, which means that there are pitfalls and challenges that design can help resolve. The initial literature review revealed that AI/ML are perceived as a new design material that calls for a new design paradigm. Additional research based on qualitative case study research was conducted to gain an overview of the relevant issues and challenges. From this, 17 themes emerged, which together with explorative expert interviews and a structured literature review, were further analyzed to produce the relevant HCD, UX and HCI themes. It became clear that designers need new processes, methods, and tools in the age of AI/ML in combination with not only design, but also data science and business expertise, which is why the proposed solution in this PhD features process modules for design, data science and business collaboration. There are seven process modules and their related activities and dependencies that serve as guidelines for practitioners who want to design intelligence. A unified framework for collecting use case exemplars was created, based on a workshop with different practitioners and researchers from the area of AI/ML to support and enrich the process modules with concrete projects examples.
Der vorliegende Handlungsleitfaden hilft zivilgesellschaftlichen Organisationen und staatlichen Einrichtungen bei der Installation eines anonymen Behandlungs- oder Krankenschein für Menschen ohne Krankenversicherung. Dabei bündelt sich hier der Erfahrungsschatz verschiedener Initiativen aus dem gesamten Bundesgebiet.
Die Auswirkungen einer Fassadenbegrünung auf den Wärmeinseleffekt in Stuttgart wurde für eine Hitzeperiode numerisch simuliert und bewertet. Die Ergebnisse zeigten positive Auswirkungen innerhalb des Simulationsgebiets sowie eine geringe Fernwirkung auf benachbarte Stadtquartiere. Diese Änderungen können zur Verbesserung des thermischen Komforts im Außenraum beitragen. Eine reduzierte Temperatur der Außenoberfläche führt darüber hinaus auch zu einer geringeren Oberflächentemperatur der Wandinnenseite, welche die Innenraumtemperatur beeinflusst. Folglich kann die thermische Behaglichkeit auch im Innenraum erhöht werden.