@phdthesis{Brehm2011, author = {Brehm, Maik}, title = {Vibration-based model updating: Reduction and quantification of uncertainties}, doi = {10.25643/bauhaus-universitaet.1465}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20110926-15553}, school = {Bauhaus-Universit{\"a}t Weimar}, year = {2011}, abstract = {Numerical models and their combination with advanced solution strategies are standard tools for many engineering disciplines to design or redesign structures and to optimize designs with the purpose to improve specific requirements. As the successful application of numerical models depends on their suitability to represent the behavior related to the intended use, they should be validated by experimentally obtained results. If the discrepancy between numerically derived and experimentally obtained results is not acceptable, a model revision or a revision of the experiment need to be considered. Model revision is divided into two classes, the model updating and the basic revision of the numerical model. The presented thesis is related to a special branch of model updating, the vibration-based model updating. Vibration-based model updating is a tool to improve the correlation of the numerical model by adjusting uncertain model input parameters by means of results extracted from vibration tests. Evidently, uncertainties related to the experiment, the numerical model, or the applied numerical solving strategies can influence the correctness of the identified model input parameters. The reduction of uncertainties for two critical problems and the quantification of uncertainties related to the investigation of several nominally identical structures are the main emphases of this thesis. First, the reduction of uncertainties by optimizing reference sensor positions is considered. The presented approach relies on predicted power spectral amplitudes and an initial finite element model as a basis to define the assessment criterion for predefined sensor positions. In combination with geometry-based design variables, which represent the sensor positions, genetic and particle swarm optimization algorithms are applied. The applicability of the proposed approach is demonstrated on a numerical benchmark study of a simply supported beam and a case study of a real test specimen. Furthermore, the theory of determining the predicted power spectral amplitudes is validated with results from vibration tests. Second, the possibility to reduce uncertainties related to an inappropriate assignment for numerically derived and experimentally obtained modes is investigated. In the context of vibration-based model updating, the correct pairing is essential. The most common criterion for indicating corresponding mode shapes is the modal assurance criterion. Unfortunately, this criterion fails in certain cases and is not reliable for automatic approaches. Hence, an alternative criterion, the energy-based modal assurance criterion, is proposed. This criterion combines the mathematical characteristic of orthogonality with the physical properties of the structure by modal strain energies. A numerical example and a case study with experimental data are presented to show the advantages of the proposed energy-based modal assurance criterion in comparison to the traditional modal assurance criterion. Third, the application of optimization strategies combined with information theory based objective functions is analyzed for the purpose of stochastic model updating. This approach serves as an alternative to the common sensitivity-based stochastic model updating strategies. Their success depends strongly on the defined initial model input parameters. In contrast, approaches based on optimization strategies can be more flexible. It can be demonstrated, that the investigated nature inspired optimization strategies in combination with Bhattacharyya distance and Kullback-Leibler divergence are appropriate. The obtained accuracies and the respective computational effort are comparable with sensitivity-based stochastic model updating strategies. The application of model updating procedures to improve the quality and suitability of a numerical model is always related to additional costs. The presented innovative approaches will contribute to reduce and quantify uncertainties within a vibration-based model updating process. Therefore, the increased benefit can compensate the additional effort, which is necessary to apply model updating procedures.}, subject = {Dynamik}, language = {en} } @phdthesis{Nickerson, author = {Nickerson, Seth}, title = {Thermo-Mechanical Behavior of Honeycomb, Porous, Microcracked Ceramics}, doi = {10.25643/bauhaus-universitaet.3975}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20190911-39753}, school = {Bauhaus-Universit{\"a}t Weimar}, abstract = {The underlying goal of this work is to reduce the uncertainty related to thermally induced stress prediction. This is accomplished by considering use of non-linear material behavior, notably path dependent thermal hysteresis behavior in the elastic properties. Primary novel factors of this work center on two aspects. 1. Broad material characterization and mechanistic material understanding, giving insight into why this class of material behaves in characteristic manners. 2. Development and implementation of a thermal hysteresis material model and its use to determine impact on overall macroscopic stress predictions. Results highlight microcracking evolution and behavior as the dominant mechanism for material property complexity in this class of materials. Additionally, it was found that for the cases studied, thermal hysteresis behavior impacts relevant peak stress predictions of a heavy-duty diesel particulate filter undergoing a drop-to-idle regeneration by less than ~15\% for all conditions tested. It is also found that path independent heating curves may be utilized for a linear solution assumption to simplify analysis. This work brings forth a newly conceived concept of a 3 state, 4 path, thermally induced microcrack evolution process; demonstrates experimental behavior that is consistent with the proposed mechanisms, develops a mathematical framework that describes the process and quantifies the impact in a real world application space.}, subject = {Keramik}, language = {en} } @phdthesis{Mai, author = {Mai, Luu}, title = {Structural Control Systems in High-speed Railway Bridges}, issn = {1610-7381}, doi = {10.25643/bauhaus-universitaet.2339}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20141223-23391}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {147}, abstract = {Structural vibration control of high-speed railway bridges using tuned mass dampers, semi-active tuned mass dampers, fluid viscous dampers and magnetorheological dampers to reduce resonant structural vibrations is studied. In this work, the addressed main issues include modeling of the dynamic interaction of the structures, optimization of the parameters of the dampers and comparison of their efficiency. A new approach to optimize multiple tuned mass damper systems on an uncertain model is proposed based on the H-infinity optimization criteria and the DK iteration procedure with norm-bounded uncertainties in frequency domain. The parameters of tuned mass dampers are optimized directly and simultaneously on different modes contributing significantly to the multi-resonant peaks to explore the different possible combinations of parameters. The effectiveness of the present method is also evaluated through comparison with a previous method. In the case of semi-active tuned mass dampers, an optimization algorithm is derived to control the magnetorheological damper in these semi-active damping systems. The use of the proposed algorithm can generate various combinations of control gains and state variables. This can lead to the improvement of the ability of MR dampers to track the desired control forces. An uncertain model to reduce detuning effects is also considered in this work. Next, for fluid viscous dampers, in order to tune the optimal parameters of fluid viscous dampers to the vicinity of the exact values, analytical formulae which can include structural damping are developed based on the perturbation method. The proposed formulae can also be considered as an improvement of the previous analytical formulae, especially for bridge beams with large structural damping. Finally, a new combination of magnetorheological dampers and a double-beam system to improve the performance of the primary structure vibration is proposed. An algorithm to control magnetorheological dampers in this system is developed by using standard linear matrix inequality techniques. Weight functions as a loop shaping procedure are also introduced in the feedback controllers to improve the tracking ability of magnetorheological damping forces. To this end, the effectiveness of magnetorheological dampers controlled by the proposed scheme, along with the effects of the uncertain and time-delay parameters on the models, are evaluated through numerical simulations. Additionally, a comparison of the dampers based on their performance is also considered in this work.}, subject = {High-speed railway bridge}, language = {en} } @phdthesis{Ghasemi, author = {Ghasemi, Hamid}, title = {Stochastic optimization of fiber reinforced composites considering uncertainties}, doi = {10.25643/bauhaus-universitaet.2704}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20161117-27042}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {140}, abstract = {Briefly, the two basic questions that this research is supposed to answer are: 1. Howmuch fiber is needed and how fibers should be distributed through a fiber reinforced composite (FRC) structure in order to obtain the optimal and reliable structural response? 2. How do uncertainties influence the optimization results and reliability of the structure? Giving answer to the above questions a double stage sequential optimization algorithm for finding the optimal content of short fiber reinforcements and their distribution in the composite structure, considering uncertain design parameters, is presented. In the first stage, the optimal amount of short fibers in a FRC structure with uniformly distributed fibers is conducted in the framework of a Reliability Based Design Optimization (RBDO) problem. Presented model considers material, structural and modeling uncertainties. In the second stage, the fiber distribution optimization (with the aim to further increase in structural reliability) is performed by defining a fiber distribution function through a Non-Uniform Rational BSpline (NURBS) surface. The advantages of using the NURBS surface as a fiber distribution function include: using the same data set for the optimization and analysis; high convergence rate due to the smoothness of the NURBS; mesh independency of the optimal layout; no need for any post processing technique and its non-heuristic nature. The output of stage 1 (the optimal fiber content for homogeneously distributed fibers) is considered as the input of stage 2. The output of stage 2 is the Reliability Index (b ) of the structure with the optimal fiber content and distribution. First order reliability method (in order to approximate the limit state function) as well as different material models including Rule of Mixtures, Mori-Tanaka, energy-based approach and stochastic multi-scales are implemented in different examples. The proposed combined model is able to capture the role of available uncertainties in FRC structures through a computationally efficient algorithm using all sequential, NURBS and sensitivity based techniques. The methodology is successfully implemented for interfacial shear stress optimization in sandwich beams and also for optimization of the internal cooling channels in a ceramic matrix composite. Finally, after some changes and modifications by combining Isogeometric Analysis, level set and point wise density mapping techniques, the computational framework is extended for topology optimization of piezoelectric / flexoelectric materials.}, subject = {Finite-Elemente-Methode}, language = {en} } @phdthesis{Liu, author = {Liu, Bokai}, title = {Stochastic multiscale modeling of polymeric nanocomposites using Data-driven techniques}, doi = {10.25643/bauhaus-universitaet.4637}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20220503-46379}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {134}, abstract = {In recent years, lightweight materials, such as polymer composite materials (PNCs) have been studied and developed due to their excellent physical and chemical properties. Structures composed of these composite materials are widely used in aerospace engineering structures, automotive components, and electrical devices. The excellent and outstanding mechanical, thermal, and electrical properties of Carbon nanotube (CNT) make it an ideal filler to strengthen polymer materials' comparable properties. The heat transfer of composite materials has very promising engineering applications in many fields, especially in electronic devices and energy storage equipment. It is essential in high-energy density systems since electronic components need heat dissipation functionality. Or in other words, in electronic devices the generated heat should ideally be dissipated by light and small heat sinks. Polymeric composites consist of fillers embedded in a polymer matrix, the first ones will significantly affect the overall (macroscopic) performance of the material. There are many common carbon-based fillers such as single-walled carbon nanotubes (SWCNT), multi-walled carbon nanotubes (MWCNT), carbon nanobuds (CNB), fullerene, and graphene. Additives inside the matrix have become a popular subject for researchers. Some extraordinary characters, such as high-performance load, lightweight design, excellent chemical resistance, easy processing, and heat transfer, make the design of polymeric nanotube composites (PNCs) flexible. Due to the reinforcing effects with different fillers on composite materials, it has a higher degree of freedom and can be designed for the structure according to specific applications' needs. As already stated, our research focus will be on SWCNT enhanced PNCs. Since experiments are timeconsuming, sometimes expensive and cannot shed light into phenomena taking place for instance at the interfaces/interphases of composites, they are often complemented through theoretical and computational analysis. While most studies are based on deterministic approaches, there is a comparatively lower number of stochastic methods accounting for uncertainties in the input parameters. In deterministic models, the output of the model is fully determined by the parameter values and the initial conditions. However, uncertainties in the input parameters such as aspect ratio, volume fraction, thermal properties of fiber and matrix need to be taken into account for reliable predictions. In this research, a stochastic multiscale method is provided to study the influence of numerous uncertain input parameters on the thermal conductivity of the composite. Therefore, a hierarchical multi-scale method based on computational homogenization is presented in to predict the macroscopic thermal conductivity based on the fine-scale structure. In order to study the inner mechanism, we use the finite element method and employ surrogate models to conduct a Global Sensitivity Analysis (GSA). The SA is performed in order to quantify the influence of the conductivity of the fiber, matrix, Kapitza resistance, volume fraction and aspect ratio on the macroscopic conductivity. Therefore, we compute first-order and total-effect sensitivity indices with different surrogate models. As stochastic multiscale models are computational expensive, surrogate approaches are commonly exploited. With the emergence of high performance computing and artificial intelligence, machine learning has become a popular modeling tool for numerous applications. Machine learning (ML) is commonly used in regression and maps data through specific rules with algorithms to build input and output models. They are particularly useful for nonlinear input-output relationships when sufficient data is available. ML has also been used in the design of new materials and multiscale analysis. For instance, Artificial neural networks and integrated learning seem to be ideally for such a task. They can theoretically simulate any non-linear relationship through the connection of neurons. Mapping relationships are employed to carry out data-driven simulations of inputs and outputs in stochastic modeling. This research aims to develop a stochastic multi-scale computational models of PNCs in heat transfer. Multi-scale stochastic modeling with uncertainty analysis and machine learning methods consist of the following components: -Uncertainty Analysis. A surrogate based global sensitivity analysis is coupled with a hierarchical multi-scale method employing computational homogenization. The effect of the conductivity of the fibers and the matrix, the Kapitza resistance, volume fraction and aspect ratio on the 'macroscopic' conductivity of the composite is systematically studied. All selected surrogate models yield consistently the conclusions that the most influential input parameters are the aspect ratio followed by the volume fraction. The Kapitza Resistance has no significant effect on the thermal conductivity of the PNCs. The most accurate surrogate model in terms of the R2 value is the moving least square (MLS). -Hybrid Machine Learning Algorithms. A combination of artificial neural network (ANN) and particle swarm optimization (PSO) is applied to estimate the relationship between variable input and output parameters. The ANN is used for modeling the composite while PSO improves the prediction performance through an optimized global minimum search. The thermal conductivity of the fibers and the matrix, the kapitza resistance, volume fraction and aspect ratio are selected as input parameters. The output is the macroscopic (homogenized) thermal conductivity of the composite. The results show that the PSO significantly improves the predictive ability of this hybrid intelligent algorithm, which outperforms traditional neural networks. -Stochastic Integrated Machine Learning. A stochastic integrated machine learning based multiscale approach for the prediction of the macroscopic thermal conductivity in PNCs is developed. Seven types of machine learning models are exploited in this research, namely Multivariate Adaptive Regression Splines (MARS), Support Vector Machine (SVM), Regression Tree (RT), Bagging Tree (Bag), Random Forest (RF), Gradient Boosting Machine (GBM) and Cubist. They are used as components of stochastic modeling to construct the relationship between the variable of the inputs' uncertainty and the macroscopic thermal conductivity of PNCs. Particle Swarm Optimization (PSO) is used for hyper-parameter tuning to find the global optimal values leading to a significant reduction in the computational cost. The advantages and disadvantages of various methods are also analyzed in terms of computing time and model complexity to finally give a recommendation for the applicability of different models.}, subject = {Polymere}, language = {en} } @phdthesis{Chan, author = {Chan, Chiu Ling}, title = {Smooth representation of thin shells and volume structures for isogeometric analysis}, doi = {10.25643/bauhaus-universitaet.4208}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20200812-42083}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {162}, abstract = {The purpose of this study is to develop self-contained methods for obtaining smooth meshes which are compatible with isogeometric analysis (IGA). The study contains three main parts. We start by developing a better understanding of shapes and splines through the study of an image-related problem. Then we proceed towards obtaining smooth volumetric meshes of the given voxel-based images. Finally, we treat the smoothness issue on the multi-patch domains with C1 coupling. Following are the highlights of each part. First, we present a B-spline convolution method for boundary representation of voxel-based images. We adopt the filtering technique to compute the B-spline coefficients and gradients of the images effectively. We then implement the B-spline convolution for developing a non-rigid images registration method. The proposed method is in some sense of "isoparametric", for which all the computation is done within the B-splines framework. Particularly, updating the images by using B-spline composition promote smooth transformation map between the images. We show the possible medical applications of our method by applying it for registration of brain images. Secondly, we develop a self-contained volumetric parametrization method based on the B-splines boundary representation. We aim to convert a given voxel-based data to a matching C1 representation with hierarchical cubic splines. The concept of the osculating circle is employed to enhance the geometric approximation, where it is done by a single template and linear transformations (scaling, translations, and rotations) without the need for solving an optimization problem. Moreover, we use the Laplacian smoothing and refinement techniques to avoid irregular meshes and to improve mesh quality. We show with several examples that the method is capable of handling complex 2D and 3D configurations. In particular, we parametrize the 3D Stanford bunny which contains irregular shapes and voids. Finally, we propose the B´ezier ordinates approach and splines approach for C1 coupling. In the first approach, the new basis functions are defined in terms of the B´ezier Bernstein polynomials. For the second approach, the new basis is defined as a linear combination of C0 basis functions. The methods are not limited to planar or bilinear mappings. They allow the modeling of solutions to fourth order partial differential equations (PDEs) on complex geometric domains, provided that the given patches are G1 continuous. Both methods have their advantages. In particular, the B´ezier approach offer more degree of freedoms, while the spline approach is more computationally efficient. In addition, we proposed partial degree elevation to overcome the C1-locking issue caused by the over constraining of the solution space. We demonstrate the potential of the resulting C1 basis functions for application in IGA which involve fourth order PDEs such as those appearing in Kirchhoff-Love shell models, Cahn-Hilliard phase field application, and biharmonic problems.}, subject = {Modellierung}, language = {en} } @phdthesis{Tan, author = {Tan, Fengjie}, title = {Shape Optimization Design of Arch Type Dams under Uncertainties}, doi = {10.25643/bauhaus-universitaet.3960}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20190819-39608}, school = {Bauhaus-Universit{\"a}t Weimar}, abstract = {Due to an increased need for hydro-electricity, water storage, and flood protection, it is assumed that a series of new dams will be built throughout the world. Comparing existing design methodologies for arch-type dams, model-based shape optimization can effectively reduce construction costs and leverage the properties of construction materials. To apply the means of shape optimization, suitable variables need to be chosen to formulate the objective function, which is the volume of the arch dam here. In order to increase the consistency with practical conditions, a great number of geometrical and behavioral constraints are included in the mathematical model. An optimization method, namely Genetic Algorithm is adopted which allows a global search. Traditional optimization techniques are realized based on a deterministic approach, which means that the material properties and loading conditions are assumed to be fixed values. As a result, the real-world structures that are optimized by these approaches suffer from uncertainties that one needs to be aware of. Hence, in any optimization process for arch dams, it is nec- essary to find a methodology that is capable of considering the influences of uncertainties and generating a solution which is robust enough against the uncertainties. The focus of this thesis is the formulation and the numerical method for the optimization of the arch dam under the uncertainties. The two main models, the probabilistic model, and non-probabilistic models are intro- duced and discussed. Classic procedures of probabilistic approaches un- der uncertainties, such as RDO (robust design optimization) and RBDO (reliability-based design optimization), are in general computationally ex- pensive and rely on estimates of the system's response variance and fail- ure probabilities. Instead, the robust optimization (RO) method which is based on the non-probabilistic model, will not follow a full probabilistic approach but works with pre-defined confidence levels. This leads to a bi-level optimization program where the volume of the dam is optimized under the worst combination of the uncertain parameters. By this, robust and reliable designs are obtained and the result is independent of any as- sumptions on stochastic properties of the random variables in the model. The optimization of an arch-type dam is realized here by a robust optimiza- tion method under load uncertainty, where hydraulic and thermal loads are considered. The load uncertainty is modeled as an ellipsoidal expression. Comparing with any traditional deterministic optimization (DO) method, which only concerns the minimum objective value and offers a solution candidate close to limit-states, the RO method provides a robust solution against uncertainties. All the above mentioned methods are applied to the optimization of the arch dam to compare with the optimal design with DO methods. The re- sults are compared and analyzed to discuss the advantages and drawbacks of each method. In order to reduce the computational cost, a ranking strategy and an ap- proximation model are further involved to do a preliminary screening. By means of these, the robust design can generate an improved arch dam structure which ensures both safety and serviceability during its lifetime.}, subject = {Wasserbau}, language = {en} } @phdthesis{Ahmad, author = {Ahmad, Sofyan}, title = {Reference Surface-Based System Identification}, doi = {10.25643/bauhaus-universitaet.2113}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20140205-21132}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {153}, abstract = {Environmental and operational variables and their impact on structural responses have been acknowledged as one of the most important challenges for the application of the ambient vibration-based damage identification in structures. The damage detection procedures may yield poor results, if the impacts of loading and environmental conditions of the structures are not considered. The reference-surface-based method, which is proposed in this thesis, is addressed to overcome this problem. In the proposed method, meta-models are used to take into account significant effects of the environmental and operational variables. The usage of the approximation models, allows the proposed method to simply handle multiple non-damaged variable effects simultaneously, which for other methods seems to be very complex. The input of the meta-model are the multiple non-damaged variables while the output is a damage indicator. The reference-surface-based method diminishes the effect of the non-damaged variables to the vibration based damage detection results. Hence, the structure condition that is assessed by using ambient vibration data at any time would be more reliable. Immediate reliable information regarding the structure condition is required to quickly respond to the event, by means to take necessary actions concerning the future use or further investigation of the structures, for instance shortly after extreme events such as earthquakes. The critical part of the proposed damage detection method is the learning phase, where the meta-models are trained by using input-output relation of observation data. Significant problems that may encounter during the learning phase are outlined and some remedies to overcome the problems are suggested. The proposed damage identification method is applied to numerical and experimental models. In addition to the natural frequencies, wavelet energy and stochastic subspace damage indicators are used.}, subject = {System Identification}, language = {en} } @phdthesis{Goswami, author = {Goswami, Somdatta}, title = {Phase field modeling of fracture with isogeometric analysis and machine learning methods}, doi = {10.25643/bauhaus-universitaet.4384}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20210304-43841}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {168}, abstract = {This thesis presents the advances and applications of phase field modeling in fracture analysis. In this approach, the sharp crack surface topology in a solid is approximated by a diffusive crack zone governed by a scalar auxiliary variable. The uniqueness of phase field modeling is that the crack paths are automatically determined as part of the solution and no interface tracking is required. The damage parameter varies continuously over the domain. But this flexibility comes with associated difficulties: (1) a very fine spatial discretization is required to represent sharp local gradients correctly; (2) fine discretization results in high computational cost; (3) computation of higher-order derivatives for improved convergence rates and (4) curse of dimensionality in conventional numerical integration techniques. As a consequence, the practical applicability of phase field models is severely limited. The research presented in this thesis addresses the difficulties of the conventional numerical integration techniques for phase field modeling in quasi-static brittle fracture analysis. The first method relies on polynomial splines over hierarchical T-meshes (PHT-splines) in the framework of isogeometric analysis (IGA). An adaptive h-refinement scheme is developed based on the variational energy formulation of phase field modeling. The fourth-order phase field model provides increased regularity in the exact solution of the phase field equation and improved convergence rates for numerical solutions on a coarser discretization, compared to the second-order model. However, second-order derivatives of the phase field are required in the fourth-order model. Hence, at least a minimum of C1 continuous basis functions are essential, which is achieved using hierarchical cubic B-splines in IGA. PHT-splines enable the refinement to remain local at singularities and high gradients, consequently reducing the computational cost greatly. Unfortunately, when modeling complex geometries, multiple parameter spaces (patches) are joined together to describe the physical domain and there is typically a loss of continuity at the patch boundaries. This decrease of smoothness is dictated by the geometry description, where C0 parameterizations are normally used to deal with kinks and corners in the domain. Hence, the application of the fourth-order model is severely restricted. To overcome the high computational cost for the second-order model, we develop a dual-mesh adaptive h-refinement approach. This approach uses a coarser discretization for the elastic field and a finer discretization for the phase field. Independent refinement strategies have been used for each field. The next contribution is based on physics informed deep neural networks. The network is trained based on the minimization of the variational energy of the system described by general non-linear partial differential equations while respecting any given law of physics, hence the name physics informed neural network (PINN). The developed approach needs only a set of points to define the geometry, contrary to the conventional mesh-based discretization techniques. The concept of `transfer learning' is integrated with the developed PINN approach to improve the computational efficiency of the network at each displacement step. This approach allows a numerically stable crack growth even with larger displacement steps. An adaptive h-refinement scheme based on the generation of more quadrature points in the damage zone is developed in this framework. For all the developed methods, displacement-controlled loading is considered. The accuracy and the efficiency of both methods are studied numerically showing that the developed methods are powerful and computationally efficient tools for accurately predicting fractures.}, subject = {Phasenfeldmodell}, language = {en} } @phdthesis{Jaouadi, author = {Jaouadi, Zouhour}, title = {Pareto and Reliability-Oriented Aeroelastic Shape Optimization of Bridge Decks}, doi = {10.25643/bauhaus-universitaet.4935}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20230303-49352}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {167}, abstract = {Due to the development of new technologies and materials, optimized bridge design has recently gained more attention. The aim is to reduce the bridge components materials and the CO2 emission from the cement manufacturing process. Thus, most long-span bridges are designed to be with high flexibility, low structural damping, and longer and slender spans. Such designs lead, however, to aeroelastic challenges. Moreover, the consideration of both the structural and aeroelastic behavior in bridges leads to contradictory solutions as the structural constraints lead to deck prototypes with high depth which provide high inertia to material volume ratios. On the other hand, considering solely the aerodynamic requirements, slender airfoil-shaped bridge box girders are recommended since they prevent vortex shedding and exhibit minimum drag. Within this framework comes this study which provides approaches to find optimal bridge deck cross-sections while considering the aerodynamic effects. Shape optimization of deck cross-section is usually formulated to minimize the amount of material by finding adequate parameters such as the depth, the height, and the thickness and while ensuring the overall stability of the structure by the application of some constraints. Codes and studies have been implemented to analyze the wind phenomena and the structural responses towards bridge deck cross-sections where simplifications have been adopted due to the complexity and the uniqueness of such components besides the difficulty of obtaining a final model of the aerodynamic behavior. In this thesis, two main perspectives have been studied; the first is fully deterministic and presents a novel framework on generating optimal aerodynamic shapes for streamlined and trapezoidal cross-sections based on the meta-modeling approach. Single and multi-objective optimizations were both carried out and a Pareto Front is generated. The performance of the optimal designs is checked afterwards. In the second part, a new strategy based on Reliability-Based Design Optimization (RBDO) to mitigate the vortex-induced vibration (VIV) on the Trans-Tokyo Bay bridge is proposed. Small changes in the leading and trailing edges are presented and uncertainties are considered in the structural system. Probabilistic constraints based on polynomial regression are evaluated and the problem is solved while applying the Reliability Index Approach (RIA) and the Performance Measure Approach (PMA). The results obtained in the first part showed that the aspect ratio has a significant effect on the aerodynamic behavior where deeper cross-sections have lower resistance against flutter and should be avoided. In the second part, the adopted RBDO approach succeeded to mitigate the VIV, and it is proven that designs with narrow or prolonged bottom-base length and featuring an abrupt surface change in the leading and trailing edges can lead to high vertical vibration amplitude. It is expected that this research will help engineers with the selections of the adequate deck cross-section layout, and encourage researchers to apply concepts of optimization regarding this field and develop the presented approaches for further studies.}, subject = {Gestaltoptimierung}, language = {en} } @phdthesis{Schemmann, author = {Schemmann, Christoph}, title = {Optimierung von radialen Verdichterlaufr{\"a}dern unter Ber{\"u}cksichtigung empirischer und analytischer Vorinformationen mittels eines mehrstufigen Sampling Verfahrens}, doi = {10.25643/bauhaus-universitaet.3974}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20190910-39748}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {233}, abstract = {Turbomachinery plays an important role in many cases of energy generation or conversion. Therefore, turbomachinery is a promising approaching point for optimization in order to increase the efficiency of energy use. In recent years, the use of automated optimization strategies in combination with numerical simulation has become increasingly popular in many fields of engineering. The complex interactions between fluid and solid mechanics encountered in turbomachines on the one hand and the high computational expense needed to calculate the performance on the other hand, have, however, prevented a widespread use of these techniques in this field of engineering. The objective of this work was the development of a strategy for efficient metamodel based optimization of centrifugal compressor impellers. In this context, the main focus is the reduction of the required numerical expense. The central idea followed in this research was the incorporation of preliminary information acquired from low-fidelity computation methods and empirical correlations into the sampling process to identify promising regions of the parameter space. This information was then used to concentrate the numerically expensive high-fidelity computations of the fluid dynamic and structure mechanic performance of the impeller in these regions while still maintaining a good coverage of the whole parameter space. The development of the optimization strategy can be divided into three main tasks. Firstly, the available preliminary information had to be researched and rated. This research identified loss models based on one dimensional flow physics and empirical correlations as the best suited method to predict the aerodynamic performance. The loss models were calibrated using available performance data to obtain a high prediction quality. As no sufficiently exact models for the prediction of the mechanical loading of the impellercould be identified, a metamodel based on finite element computations was chosen for this estimation. The second task was the development of a sampling method which concentrates samples in regions of the parameter space where high quality designs are predicted by the preliminary information while maintaining a good overall coverage. As available methods like rejection sampling or Markov-chain Monte-Carlo methods did not meet the requirements in terms of sample distribution and input correlation, a new multi-fidelity sampling method called "Filtered Sampling"has been developed. The last task was the development of an automated computational workflow. This workflow encompasses geometry parametrization, geometry generation, grid generation and computation of the aerodynamic performance and the structure mechanic loading. Special emphasis was put into the development of a geometry parametrization strategy based on fluid mechanic considerations to prevent the generation of physically inexpedient designs. Finally, the optimization strategy, which utilizes the previously developed tools, was successfully employed to carry out three optimization tasks. The efficiency of the method was proven by the first and second testcase where an existing compressor design was optimized by the presented method. The results were comparable to optimizations which did not take preliminary information into account, while the required computational expense cloud be halved. In the third testcase, the method was applied to generate a new impeller design. In contrast to the previous examples, this optimization featuredlargervariationsoftheimpellerdesigns. Therefore, theapplicability of the method to parameter spaces with significantly varying designs could be proven, too.}, subject = {Simulation}, language = {en} } @phdthesis{Zabel, author = {Zabel, Volkmar}, title = {Operational modal analysis - Theory and aspects of application in civil engineering}, editor = {K{\"o}nke, Carsten and Lahmer, Tom and Rabczuk, Timon}, doi = {10.25643/bauhaus-universitaet.4006}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20191030-40061}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {263}, abstract = {In recent years the demand on dynamic analyses of existing structures in civil engineering has remarkably increased. These analyses are mainly based on numerical models. Accordingly, the generated results depend on the quality of the used models. Therefore it is very important that the models describe the considered systems such that the behaviour of the physical structure is realistically represented. As any model is based on assumptions, there is always a certain degree of uncertainty present in the results of a simulation based on the respective numerical model. To minimise these uncertainties in the prediction of the response of a structure to a certain loading, it has become common practice to update or calibrate the parameters of a numerical model based on observations of the structural behaviour of the respective existing system. The determination of the behaviour of an existing structure requires experimental investigations. If the numerical analyses concern the dynamic response of a structure it is sensible to direct the experimental investigations towards the identification of the dynamic structural behaviour which is determined by the modal parameters of the system. In consequence, several methods for the experimental identification of modal parameters have been developed since the 1980ies. Due to various technical restraints in civil engineering which limit the possibilities to excitate a structure with economically reasonable effort, several methods have been developed that allow a modal identification form tests with an ambient excitation. The approach of identifying modal parameters only from measurements of the structural response without precise knowledge of the excitation is known as output-only or operational modal analysis. Since operational modal analysis (OMA) can be considered as a link between numerical modelling and simulation on the one hand and the dynamic behaviour of an existing structure on the other hand, the respective algorithms connect both the concepts of structural dynamics and mathematical tools applied within the processing of experimental data. Accordingly, the related theoretical topics are revised after an introduction into the topic. Several OMA methods have been developed over the last decades. The most established algorithms are presented here and their application is illustrated by means of both a small numerical and an experimental example. Since experimentally obtained results always underly manifold influences, an appropriate postprocessing of the results is necessary for a respective quality assessment. This quality assessment does not only require respective indicators but should also include the quantification of uncertainties. One special feature in modal testing is that it is common to instrument the structure in different sensor setups to improve the spacial resolution of identified mode shapes. The modal information identified from tests in several setups needs to be merged a posteriori. Algorithms to cope with this problem are also presented. Due to the fact that the amount of data generated in modal tests can become very large, manual processing can become extremely expensive or even impossible, for example in the case of a long-term continuous structural monitoring. In these situations an automated analysis and postprocessing are essential. Descriptions of respective methodologies are therefore also included in this work. Every structural system in civil engineering is unique and so also every identification of modal parameters has its specific challenges. Some aspects that can be faced in practical applications of operational modal analysis are presented and discussed in a chapter that is dedicated specific problems that an analyst may have to overcome. Case studies of systems with very close modes, with limited accessibility as well as the application of different OMA methods are described and discussed. In this context the focus is put on several types of uncertainty that may occur in the multiple stages of an operational modal analysis. In literature only very specific uncertainties at certain stages of the analysis are addressed. Here, the topic of uncertainties has been considered in a broader sense and approaches for treating respective problems are suggested. Eventually, it is concluded that the methodologies of operatinal modal analysis and related technical solutions have been well-engineered already. However, as in any discipline that includes experiments, a certain degree of uncertainty always remains in the results. From these conclusions has been derived a demand for further research and development that should be directed towards the minimisation of these uncertainties and to a respective optimisation of the steps and corresponding parameters included in an operational modal analysis.}, subject = {Modalanalyse}, language = {en} } @phdthesis{Hamdia, author = {Hamdia, Khader}, title = {On the fracture toughness of polymeric nanocomposites: Comprehensive stochastic and numerical studies}, doi = {10.25643/bauhaus-universitaet.3765}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20180712-37652}, school = {Bauhaus-Universit{\"a}t Weimar}, abstract = {Polymeric nanocomposites (PNCs) are considered for numerous nanotechnology such as: nano-biotechnology, nano-systems, nanoelectronics, and nano-structured materials. Commonly , they are formed by polymer (epoxy) matrix reinforced with a nanosized filler. The addition of rigid nanofillers to the epoxy matrix has offered great improvements in the fracture toughness without sacrificing other important thermo-mechanical properties. The physics of the fracture in PNCs is rather complicated and is influenced by different parameters. The presence of uncertainty in the predicted output is expected as a result of stochastic variance in the factors affecting the fracture mechanism. Consequently, evaluating the improved fracture toughness in PNCs is a challenging problem. Artificial neural network (ANN) and adaptive neuro-fuzzy inference system (ANFIS) have been employed to predict the fracture energy of polymer/particle nanocomposites. The ANN and ANFIS models were constructed, trained, and tested based on a collection of 115 experimental datasets gathered from the literature. The performance evaluation indices of the developed ANN and ANFIS showed relatively small error, with high coefficients of determination (R2), and low root mean square error and mean absolute percentage error. In the framework for uncertainty quantification of PNCs, a sensitivity analysis (SA) has been conducted to examine the influence of uncertain input parameters on the fracture toughness of polymer/clay nanocomposites (PNCs). The phase-field approach is employed to predict the macroscopic properties of the composite considering six uncertain input parameters. The efficiency, robustness, and repeatability are compared and evaluated comprehensively for five different SA methods. The Bayesian method is applied to develop a methodology in order to evaluate the performance of different analytical models used in predicting the fracture toughness of polymeric particles nanocomposites. The developed method have considered the model and parameters uncertainties based on different reference data (experimental measurements) gained from the literature. Three analytical models differing in theory and assumptions were examined. The coefficients of variation of the model predictions to the measurements are calculated using the approximated optimal parameter sets. Then, the model selection probability is obtained with respect to the different reference data. Stochastic finite element modeling is implemented to predict the fracture toughness of polymer/particle nanocomposites. For this purpose, 2D finite element model containing an epoxy matrix and rigid nanoparticles surrounded by an interphase zone is generated. The crack propagation is simulated by the cohesive segments method and phantom nodes. Considering the uncertainties in the input parameters, a polynomial chaos expansion (PCE) surrogate model is construed followed by a sensitivity analysis.}, subject = {Bruch}, language = {en} } @phdthesis{Hatahet, author = {Hatahet, Tareq}, title = {On the Analysis of the Disproportionate Structural Collapse in RC Buildings}, doi = {10.25643/bauhaus-universitaet.3740}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20180329-37405}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {243}, abstract = {Increasing structural robustness is the goal which is of interest for structural engineering community. The partial collapse of RC buildings is subject of this dissertation. Understanding the robustness of RC buildings will guide the development of safer structures against abnormal loading scenarios such as; explosions, earthquakes, fine, and/or long-term accumulation effects leading to deterioration or fatigue. Any of these may result in local immediate structural damage, that can propagate to the rest of the structure causing what is known by the disproportionate collapse. This work handels collapse propagation through various analytical approaches which simplifies the mechanical description of damaged reinfoced concrete structures due to extreme acidental event.}, subject = {Beton}, language = {en} } @phdthesis{Itam, author = {Itam, Zarina}, title = {Numerical Simulation of Thermo-Chemo-Hygro-Mechanical Alkali-Silica Reaction Model in Concrete at the Mesoscale and Macroscale}, doi = {10.25643/bauhaus-universitaet.2335}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20141218-23352}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {164}, abstract = {Alkali-silica reaction causes major problems in concrete structures due to the rapidity of its deformation which leads to the serviceability limit of the structure being reached well before its time. Factors that affect ASR vary greatly, including alkali and silica content, relative humidity, temperature and porosity of the cementitious matrix,all these making it a very complex phenomenon to consider explicitly. With this in mind, the finite element technique was used to build models and generate expansive pressures and damage propagation due to ASR under the influence of thermo-hygrochemoelastic loading. Since ASR initializes in the mesoscopic regions of the concrete, the accumulative effects of its expansion escalates onto the macroscale level with the development of web cracking on the concrete surface, hence solution of the damage model as well as simulation of the ASR phenomenon at both the macroscale and mesoscale levels have been performed. The macroscale model realizes the effects of ASR expansion as a whole and shows how it develops under the influence of moisture, thermal and mechanical loading. Results of the macroscale modeling are smeared throughout the structure and are sufficient to show how damage due to ASR expansion orientates. As opposed to the mesoscale model, the heterogeneity of the model shows us how difference in material properties between aggregates and the cementitious matrix facilitates ASR expansion. With both these models, the ASR phenomenon under influence of thermo-chemo-hygro-mechanical loading can be better understood.}, subject = {Strukturmechanik}, language = {en} } @phdthesis{Zacharias, author = {Zacharias, Christin}, title = {Numerical Simulation Models for Thermoelastic Damping Effects}, doi = {10.25643/bauhaus-universitaet.4735}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20221116-47352}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {191}, abstract = {Finite Element Simulations of dynamically excited structures are mainly influenced by the mass, stiffness, and damping properties of the system, as well as external loads. The prediction quality of dynamic simulations of vibration-sensitive components depends significantly on the use of appropriate damping models. Damping phenomena have a decisive influence on the vibration amplitude and the frequencies of the vibrating structure. However, developing realistic damping models is challenging due to the multiple sources that cause energy dissipation, such as material damping, different types of friction, or various interactions with the environment. This thesis focuses on thermoelastic damping, which is the main cause of material damping in homogeneous materials. The effect is caused by temperature changes due to mechanical strains. In vibrating structures, temperature gradients arise in adjacent tension and compression areas. Depending on the vibration frequency, they result in heat flows, leading to increased entropy and the irreversible transformation of mechanical energy into thermal energy. The central objective of this thesis is the development of efficient simulation methods to incorporate thermoelastic damping in finite element analyses based on modal superposition. The thermoelastic loss factor is derived from the structure's mechanical mode shapes and eigenfrequencies. In subsequent analyses that are performed in the time and frequency domain, it is applied as modal damping. Two approaches are developed to determine the thermoelastic loss in thin-walled plate structures, as well as three-dimensional solid structures. The realistic representation of the dissipation effects is verified by comparing the simulation results with experimentally determined data. Therefore, an experimental setup is developed to measure material damping, excluding other sources of energy dissipation. The three-dimensional solid approach is based on the determination of the generated entropy and therefore the generated heat per vibration cycle, which is a measure for thermoelastic loss in relation to the total strain energy. For thin plate structures, the amount of bending energy in a modal deformation is calculated and summarized in the so-called Modal Bending Factor (MBF). The highest amount of thermoelastic loss occurs in the state of pure bending. Therefore, the MBF enables a quantitative classification of the mode shapes concerning the thermoelastic damping potential. The results of the developed simulations are in good agreement with the experimental results and are appropriate to predict thermoelastic loss factors. Both approaches are based on modal superposition with the advantage of a high computational efficiency. Overall, the modeling of thermoelastic damping represents an important component in a comprehensive damping model, which is necessary to perform realistic simulations of vibration processes.}, subject = {Werkstoffd{\"a}mpfung}, language = {en} } @phdthesis{Unger2009, author = {Unger, J{\"o}rg F.}, title = {Neural networks in a multiscale approach for concrete}, doi = {10.25643/bauhaus-universitaet.1392}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20090626-14763}, school = {Bauhaus-Universit{\"a}t Weimar}, year = {2009}, abstract = {From a macroscopic point of view, failure within concrete structures is characterized by the initiation and propagation of cracks. In the first part of the thesis, a methodology for macroscopic crack growth simulations for concrete structures using a cohesive discrete crack approach based on the extended finite element method is introduced. Particular attention is turned to the investigation of criteria for crack initiation and crack growth. A drawback of the macroscopic simulation is that the real physical phenomena leading to the nonlinear behavior are only modeled phenomenologically. For concrete, the nonlinear behavior is characterized by the initiation of microcracks which coalesce into macroscopic cracks. In order to obtain a higher resolution of this failure zones, a mesoscale model for concrete is developed that models particles, mortar matrix and the interfacial transition zone (ITZ) explicitly. The essential features are a representation of particles using a prescribed grading curve, a material formulation based on a cohesive approach for the ITZ and a combined model with damage and plasticity for the mortar matrix. Compared to numerical simulations, the response of real structures exhibits a stochastic scatter. This is e.g. due to the intrinsic heterogeneities of the structure. For mesoscale models, these intrinsic heterogeneities are simulated by using a random distribution of particles and by a simulation of spatially variable material parameters using random fields. There are two major problems related to numerical simulations on the mesoscale. First of all, the material parameters for the constitutive description of the materials are often difficult to measure directly. In order to estimate material parameters from macroscopic experiments, a parameter identification procedure based on Bayesian neural networks is developed which is universally applicable to any parameter identification problem in numerical simulations based on experimental results. This approach offers information about the most probable set of material parameters based on experimental data and information about the accuracy of the estimate. Consequently, this approach can be used a priori to determine a set of experiments to be carried out in order to fit the parameters of a numerical model to experimental data. The second problem is the computational effort required for mesoscale simulations of a full macroscopic structure. For this purpose, a coupling between mesoscale and macroscale model is developed. Representative mesoscale simulations are used to train a metamodel that is finally used as a constitutive model in a macroscopic simulation. Special focus is placed on the ability of appropriately simulating unloading.}, subject = {Beton}, language = {en} } @phdthesis{Wang, author = {Wang, Cuixia}, title = {Nanomechanical Resonators Based on Quasi-two-dimensional Materials}, doi = {10.25643/bauhaus-universitaet.3760}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20180709-37609}, school = {Bauhaus-Universit{\"a}t Weimar}, abstract = {Advances in nanotechnology lead to the development of nano-electro-mechanical systems (NEMS) such as nanomechanical resonators with ultra-high resonant frequencies. The ultra-high-frequency resonators have recently received significant attention for wide-ranging applications such as molecular separation, molecular transportation, ultra-high sensitive sensing, high-frequency signal processing, and biological imaging. It is well known that for micrometer length scale, first-principles technique, the most accurate approach, poses serious limitations for comparisons with experimental studies. For such larger size, classical molecular dynamics (MD) simulations are desirable, which require interatomic potentials. Additionally, a mesoscale method such as the coarse-grained (CG) method is another useful method to support simulations for even larger system sizes. Furthermore, quasi-two-dimensional (Q2D) materials have attracted intensive research interest due to their many novel properties over the past decades. However, the energy dissipation mechanisms of nanomechanical resonators based on several Q2D materials are still unknown. In this work, the addressed main issues include the development of the CG models for molybdenum disulphide (MoS2), investigation of the mechanism effects on black phosphorus (BP) nanoresonators and the application of graphene nanoresonators. The primary coverage and results of the dissertation are as follows: Method development. Firstly, a two-dimensional (2D) CG model for single layer MoS2 (SLMoS2) is analytically developed. The Stillinger-Weber (SW) potential for this 2D CG model is further parametrized, in which all SW geometrical parameters are determined analytically according to the equilibrium condition for each individual potential term, while the SW energy parameters are derived analytically based on the valence force field model. Next, the 2D CG model is further simplified to one-dimensional (1D) CG model, which describes the 2D SLMoS2 structure using a 1D chain model. This 1D CG model is applied to investigate the relaxed configuration and the resonant oscillation of the folded SLMoS2. Owning to the simplicity nature of the 1D CG model, the relaxed configuration of the folded SLMoS2 is determined analytically, and the resonant oscillation frequency is derived analytically. Considering the increasing interest in studying the properties of other 2D layered materials, and in particular those in the semiconducting transition metal dichalcogenide class like MoS2, the CG models proposed in current work provide valuable simulation approaches. Mechanism understanding. Two energy dissipation mechanisms of BP nanoresonators are focused exclusively, i.e. mechanical strain effects and defect effects (including vacancy and oxidation). Vacancy defect is intrinsic damping factor for the quality (Q)-factor, while mechanical strain and oxidation are extrinsic damping factors. Intrinsic dissipation (induced by thermal vibrations) in BP resonators (BPRs) is firstly investigated. Specifically, classical MD simulations are performed to examine the temperature dependence for the Q-factor of the single layer BPR (SLBPR) along the armchair and zigzag directions, where two-step fitting procedure is used to extract the frequency and Q-factor from the kinetic energy time history. The Q-factors of BPRs are evaluated through comparison with those of graphene and MoS2 nanoresonators. Next, effects of mechanical strain, vacancy and oxidation on BP nanoresonators are investigated in turn. Considering the increasing interest in studying the properties of BP, and in particular the lack of theoretical study for the BPRs, the results in current work provide a useful reference. Application. A novel application for graphene nanoresonators, using them to self-assemble small nanostructures such as water chains, is proposed. All of the underlying physics enabling this phenomenon is elucidated. In particular, by drawing inspiration from macroscale self-assembly using the higher order resonant modes of Chladni plates, classical MD simulations are used to investigate the self-assembly of water molecules using graphene nanoresonators. An analytic formula for the critical resonant frequency based on the interaction between water molecules and graphene is provided. Furthermore, the properties of the water chains assembled by the graphene nanoresonators are studied.}, subject = {Nanomechanik}, language = {en} } @phdthesis{Zhao, author = {Zhao, Jun-Hua}, title = {Multiscale modeling of nanodevices based on carbon nanotubes and polymers}, doi = {10.25643/bauhaus-universitaet.2107}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20140130-21078}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {175}, abstract = {This thesis concerns the physical and mechanical interactions on carbon nanotubes and polymers by multiscale modeling. CNTs have attracted considerable interests in view of their unique mechanical, electronic, thermal, optical and structural properties, which enable them to have many potential applications. Carbon nanotube exists in several structure forms, from individual single-walled carbon nanotubes (SWCNTs) and multi-walled carbon nanotubes (MWCNTs) to carbon nanotube bundles and networks. The mechanical properties of SWCNTs and MWCNTs have been extensively studied by continuum modeling and molecular dynamics (MD) simulations in the past decade since the properties could be important in the CNT-based devices. CNT bundles and networks feature outstanding mechanical performance and hierarchical structures and network topologies, which have been taken as a potential saving-energy material. In the synthesis of nanocomposites, the formation of the CNT bundles and networks is a challenge to remain in understanding how to measure and predict the properties of such large systems. Therefore, a mesoscale method such as a coarse-grained (CG) method should be developed to study the nanomechanical characterization of CNT bundles and networks formation. In this thesis, the main contributions can be written as follows: (1) Explicit solutions for the cohesive energy between carbon nanotubes, graphene and substrates are obtained through continuum modeling of the van der Waals interaction between them. (2) The CG potentials of SWCNTs are established by a molecular mechanics model. (3) The binding energy between two parallel and crossing SWCNTs and MWCNTs is obtained by continuum modeling of the van der Waals interaction between them. Crystalline and amorphous polymers are increasingly used in modern industry as tructural materials due to its important mechanical and physical properties. For crystalline polyethylene (PE), despite its importance and the studies of available MD simulations and continuum models, the link between molecular and continuum descriptions of its mechanical properties is still not well established. For amorphous polymers, the chain length and temperature effect on their elastic and elastic-plastic properties has been reported based on the united-atom (UA) and CG MD imulations in our previous work. However, the effect of the CL and temperature on the failure behavior is not understood well yet. Especially, the failure behavior under shear has been scarcely reported in previous work. Therefore, understanding the molecular origins of macroscopic fracture behavior such as fracture energy is a fundamental scientific challenge. In this thesis, the main contributions can be written as follows: (1) An analytical molecular mechanics model is developed to obtain the size-dependent elastic properties of crystalline PE. (2) We show that the two molecular mechanics models, the stick-spiral and the beam models, predict considerably different mechanical properties of materials based on energy equivalence. The difference between the two models is independent of the materials. (3) The tensile and shear failure behavior dependence on chain length and temperature in amorphous polymers are scrutinized using molecular dynamics simulations. Finally, the influence of polymer wrapped two neighbouring SWNTs' dispersion on their load transfer is investigated by molecular dynamics (MD) simulations, in which the SWNTs' position, the polymer chain length and the temperature on the interaction force is systematically studied.}, subject = {Mehrskalenmodell}, language = {en} } @phdthesis{Kessler2018, author = {Keßler, Andrea}, title = {Matrix-free voxel-based finite element method for materials with heterogeneous microstructures}, doi = {10.25643/bauhaus-universitaet.3844}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20190116-38448}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {113}, year = {2018}, abstract = {Modern image detection techniques such as micro computer tomography (μCT), magnetic resonance imaging (MRI) and scanning electron microscopy (SEM) provide us with high resolution images of the microstructure of materials in a non-invasive and convenient way. They form the basis for the geometrical models of high-resolution analysis, so called image-based analysis. However especially in 3D, discretizations of these models reach easily the size of 100 Mill. degrees of freedoms and require extensive hardware resources in terms of main memory and computing power to solve the numerical model. Consequently, the focus of this work is to combine and adapt numerical solution methods to reduce the memory demand first and then the computation time and therewith enable an execution of the image-based analysis on modern computer desktops. Hence, the numerical model is a straightforward grid discretization of the voxel-based (pixels with a third dimension) geometry which omits the boundary detection algorithms and allows reduced storage of the finite element data structure and a matrix-free solution algorithm. This in turn reduce the effort of almost all applied grid-based solution techniques and results in memory efficient and numerically stable algorithms for the microstructural models. Two variants of the matrix-free algorithm are presented. The efficient iterative solution method of conjugate gradients is used with matrix-free applicable preconditioners such as the Jacobi and the especially suited multigrid method. The jagged material boundaries of the voxel-based mesh are smoothed through embedded boundary elements which contain different material information at the integration point and are integrated sub-cell wise though without additional boundary detection. The efficiency of the matrix-free methods can be retained.}, subject = {Dissertation}, language = {en} } @phdthesis{ShaabanMohamed, author = {Shaaban Mohamed, Ahmed Mostafa}, title = {Isogeometric boundary element analysis and structural shape optimization for Helmholtz acoustic problems}, doi = {10.25643/bauhaus-universitaet.4703}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20220816-47030}, school = {Bauhaus-Universit{\"a}t Weimar}, abstract = {In this thesis, a new approach is developed for applications of shape optimization on the time harmonic wave propagation (Helmholtz equation) for acoustic problems. This approach is introduced for different dimensional problems: 2D, 3D axi-symmetric and fully 3D problems. The boundary element method (BEM) is coupled with the isogeometric analysis (IGA) forming the so-called (IGABEM) which speeds up meshing and gives higher accuracy in comparison with standard BEM. BEM is superior for handling unbounded domains by modeling only the inner boundaries and avoiding the truncation error, present in the finite element method (FEM) since BEM solutions satisfy the Sommerfeld radiation condition automatically. Moreover, BEM reduces the space dimension by one from a volumetric three-dimensional problem to a surface two-dimensional problem, or from a surface two-dimensional problem to a perimeter one-dimensional problem. Non-uniform rational B-splines basis functions (NURBS) are used in an isogeometric setting to describe both the CAD geometries and the physical fields. IGABEM is coupled with one of the gradient-free optimization methods, the Particle Swarm Optimization (PSO) for structural shape optimization problems. PSO is a straightforward method since it does not require any sensitivity analysis but it has some trade-offs with regard to the computational cost. Coupling IGA with optimization problems enables the NURBS basis functions to represent the three models: shape design, analysis and optimization models, by a definition of a set of control points to be the control variables and the optimization parameters as well which enables an easy transition between the three models. Acoustic shape optimization for various frequencies in different mediums is performed with PSO and the results are compared with the benchmark solutions from the literature for different dimensional problems proving the efficiency of the proposed approach with the following remarks: - In 2D problems, two BEM methods are used: the conventional isogeometric boundary element method (IGABEM) and the eXtended IGABEM (XIBEM) enriched with the partition-of-unity expansion using a set of plane waves, where the results are generally in good agreement with the linterature with some computation advantage to XIBEM which allows coarser meshes. -In 3D axi-symmetric problems, the three-dimensional problem is simplified in BEM from a surface integral to a combination of two 1D integrals. The first is the line integral similar to a two-dimensional BEM problem. The second integral is performed over the angle of revolution. The discretization is applied only to the former integration. This leads to significant computational savings and, consequently, better treatment for higher frequencies over the full three-dimensional models. - In fully 3D problems, a detailed comparison between two BEM methods: the conventional boundary integral equation (CBIE) and Burton-Miller (BM) is provided including the computational cost. The proposed models are enhanced with a modified collocation scheme with offsets to Greville abscissae to avoid placing collocation points at the corners. Placing collocation points on smooth surface enables accurate evaluation of normals for BM formulation in addition to straightforward prediction of jump-terms and avoids singularities in \$\mathcal{O} (1/r)\$ integrals eliminating the need for polar integration. Furthermore, no additional special treatment is required for the hyper-singular integral while collocating on highly distorted elements, such as those containing sphere poles. The obtained results indicate that, CBIE with PSO is a feasible alternative (except for a small number of fictitious frequencies) which is easier to implement. Furthermore, BM presents an outstanding treatment of the complicated geometry of mufflers with internal extended inlet/outlet tube as an interior 3D Helmholtz acoustic problem instead of using mixed or dual BEM.}, subject = {Randelemente-Methode}, language = {en} } @phdthesis{LopezZermeno, author = {L{\´o}pez Zerme{\~n}o, Jorge Alberto}, title = {Isogeometric and CAD-based methods for shape and topology optimization: Sensitivity analysis, B{\´e}zier elements and phase-field approaches}, doi = {10.25643/bauhaus-universitaet.4710}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20220831-47102}, school = {Bauhaus-Universit{\"a}t Weimar}, abstract = {The Finite Element Method (FEM) is widely used in engineering for solving Partial Differential Equations (PDEs) over complex geometries. To this end, it is required to provide the FEM software with a geometric model that is typically constructed in a Computer-Aided Design (CAD) software. However, FEM and CAD use different approaches for the mathematical description of the geometry. Thus, it is required to generate a mesh, which is suitable for FEM, based on the CAD model. Nonetheless, this procedure is not a trivial task and it can be time consuming. This issue becomes more significant for solving shape and topology optimization problems, which consist in evolving the geometry iteratively. Therefore, the computational cost associated to the mesh generation process is increased exponentially for this type of applications. The main goal of this work is to investigate the integration of CAD and CAE in shape and topology optimization. To this end, numerical tools that close the gap between design and analysis are presented. The specific objectives of this work are listed below: • Automatize the sensitivity analysis in an isogeometric framework for applications in shape optimization. Applications for linear elasticity are considered. • A methodology is developed for providing a direct link between the CAD model and the analysis mesh. In consequence, the sensitivity analysis can be performed in terms of the design variables located in the design model. • The last objective is to develop an isogeometric method for shape and topological optimization. This method should take advantage of using Non-Uniform Rational B-Splines (NURBS) with higher continuity as basis functions. Isogeometric Analysis (IGA) is a framework designed to integrate the design and analysis in engineering problems. The fundamental idea of IGA is to use the same basis functions for modeling the geometry, usually NURBS, for the approximation of the solution fields. The advantage of integrating design and analysis is two-fold. First, the analysis stage is more accurate since the system of PDEs is not solved using an approximated geometry, but the exact CAD model. Moreover, providing a direct link between the design and analysis discretizations makes possible the implementation of efficient sensitivity analysis methods. Second, the computational time is significantly reduced because the mesh generation process can be avoided. Sensitivity analysis is essential for solving optimization problems when gradient-based optimization algorithms are employed. Automatic differentiation can compute exact gradients, automatically by tracking the algebraic operations performed on the design variables. For the automation of the sensitivity analysis, an isogeometric framework is used. Here, the analysis mesh is obtained after carrying out successive refinements, while retaining the coarse geometry for the domain design. An automatic differentiation (AD) toolbox is used to perform the sensitivity analysis. The AD toolbox takes the code for computing the objective and constraint functions as input. Then, using a source code transformation approach, it outputs a code for computing the objective and constraint functions, and their sensitivities as well. The sensitivities obtained from the sensitivity propagation method are compared with analytical sensitivities, which are computed using a full isogeometric approach. The computational efficiency of AD is comparable to that of analytical sensitivities. However, the memory requirements are larger for AD. Therefore, AD is preferable if the memory requirements are satisfied. Automatic sensitivity analysis demonstrates its practicality since it simplifies the work of engineers and designers. Complex geometries with sharp edges and/or holes cannot easily be described with NURBS. One solution is the use of unstructured meshes. Simplex-elements (triangles and tetrahedra for two and three dimensions respectively) are particularly useful since they can automatically parameterize a wide variety of domains. In this regard, unstructured B{\´e}zier elements, commonly used in CAD, can be employed for the exact modelling of CAD boundary representations. In two dimensions, the domain enclosed by NURBS curves is parameterized with B{\´e}zier triangles. To describe exactly the boundary of a two-dimensional CAD model, the continuity of a NURBS boundary representation is reduced to C^0. Then, the control points are used to generate a triangulation such that the boundary of the domain is identical to the initial CAD boundary representation. Thus, a direct link between the design and analysis discretizations is provided and the sensitivities can be propagated to the design domain. In three dimensions, the initial CAD boundary representation is given as a collection of NURBS surfaces that enclose a volume. Using a mesh generator (Gmsh), a tetrahedral mesh is obtained. The original surface is reconstructed by modifying the location of the control points of the tetrahedral mesh using B{\´e}zier tetrahedral elements and a point inversion algorithm. This method offers the possibility of computing the sensitivity analysis using the analysis mesh. Then, the sensitivities can be propagated into the design discretization. To reuse the mesh originally generated, a moving B{\´e}zier tetrahedral mesh approach was implemented. A gradient-based optimization algorithm is employed together with a sensitivity propagation procedure for the shape optimization cases. The proposed shape optimization approaches are used to solve some standard benchmark problems in structural mechanics. The results obtained show that the proposed approach can compute accurate gradients and evolve the geometry towards optimal solutions. In three dimensions, the moving mesh approach results in faster convergence in terms of computational time and avoids remeshing at each optimization step. For considering topological changes in a CAD-based framework, an isogeometric phase-field based shape and topology optimization is developed. In this case, the diffuse interface of a phase-field variable over a design domain implicitly describes the boundaries of the geometry. The design variables are the local values of the phase-field variable. The descent direction to minimize the objective function is found by using the sensitivities of the objective function with respect to the design variables. The evolution of the phase-field is determined by solving the time dependent Allen-Cahn equation. Especially for topology optimization problems that require C^1 continuity, such as for flexoelectric structures, the isogeometric phase field method is of great advantage. NURBS can achieve the desired continuity more efficiently than the traditional employed functions. The robustness of the method is demonstrated when applied to different geometries, boundary conditions, and material configurations. The applications illustrate that compared to piezoelectricity, the electrical performance of flexoelectric microbeams is larger under bending. In contrast, the electrical power for a structure under compression becomes larger with piezoelectricity.}, subject = {CAD}, language = {en} } @phdthesis{NguyenThanh, author = {Nguyen-Thanh, Nhon}, title = {Isogeometric analysis based on rational splines over hierarchical T-mesh and alpha finite element method for structural analysis}, issn = {1610-7381}, doi = {10.25643/bauhaus-universitaet.2078}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20131125-20781}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {196}, abstract = {This thesis presents two new methods in finite elements and isogeometric analysis for structural analysis. The first method proposes an alternative alpha finite element method using triangular elements. In this method, the piecewise constant strain field of linear triangular finite element method models is enhanced by additional strain terms with an adjustable parameter a, which results in an effectively softer stiffness formulation compared to a linear triangular element. In order to avoid the transverse shear locking of Reissner-Mindlin plates analysis the alpha finite element method is coupled with a discrete shear gap technique for triangular elements to significantly improve the accuracy of the standard triangular finite elements. The basic idea behind this element formulation is to approximate displacements and rotations as in the standard finite element method, but to construct the bending, geometrical and shear strains using node-based smoothing domains. Several numerical examples are presented and show that the alpha FEM gives a good agreement compared to several other methods in the literature. Second method, isogeometric analysis based on rational splines over hierarchical T-meshes (RHT-splines) is proposed. The RHT-splines are a generalization of Non-Uniform Rational B-splines (NURBS) over hierarchical T-meshes, which is a piecewise bicubic polynomial over a hierarchical T-mesh. The RHT-splines basis functions not only inherit all the properties of NURBS such as non-negativity, local support and partition of unity but also more importantly as the capability of joining geometric objects without gaps, preserving higher order continuity everywhere and allow local refinement and adaptivity. In order to drive the adaptive refinement, an efficient recovery-based error estimator is employed. For this problem an imaginary surface is defined. The imaginary surface is basically constructed by RHT-splines basis functions which is used for approximation and interpolation functions as well as the construction of the recovered stress components. Numerical investigations prove that the proposed method is capable to obtain results with higher accuracy and convergence rate than NURBS results.}, subject = {Isogeometric analysis}, language = {en} } @phdthesis{Schwedler, author = {Schwedler, Michael}, title = {Integrated structural analysis using isogeometric finite element methods}, doi = {10.25643/bauhaus-universitaet.2737}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170130-27372}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {209}, abstract = {The gradual digitization in the architecture, engineering, and construction industry over the past fifty years led to an extremely heterogeneous software environment, which today is embodied by the multitude of different digital tools and proprietary data formats used by the many specialists contributing to the design process in a construction project. Though these projects become increasingly complex, the demands on financial efficiency and the completion within a tight schedule grow at the same time. The digital collaboration of project partners has been identified as one key issue in successfully dealing with these challenges. Yet currently, the numerous software applications and their respective individual views on the design process severely impede that collaboration. An approach to establish a unified basis for the digital collaboration, regardless of the existing software heterogeneity, is a comprehensive digital building model contributed to by all projects partners. This type of data management known as building information modeling (BIM) has many benefits, yet its adoption is associated with many difficulties and thus, proceeds only slowly. One aspect in the field of conflicting requirements on such a digital model is the cooperation of architects and structural engineers. Traditionally, these two disciplines use different abstractions of reality for their models that in consequence lead to incompatible digital representations thereof. The onset of isogeometric analysis (IGA) promised to ease the discrepancy in design and analysis model representations. Yet, that initial focus quickly shifted towards using these methods as a more powerful basis for numerical simulations. Furthermore, the isogeometric representation alone is not capable of solving the model abstraction problem. It is thus the intention of this work to contribute to an improved digital collaboration of architects and engineers by exploring an integrated analysis approach on the basis of an unified digital model and solid geometry expressed by splines. In the course of this work, an analysis framework is developed that utilizes such models to automatically conduct numerical simulations commonly required in construction projects. In essence, this allows to retrieve structural analysis results from BIM models in a fast and simple manner, thereby facilitating rapid design iterations and profound design feedback. The BIM implementation Industry Foundation Classes (IFC) is reviewed with regard to its capabilities of representing the unified model. The current IFC schema strongly supports the use of redundant model data, a major pitfall in digital collaboration. Additionally, it does not allow to describe the geometry by volumetric splines. As the pursued approach builds upon a unique model for both, architectural and structural design, and furthermore requires solid geometry, necessary schema modifications are suggested. Structural entities are modeled by volumetric NURBS patches, each of which constitutes an individual subdomain that, with regard to the analysis, is incompatible with the remaining full model. The resulting consequences for numerical simulation are elaborated in this work. The individual subdomains have to be weakly coupled, for which the mortar method is used. Different approaches to discretize the interface traction fields are implemented and their respective impact on the analysis results is evaluated. All necessary coupling conditions are automatically derived from the related geometry model. The weak coupling procedure leads to a linear system of equations in saddle point form, which, owed to the volumetric modeling, is large in size and, the associated coefficient matrix has, due to the use of higher degree basis functions, a high bandwidth. The peculiarities of the system require adapted solution methods that generally cause higher numerical costs than the standard procedures for symmetric, positive-definite systems do. Different methods to solve the specific system are investigated and an efficient parallel algorithm is finally proposed. When the structural analysis model is derived from the unified model in the BIM data, it does in general initially not meet the requirements on the discretization that are necessary to obtain sufficiently accurate analysis results. The consequently necessary patch refinements must be controlled automatically to allowfor an entirely automatic analysis procedure. For that purpose, an empirical refinement scheme based on the geometrical and possibly mechanical properties of the specific entities is proposed. The level of refinement may be selectively manipulated by the structural engineer in charge. Furthermore, a Zienkiewicz-Zhu type error estimator is adapted for the use with isogeometric analysis results. It is shown that also this estimator can be used to steer an adaptive refinement procedure.}, subject = {Finite-Elemente-Methode}, language = {en} } @phdthesis{Abeltshauser, author = {Abeltshauser, Rainer}, title = {Identification and separation of physical effects of coupled systems by using defined model abstractions}, doi = {10.25643/bauhaus-universitaet.2860}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20170314-28600}, school = {Bauhaus-Universit{\"a}t Weimar}, abstract = {The thesis investigates at the computer aided simulation process for operational vibration analysis of complex coupled systems. As part of the internal methods project "Absolute Values" of the BMW Group, the thesis deals with the analysis of the structural dynamic interactions and excitation interactions. The overarching aim of the methods project is to predict the operational vibrations of engines. Simulations are usually used to analyze technical aspects (e. g. operational vibrations, strength, ...) of single components in the industrial development. The boundary conditions of submodels are mostly based on experiences. So the interactions with neighboring components and systems are neglected. To get physically more realistic results but still efficient simulations, this work wants to support the engineer during the preprocessing phase by useful criteria. At first suitable abstraction levels based on the existing literature are defined to identify structural dynamic interactions and excitation interactions of coupled systems. So it is possible to separate different effects of the coupled subsystems. On this basis, criteria are derived to assess the influence of interactions between the considered systems. These criteria can be used during the preprocessing phase and help the engineer to build up efficient models with respect to the interactions with neighboring systems. The method was developed by using several models with different complexity levels. Furthermore, the method is proved for the application in the industrial environment by using the example of a current combustion engine.}, subject = {Strukturdynamik}, language = {en} } @phdthesis{Schrader, author = {Schrader, Kai}, title = {Hybrid 3D simulation methods for the damage analysis of multiphase composites}, doi = {10.25643/bauhaus-universitaet.2059}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20131021-20595}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {174}, abstract = {Modern digital material approaches for the visualization and simulation of heterogeneous materials allow to investigate the behavior of complex multiphase materials with their physical nonlinear material response at various scales. However, these computational techniques require extensive hardware resources with respect to computing power and main memory to solve numerically large-scale discretized models in 3D. Due to a very high number of degrees of freedom, which may rapidly be increased to the two-digit million range, the limited hardware ressources are to be utilized in a most efficient way to enable an execution of the numerical algorithms in minimal computation time. Hence, in the field of computational mechanics, various methods and algorithms can lead to an optimized runtime behavior of nonlinear simulation models, where several approaches are proposed and investigated in this thesis. Today, the numerical simulation of damage effects in heterogeneous materials is performed by the adaption of multiscale methods. A consistent modeling in the three-dimensional space with an appropriate discretization resolution on each scale (based on a hierarchical or concurrent multiscale model), however, still contains computational challenges in respect to the convergence behavior, the scale transition or the solver performance of the weak coupled problems. The computational efficiency and the distribution among available hardware resources (often based on a parallel hardware architecture) can significantly be improved. In the past years, high-performance computing (HPC) and graphics processing unit (GPU) based computation techniques were established for the investigationof scientific objectives. Their application results in the modification of existing and the development of new computational methods for the numerical implementation, which enables to take advantage of massively clustered computer hardware resources. In the field of numerical simulation in material science, e.g. within the investigation of damage effects in multiphase composites, the suitability of such models is often restricted by the number of degrees of freedom (d.o.f.s) in the three-dimensional spatial discretization. This proves to be difficult for the type of implementation method used for the nonlinear simulation procedure and, simultaneously has a great influence on memory demand and computational time. In this thesis, a hybrid discretization technique has been developed for the three-dimensional discretization of a three-phase material, which is respecting the numerical efficiency of nonlinear (damage) simulations of these materials. The increase of the computational efficiency is enabled by the improved scalability of the numerical algorithms. Consequently, substructuring methods for partitioning the hybrid mesh were implemented, tested and adapted to the HPC computing framework using several hundred CPU (central processing units) nodes for building the finite element assembly. A memory-efficient iterative and parallelized equation solver combined with a special preconditioning technique for solving the underlying equation system was modified and adapted to enable combined CPU and GPU based computations. Hence, it is recommended by the author to apply the substructuring method for hybrid meshes, which respects different material phases and their mechanical behavior and which enables to split the structure in elastic and inelastic parts. However, the consideration of the nonlinear material behavior, specified for the corresponding phase, is limited to the inelastic domains only, and by that causes a decreased computing time for the nonlinear procedure. Due to the high numerical effort for such simulations, an alternative approach for the nonlinear finite element analysis, based on the sequential linear analysis, was implemented in respect to scalable HPC. The incremental-iterative procedure in finite element analysis (FEA) during the nonlinear step was then replaced by a sequence of linear FE analysis when damage in critical regions occured, known in literature as saw-tooth approach. As a result, qualitative (smeared) crack initiation in 3D multiphase specimens has efficiently been simulated.}, subject = {high-performance computing}, language = {en} } @phdthesis{Habtemariam, author = {Habtemariam, Abinet Kifle}, title = {Generalized Beam Theory for the analysis of thin-walled circular pipe members}, doi = {10.25643/bauhaus-universitaet.4572}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20220127-45723}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {188}, abstract = {The detailed structural analysis of thin-walled circular pipe members often requires the use of a shell or solid-based finite element method. Although these methods provide a very good approximation of the deformations, they require a higher degree of discretization which causes high computational costs. On the other hand, the analysis of thin-walled circular pipe members based on classical beam theories is easy to implement and needs much less computation time, however, they are limited in their ability to approximate the deformations as they cannot consider the deformation of the cross-section. This dissertation focuses on the study of the Generalized Beam Theory (GBT) which is both accurate and efficient in analyzing thin-walled members. This theory is based on the separation of variables in which the displacement field is expressed as a combination of predetermined deformation modes related to the cross-section, and unknown amplitude functions defined on the beam's longitudinal axis. Although the GBT was initially developed for long straight members, through the consideration of complementary deformation modes, which amend the null transverse and shear membrane strain assumptions of the classical GBT, problems involving short members, pipe bends, and geometrical nonlinearity can also be analyzed using GBT. In this dissertation, the GBT formulation for the analysis of these problems is developed and the application and capabilities of the method are illustrated using several numerical examples. Furthermore, the displacement and stress field results of these examples are verified using an equivalent refined shell-based finite element model. The developed static and dynamic GBT formulations for curved thin-walled circular pipes are based on the linear kinematic description of the curved shell theory. In these formulations, the complex problem in pipe bends due to the strong coupling effect of the longitudinal bending, warping and the cross-sectional ovalization is handled precisely through the derivation of the coupling tensors between the considered GBT deformation modes. Similarly, the geometrically nonlinear GBT analysis is formulated for thin-walled circular pipes based on the nonlinear membrane kinematic equations. Here, the initial linear and quadratic stress and displacement tangent stiffness matrices are built using the third and fourth-order GBT deformation mode coupling tensors. Longitudinally, the formulation of the coupled GBT element stiffness and mass matrices are presented using a beam-based finite element formulation. Furthermore, the formulated GBT elements are tested for shear and membrane locking problems and the limitations of the formulations regarding the membrane locking problem are discussed.}, subject = {Finite-Elemente-Methode}, language = {en} } @phdthesis{Ashour, author = {Ashour, Mohammed}, title = {Electromechanics and Hydrodynamics of Single Vesicles and Vesicle Doublet Using Phase-Field Isogeometric Analysis}, doi = {10.25643/bauhaus-universitaet.6400}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20230628-64003}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {175}, abstract = {Biomembranes are selectively permeable barriers that separate the internal components of the cell from its surroundings. They have remarkable mechanical behavior which is characterized by many phenomena, but most noticeably their fluid-like in-plane behavior and solid-like out-of-plane behavior. Vesicles have been studied in the context of discrete models, such as Molecular Dynamics, Monte Carlo methods, Dissipative Particle Dynamics, and Brownian Dynamics. Those methods, however, tend to have high computational costs, which limited their uses for studying atomistic details. In order to broaden the scope of this research, we resort to the continuum models, where the atomistic details of the vesicles are neglected, and the focus shifts to the overall morphological evolution. Under the umbrella of continuum models, vesicles morphology has been studied extensively. However, most of those studies were limited to the mechanical response of vesicles by considering only the bending energy and aiming for the solution by minimizing the total energy of the system. Most of the literature is divided between two geometrical representation methods; the sharp interface methods and the diffusive interface methods. Both of those methods track the boundaries and interfaces implicitly. In this research, we focus our attention on solving two non-trivial problems. In the first one, we study a constrained Willmore problem coupled with an electrical field, and in the second one, we investigate the hydrodynamics of a vesicle doublet suspended in an external viscous fluid flow. For the first problem, we solve a constrained Willmore problem coupled with an electrical field using isogeometric analysis to study the morphological evolution of vesicles subjected to static electrical fields. The model comprises two phases, the lipid bilayer, and the electrolyte. This two-phase problem is modeled using the phase-field method, which is a subclass of the diffusive interface methods mentioned earlier. The bending, flexoelectric, and dielectric energies of the model are reformulated using the phase-field parameter. A modified Augmented-Lagrangian (ALM) approach was used to satisfy the constraints while maintaining numerical stability and a relatively large time step. This approach guarantees the satisfaction of the constraints at each time step over the entire temporal domain. In the second problem, we study the hydrodynamics of vesicle doublet suspended in an external viscous fluid flow. Vesicles in this part of the research are also modeled using the phase-field model. The bending energy and energies associated with enforcing the global volume and area are considered. In addition, the local inextensibility condition is ensured by introducing an additional equation to the system. To prevent the vesicles from numerically overlapping, we deploy an interaction energy definition to maintain a short-range repulsion between the vesicles. The fluid flow is modeled using the incompressible Navier-Stokes equations and the vesicle evolution in time is modeled using two advection equations describing the process of advecting each vesicle by the fluid flow. To overcome the velocity-pressure saddle point system, we apply the Residual-Based Variational MultiScale (RBVMS) method to the Navier-Stokes equations and solve the coupled systems using isogeometric analysis. We study vesicle doublet hydrodynamics in shear flow, planar extensional flow, and parabolic flow under various configurations and boundary conditions. The results reveal several interesting points about the electrodynamics and hydrodynamics responses of single vesicles and vesicle doublets. But first, it can be seen that isogeometric analysis as a numerical tool has the ability to model and solve 4th-order PDEs in a primal variational framework at extreme efficiency and accuracy due to the abilities embedded within the NURBS functions without the need to reduce the order of the PDE by creating an intermediate environment. Refinement whether by knot insertion, order increasing or both is far easier to obtain than traditional mesh-based methods. Given the wide variety of phenomena in natural sciences and engineering that are mathematically modeled by high-order PDEs, the isogeometric analysis is among the most robust methods to address such problems as the basis functions can easily attain high global continuity. On the applicational side, we study the vesicle morphological evolution based on the electromechanical liquid-crystal model in 3D settings. This model describing the evolution of vesicles is composed of time-dependent, highly nonlinear, high-order PDEs, which are nontrivial to solve. Solving this problem requires robust numerical methods, such as isogeometric analysis. We concluded that the vesicle tends to deform under increasing magnitudes of electric fields from the original sphere shape to an oblate-like shape. This evolution is affected by many factors and requires fine-tuning of several parameters, mainly the regularization parameter which controls the thickness of the diffusive interface width. But it is most affected by the method used for enforcing the constraints. The penalty method in presence of an electrical field tends to lock on the initial phase-field and prevent any evolution while a modified version of the ALM has proven to be sufficiently stable and accurate to let the phase-field evolve while satisfying the constraints over time at each time step. We show additionally the effect of including the flexoelectric nature of the Biomembranes in the computation and how it affects the shape evolution as well as the effect of having different conductivity ratios. All the examples were solved based on a staggered scheme, which reduces the computational cost significantly. For the second part of the research, we consider vesicle doublet suspended in a shear flow, in a planar extensional flow, and in a parabolic flow. When the vesicle doublet is suspended in a shear flow, it can either slip past each other or slide on top of each other based on the value of the vertical displacement, that is the vertical distance between the center of masses between the two vesicles, and the velocity profile applied. When the vesicle doublet is suspended in a planar extensional flow in a configuration that resembles a junction, the time in which both vesicles separate depends largely on the value of the vertical displacement after displacing as much fluid from between the two vesicles. However, when the vesicles are suspended in a tubular channel with a parabolic fluid flow, they develop a parachute-like shape upon converging towards each other before exiting the computational domain from the predetermined outlets. This shape however is affected largely by the height of the tubular channel in which the vesicle is suspended. The velocity essential boundary conditions are imposed weakly and strongly. The weak implementation of the boundary conditions was used when the velocity profile was defined on the entire boundary, while the strong implementation was used when the velocity profile was defined on a part of the boundary. The strong implementation of the essential boundary conditions was done by selectively applying it to the predetermined set of elements in a parallel-based code. This allowed us to simulate vesicle hydrodynamics in a computational domain with multiple inlets and outlets. We also investigate the hydrodynamics of oblate-like shape vesicles in a parabolic flow. This work has been done in 2D configuration because of the immense computational load resulting from a large number of degrees of freedom, but we are actively seeking to expand it to 3D settings and test a broader set of parameters and geometrical configurations.}, subject = {Isogeometrische Analyse}, language = {en} } @phdthesis{Valizadeh, author = {Valizadeh, Navid}, title = {Developments in Isogeometric Analysis and Application to High-Order Phase-Field Models of Biomembranes}, doi = {10.25643/bauhaus-universitaet.4565}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20220114-45658}, school = {Bauhaus-Universit{\"a}t Weimar}, abstract = {Isogeometric analysis (IGA) is a numerical method for solving partial differential equations (PDEs), which was introduced with the aim of integrating finite element analysis with computer-aided design systems. The main idea of the method is to use the same spline basis functions which describe the geometry in CAD systems for the approximation of solution fields in the finite element method (FEM). Originally, NURBS which is a standard technology employed in CAD systems was adopted as basis functions in IGA but there were several variants of IGA using other technologies such as T-splines, PHT splines, and subdivision surfaces as basis functions. In general, IGA offers two key advantages over classical FEM: (i) by describing the CAD geometry exactly using smooth, high-order spline functions, the mesh generation process is simplified and the interoperability between CAD and FEM is improved, (ii) IGA can be viewed as a high-order finite element method which offers basis functions with high inter-element continuity and therefore can provide a primal variational formulation of high-order PDEs in a straightforward fashion. The main goal of this thesis is to further advance isogeometric analysis by exploiting these major advantages, namely precise geometric modeling and the use of smooth high-order splines as basis functions, and develop robust computational methods for problems with complex geometry and/or complex multi-physics. As the first contribution of this thesis, we leverage the precise geometric modeling of isogeometric analysis and propose a new method for its coupling with meshfree discretizations. We exploit the strengths of both methods by using IGA to provide a smooth, geometrically-exact surface discretization of the problem domain boundary, while the Reproducing Kernel Particle Method (RKPM) discretization is used to provide the volumetric discretization of the domain interior. The coupling strategy is based upon the higher-order consistency or reproducing conditions that are directly imposed in the physical domain. The resulting coupled method enjoys several favorable features: (i) it preserves the geometric exactness of IGA, (ii) it circumvents the need for global volumetric parameterization of the problem domain, (iii) it achieves arbitrary-order approximation accuracy while preserving higher-order smoothness of the discretization. Several numerical examples are solved to show the optimal convergence properties of the coupled IGA-RKPM formulation, and to demonstrate its effectiveness in constructing volumetric discretizations for complex-geometry objects. As for the next contribution, we exploit the use of smooth, high-order spline basis functions in IGA to solve high-order surface PDEs governing the morphological evolution of vesicles. These governing equations are often consisted of geometric PDEs, high-order PDEs on stationary or evolving surfaces, or a combination of them. We propose an isogeometric formulation for solving these PDEs. In the context of geometric PDEs, we consider phase-field approximations of mean curvature flow and Willmore flow problems and numerically study the convergence behavior of isogeometric analysis for these problems. As a model problem for high-order PDEs on stationary surfaces, we consider the Cahn-Hilliard equation on a sphere, where the surface is modeled using a phase-field approach. As for the high-order PDEs on evolving surfaces, a phase-field model of a deforming multi-component vesicle, which consists of two fourth-order nonlinear PDEs, is solved using the isogeometric analysis in a primal variational framework. Through several numerical examples in 2D, 3D and axisymmetric 3D settings, we show the robustness of IGA for solving the considered phase-field models. Finally, we present a monolithic, implicit formulation based on isogeometric analysis and generalized-alpha time integration for simulating hydrodynamics of vesicles according to a phase-field model. Compared to earlier works, the number of equations of the phase-field model which need to be solved is reduced by leveraging high continuity of NURBS functions, and the algorithm is extended to 3D settings. We use residual-based variational multi-scale method (RBVMS) for solving Navier-Stokes equations, while the rest of PDEs in the phase-field model are treated using a standard Galerkin-based IGA. We introduce the resistive immersed surface (RIS) method into the formulation which can be employed for an implicit description of complex geometries using a diffuse-interface approach. The implementation highlights the robustness of the RBVMS method for Navier-Stokes equations of incompressible flows with non-trivial localized forcing terms including bending and tension forces of the vesicle. The potential of the phase-field model and isogeometric analysis for accurate simulation of a variety of fluid-vesicle interaction problems in 2D and 3D is demonstrated.}, subject = {Phasenfeldmodell}, language = {en} } @phdthesis{Vollmering, author = {Vollmering, Max}, title = {Damage Localization of Mechanical Structures by Subspace Identification and Krein Space Based H-infinity Estimation}, doi = {10.25643/bauhaus-universitaet.3772}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20180730-37728}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {205}, abstract = {This dissertation is devoted to the theoretical development and experimental laboratory verification of a new damage localization method: The state projection estimation error (SP2E). This method is based on the subspace identification of mechanical structures, Krein space based H-infinity estimation and oblique projections. To explain method SP2E, several theories are discussed and laboratory experiments have been conducted and analysed. A fundamental approach of structural dynamics is outlined first by explaining mechanical systems based on first principles. Following that, a fundamentally different approach, subspace identification, is comprehensively explained. While both theories, first principle and subspace identification based mechanical systems, may be seen as widespread methods, barely known and new techniques follow up. Therefore, the indefinite quadratic estimation theory is explained. Based on a Popov function approach, this leads to the Krein space based H-infinity theory. Subsequently, a new method for damage identification, namely SP2E, is proposed. Here, the introduction of a difference process, the analysis by its average process power and the application of oblique projections is discussed in depth. Finally, the new method is verified in laboratory experiments. Therefore, the identification of a laboratory structure at Leipzig University of Applied Sciences is elaborated. Then structural alterations are experimentally applied, which were localized by SP2E afterwards. In the end four experimental sensitivity studies are shown and discussed. For each measurement series the structural alteration was increased, which was successfully tracked by SP2E. The experimental results are plausible and in accordance with the developed theories. By repeating these experiments, the applicability of SP2E for damage localization is experimentally proven.}, subject = {Strukturmechanik}, language = {en} } @phdthesis{ZHANG2018, author = {ZHANG, CHAO}, title = {Crack Identification using Dynamic Extended Finite Element Method and Thermal Conductivity Engineering for Nanomaterials}, doi = {10.25643/bauhaus-universitaet.3847}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20190119-38478}, school = {Bauhaus-Universit{\"a}t Weimar}, year = {2018}, abstract = {Identification of flaws in structures is a critical element in the management of maintenance and quality assurance processes in engineering. Nondestructive testing (NDT) techniques based on a wide range of physical principles have been developed and are used in common practice for structural health monitoring. However, basic NDT techniques are usually limited in their ability to provide the accurate information on locations, dimensions and shapes of flaws. One alternative to extract additional information from the results of NDT is to append it with a computational model that provides detailed analysis of the physical process involved and enables the accurate identification of the flaw parameters. The aim here is to develop the strategies to uniquely identify cracks in two-dimensional 2D) structures under dynamic loadings. A local NDT technique combined eXtended Finite Element Method (XFEM) with dynamic loading in order to identify the cracks in the structures quickly and accurately is developed in this dissertation. The Newmark-b time integration method with Rayleigh damping is used for the time integration. We apply Nelder-Mead (NM)and Quasi-Newton (QN) methods for identifying the crack tip in plate. The inverse problem is solved iteratively, in which XFEM is used for solving the forward problem in each iteration. For a timeharmonic excitation with a single frequency and a short-duration signal measured along part of the external boundary, the crack is detected through the solution of an inverse time-dependent problem. Compared to the static load, we show that the dynamic loads are more effective for crack detection problems. Moreover, we tested different dynamic loads and find that NM method works more efficient under the harmonic load than the pounding load while the QN method achieves almost the same results for both load types. A global strategy, Multilevel Coordinate Search (MCS) with XFEM (XFEM-MCS) methodology under the dynamic electric load, to detect multiple cracks in 2D piezoelectric plates is proposed in this dissertation. The Newmark-b method is employed for the time integration and in each iteration the forward problem is solved by XFEM for various cracks. The objective functional is minimized by using a global search algorithm MCS. The test problems show that the XFEM-MCS algorithm under the dynamic electric load can be effectively employed for multiple cracks detection in piezoelectric materials, and it proves to be robust in identifying defects in piezoelectric structures. Fiber-reinforced composites (FRCs) are extensively applied in practical engineering since they have high stiffness and strength. Experiments reveal a so-called interphase zone, i.e. the space between the outside interface of the fiber and the inside interface of the matrix. The interphase strength between the fiber and the matrix strongly affects the mechanical properties as a result of the large ratio of interface/volume. For the purpose of understanding the mechanical properties of FRCs with functionally graded interphase (FGI), a closed-form expression of the interface strength between a fiber and a matrix is obtained in this dissertation using a continuum modeling approach according to the ver derWaals (vdW) forces. Based on the interatomic potential, we develop a new modified nonlinear cohesive law, which is applied to study the interface delamination of FRCs with FGI under different loadings. The analytical solutions show that the delamination behavior strongly depends on the interphase thickness, the fiber radius, the Young's moduli and Poisson's ratios of the fiber and the matrix. Thermal conductivity is the property of a material to conduct heat. With the development and deep research of 2D materials, especially graphene and molybdenum disulfide (MoS2), the thermal conductivity of 2D materials attracts wide attentions. The thermal conductivity of graphene nanoribbons (GNRs) is found to appear a tendency of decreasing under tensile strain by classical molecular dynamics (MD) simulations. Hence, the strain effects of graphene can play a key role in the continuous tunability and applicability of its thermal conductivity property at nanoscale, and the dissipation of thermal conductivity is an obstacle for the applications of thermal management. Up to now, the thermal conductivity of graphene under shear deformation has not been investigated yet. From a practical point of view, good thermal managements of GNRs have significantly potential applications of future GNR-based thermal nanodevices, which can greatly improve performances of the nanosized devices due to heat dissipations. Meanwhile, graphene is a thin membrane structure, it is also important to understand the wrinkling behavior under shear deformation. MoS2 exists in the stable semiconducting 1H phase (1H-MoS2) while the metallic 1T phase (1T-MoS2) is unstable at ambient conditions. As it's well known that much attention has been focused on studying the nonlinear optical properties of the 1H-MoS2. In a very recent research, the 1T-type monolayer crystals of TMDCs, MX2 (MoS2, WS2 ...) was reported having an intrinsic in-plane negative Poisson's ratio. Luckily, nearly at the same time, unprecedented long-term (>3months) air stability of the 1T-MoS2 can be achieved by using the donor lithium hydride (LiH). Therefore, it's very important to study the thermal conductivity of 1T-MoS2. The thermal conductivity of graphene under shear strain is systematically studied in this dissertation by MD simulations. The results show that, in contrast to the dramatic decrease of thermal conductivity of graphene under uniaxial tensile, the thermal conductivity of graphene is not sensitive to the shear strain, and the thermal conductivity decreases only 12-16\%. The wrinkle evolves when the shear strain is around 5\%-10\%, but the thermal conductivity barely changes. The thermal conductivities of single-layer 1H-MoS2(1H-SLMoS2) and single-layer 1T-MoS2 (1T-SLMoS2) with different sample sizes, temperatures and strain rates have been studied systematically in this dissertation. We find that the thermal conductivities of 1H-SLMoS2 and 1T-SLMoS2 in both the armchair and the zigzag directions increase with the increasing of the sample length, while the increase of the width of the sample has minor effect on the thermal conductions of these two structures. The thermal conductivity of 1HSLMoS2 is smaller than that of 1T-SLMoS2 under size effect. Furthermore, the temperature effect results show that the thermal conductivities of both 1H-SLMoS2 and 1T-SLMoS2 decrease with the increasing of the temperature. The thermal conductivities of 1HSLMoS2 and 1T-SLMoS2 are nearly the same (difference <6\%) in both of the chiral orientations under corresponding temperatures, especially in the armchair direction (difference <2.8\%). Moreover, we find that the strain effects on the thermal conductivity of 1HSLMoS2 and 1T-SLMoS2 are different. More specifically, the thermal conductivity decreases with the increasing tensile strain rate for 1T-SLMoS2, while fluctuates with the growth of the strain for 1HSLMoS2. Finally, we find that the thermal conductivity of same sized 1H-SLMoS2 is similar with that of the strained 1H-SLMoS2 structure.}, subject = {crack}, language = {en} } @phdthesis{Bianco, author = {Bianco, Marcelo Jos{\´e}}, title = {Coupling between Shell and Generalized Beam Theory (GBT) elements}, doi = {10.25643/bauhaus-universitaet.4391}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20210315-43914}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {265}, abstract = {In the last decades, Finite Element Method has become the main method in statics and dynamics analysis in engineering practice. For current problems, this method provides a faster, more flexible solution than the analytic approach. Prognoses of complex engineer problems that used to be almost impossible to solve are now feasible. Although the finite element method is a robust tool, it leads to new questions about engineering solutions. Among these new problems, it is possible to divide into two major groups: the first group is regarding computer performance; the second one is related to understanding the digital solution. Simultaneously with the development of the finite element method for numerical solutions, a theory between beam theory and shell theory was developed: Generalized Beam Theory, GBT. This theory has not only a systematic and analytical clear presentation of complicated structural problems, but also a compact and elegant calculation approach that can improve computer performance. Regrettably, GBT was not internationally known since the most publications of this theory were written in German, especially in the first years. Only in recent years, GBT has gradually become a fertile research topic, with developments from linear to non-linear analysis. Another reason for the misuse of GBT is the isolated application of the theory. Although recently researches apply finite element method to solve the GBT's problems numerically, the coupling between finite elements of GBT and other theories (shell, solid, etc) is not the subject of previous research. Thus, the main goal of this dissertation is the coupling between GBT and shell/membrane elements. Consequently, one achieves the benefits of both sides: the versatility of shell elements with the high performance of GBT elements. Based on the assumptions of GBT, this dissertation presents how the separation of variables leads to two calculation's domains of a beam structure: a cross-section modal analysis and the longitudinal amplification axis. Therefore, there is the possibility of applying the finite element method not only in the cross-section analysis, but also the development for an exact GBT's finite element in the longitudinal direction. For the cross-section analysis, this dissertation presents the solution of the quadratic eigenvalue problem with an original separation between plate and membrane mechanism. Subsequently, one obtains a clearer representation of the deformation mode, as well as a reduced quadratic eigenvalue problem. Concerning the longitudinal direction, this dissertation develops the novel exact elements, based on hyperbolic and trigonometric shape functions. Although these functions do not have trivial expressions, they provide a recursive procedure that allows periodic derivatives to systematise the development of stiffness matrices. Also, these shape functions enable a single-element discretisation of the beam structure and ensure a smooth stress field. From these developments, this dissertation achieves the formulation of its primary objective: the connection of GBT and shell elements in a mixed model. Based on the displacement field, it is possible to define the coupling equations applied in the master-slave method. Therefore, one can model the structural connections and joints with finite shell elements and the structural beams and columns with GBT finite element. As a side effect, the coupling equations limit the displacement field of the shell elements under the assumptions of GBT, in particular in the neighbourhood of the coupling cross-section. Although these side effects are almost unnoticeable in linear analysis, they lead to cumulative errors in non-linear analysis. Therefore, this thesis finishes with the evaluation of the mixed GBT-shell models in non-linear analysis.}, subject = {Biegetheorie}, language = {en} } @phdthesis{AbuBakar, author = {Abu Bakar, Ilyani Akmar}, title = {Computational Analysis of Woven Fabric Composites: Single- and Multi-Objective Optimizations and Sensitivity Analysis in Meso-scale Structures}, issn = {1610-7381}, doi = {10.25643/bauhaus-universitaet.4176}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20200605-41762}, school = {Bauhaus-Universit{\"a}t Weimar}, pages = {151}, abstract = {This study permits a reliability analysis to solve the mechanical behaviour issues existing in the current structural design of fabric structures. Purely predictive material models are highly desirable to facilitate an optimized design scheme and to significantly reduce time and cost at the design stage, such as experimental characterization. The present study examined the role of three major tasks; a) single-objective optimization, b) sensitivity analyses and c) multi-objective optimization on proposed weave structures for woven fabric composites. For single-objective optimization task, the first goal is to optimize the elastic properties of proposed complex weave structure under unit cells basis based on periodic boundary conditions. We predict the geometric characteristics towards skewness of woven fabric composites via Evolutionary Algorithm (EA) and a parametric study. We also demonstrate the effect of complex weave structures on the fray tendency in woven fabric composites via tightness evaluation. We utilize a procedure which does not require a numerical averaging process for evaluating the elastic properties of woven fabric composites. The fray tendency and skewness of woven fabrics depends upon the behaviour of the floats which is related to the factor of weave. Results of this study may suggest a broader view for further research into the effects of complex weave structures or may provide an alternative to the fray and skewness problems of current weave structure in woven fabric composites. A comprehensive study is developed on the complex weave structure model which adopts the dry woven fabric of the most potential pattern in singleobjective optimization incorporating the uncertainties parameters of woven fabric composites. The comprehensive study covers the regression-based and variance-based sensitivity analyses. The second task goal is to introduce the fabric uncertainties parameters and elaborate how they can be incorporated into finite element models on macroscopic material parameters such as elastic modulus and shear modulus of dry woven fabric subjected to uni-axial and biaxial deformations. Significant correlations in the study, would indicate the need for a thorough investigation of woven fabric composites under uncertainties parameters. The study describes here could serve as an alternative to identify effective material properties without prolonged time consumption and expensive experimental tests. The last part focuses on a hierarchical stochastic multi-scale optimization approach (fine-scale and coarse-scale optimizations) under geometrical uncertainties parameters for hybrid composites considering complex weave structure. The fine-scale optimization is to determine the best lamina pattern that maximizes its macroscopic elastic properties, conducted by EA under the following uncertain mesoscopic parameters: yarn spacing, yarn height, yarn width and misalignment of yarn angle. The coarse-scale optimization has been carried out to optimize the stacking sequences of symmetric hybrid laminated composite plate with uncertain mesoscopic parameters by employing the Ant Colony Algorithm (ACO). The objective functions of the coarse-scale optimization are to minimize the cost (C) and weight (W) of the hybrid laminated composite plate considering the fundamental frequency and the buckling load factor as the design constraints. Based on the uncertainty criteria of the design parameters, the appropriate variation required for the structural design standards can be evaluated using the reliability tool, and then an optimized design decision in consideration of cost can be subsequently determined.}, subject = {Verbundwerkstoff}, language = {en} } @phdthesis{Mojahedin, author = {Mojahedin, Arvin}, title = {Analysis of Functionally Graded Porous Materials Using Deep Energy Method and Analytical Solution}, doi = {10.25643/bauhaus-universitaet.4867}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20221220-48674}, school = {Bauhaus-Universit{\"a}t Weimar}, abstract = {Porous materials are an emerging branch of engineering materials that are composed of two elements: One element is a solid (matrix), and the other element is either liquid or gas. Pores can be distributed within the solid matrix of porous materials with different shapes and sizes. In addition, porous materials are lightweight, and flexible, and have higher resistance to crack propagation and specific thermal, mechanical, and magnetic properties. These properties are necessary for manufacturing engineering structures such as beams and other engineering structures. These materials are widely used in solid mechanics and are considered a good replacement for classical materials by many researchers recently. Producing lightweight materials has been developed because of the possibility of exploiting the properties of these materials. Various types of porous material are generated naturally or artificially for a specific application such as bones and foams. Like functionally graded materials, pore distribution patterns can be uniform or non-uniform. Biot's theory is a well-developed theory to study the behavior of poroelastic materials which investigates the interaction between fluid and solid phases of a fluid-saturated porous medium. Functionally graded porous materials (FGPM) are widely used in modern industries, such as aerospace, automotive, and biomechanics. These advanced materials have some specific properties compared to materials with a classic structure. They are extremely light, while they have specific strength in mechanical and high-temperature environments. FGPMs are characterized by a gradual variation of material parameters over the volume. Although these materials can be made naturally, it is possible to design and manufacture them for a specific application. Therefore, many studies have been done to analyze the mechanical and thermal properties of FGPM structures, especially beams. Biot was the pioneer in formulating the linear elasticity and thermoelasticity equations of porous material. Since then, Biot's formulation has been developed in continuum mechanics which is named poroelasticity. There are obstacles to analyzing the behavior of these materials accurately like the shape of the pores, the distribution of pores in the material, and the behavior of the fluid (or gas) that saturated pores. Indeed, most of the engineering structures made of FGPM have nonlinear governing equations. Therefore, it is difficult to study engineering structures by solving these complicated equations. The main purpose of this dissertation is to analyze porous materials in engineering structures. For this purpose, the complex equations of porous materials have been simplified and applied to engineering problems so that the effect of all parameters of porous materials on the behavior of engineering structure has been investigated. The effect of important parameters of porous materials on beam behavior including pores compressibility, porosity distribution, thermal expansion of fluid within pores, the interaction of stresses between pores and material matrix due to temperature increase, effects of pore size, material thickness, and saturated pores with fluid and unsaturated conditions are investigated. Two methods, the deep energy method, and the exact solution have been used to reduce the problem hypotheses, increase accuracy, increase processing speed, and apply these in engineering structures. In both methods, they are analyzed nonlinear and complex equations of porous materials. To increase the accuracy of analysis and study of the effect of shear forces, Timoshenko and Reddy's beam theories have been used. Also, neural networks such as residual and fully connected networks are designed to have high accuracy and less processing time than other computational methods.}, subject = {Por{\"o}ser Stoff}, language = {en} } @phdthesis{Alalade, author = {Alalade, Muyiwa}, title = {An Enhanced Full Waveform Inversion Method for the Structural Analysis of Dams}, doi = {10.25643/bauhaus-universitaet.3956}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20190813-39566}, school = {Bauhaus-Universit{\"a}t Weimar}, abstract = {Since the Industrial Revolution in the 1700s, the high emission of gaseous wastes into the atmosphere from the usage of fossil fuels has caused a general increase in temperatures globally. To combat the environmental imbalance, there is an increase in the demand for renewable energy sources. Dams play a major role in the generation of "green" energy. However, these structures require frequent and strict monitoring to ensure safe and efficient operation. To tackle the challenges faced in the application of convention dam monitoring techniques, this work proposes the inverse analysis of numerical models to identify damaged regions in the dam. Using a dynamic coupled hydro-mechanical Extended Finite Element Method (XFEM) model and a global optimization strategy, damage (crack) in the dam is identified. By employing seismic waves to probe the dam structure, a more detailed information on the distribution of heterogeneous materials and damaged regions are obtained by the application of the Full Waveform Inversion (FWI) method. The FWI is based on a local optimization strategy and thus it is highly dependent on the starting model. A variety of data acquisition setups are investigated, and an optimal setup is proposed. The effect of different starting models and noise in the measured data on the damage identification is considered. Combining the non-dependence of a starting model of the global optimization strategy based dynamic coupled hydro-mechanical XFEM method and the detailed output of the local optimization strategy based FWI method, an enhanced Full Waveform Inversion is proposed for the structural analysis of dams.}, subject = {Talsperre}, language = {en} } @phdthesis{Budarapu, author = {Budarapu, Pattabhi Ramaiah}, title = {Adaptive multiscale methods for fracture}, doi = {10.25643/bauhaus-universitaet.2391}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20150507-23918}, school = {Bauhaus-Universit{\"a}t Weimar}, abstract = {One major research focus in the Material Science and Engineering Community in the past decade has been to obtain a more fundamental understanding on the phenomenon 'material failure'. Such an understanding is critical for engineers and scientists developing new materials with higher strength and toughness, developing robust designs against failure, or for those concerned with an accurate estimate of a component's design life. Defects like cracks and dislocations evolve at nano scales and influence the macroscopic properties such as strength, toughness and ductility of a material. In engineering applications, the global response of the system is often governed by the behaviour at the smaller length scales. Hence, the sub-scale behaviour must be computed accurately for good predictions of the full scale behaviour. Molecular Dynamics (MD) simulations promise to reveal the fundamental mechanics of material failure by modeling the atom to atom interactions. Since the atomistic dimensions are of the order of Angstroms ( A), approximately 85 billion atoms are required to model a 1 micro- m^3 volume of Copper. Therefore, pure atomistic models are prohibitively expensive with everyday engineering computations involving macroscopic cracks and shear bands, which are much larger than the atomistic length and time scales. To reduce the computational effort, multiscale methods are required, which are able to couple a continuum description of the structure with an atomistic description. In such paradigms, cracks and dislocations are explicitly modeled at the atomistic scale, whilst a self-consistent continuum model elsewhere. Many multiscale methods for fracture are developed for "fictitious" materials based on "simple" potentials such as the Lennard-Jones potential. Moreover, multiscale methods for evolving cracks are rare. Efficient methods to coarse grain the fine scale defects are missing. However, the existing multiscale methods for fracture do not adaptively adjust the fine scale domain as the crack propagates. Most methods, therefore only "enlarge" the fine scale domain and therefore drastically increase computational cost. Adaptive adjustment requires the fine scale domain to be refined and coarsened. One of the major difficulties in multiscale methods for fracture is to up-scale fracture related material information from the fine scale to the coarse scale, in particular for complex crack problems. Most of the existing approaches therefore were applied to examples with comparatively few macroscopic cracks. Key contributions The bridging scale method is enhanced using the phantom node method so that cracks can be modeled at the coarse scale. To ensure self-consistency in the bulk, a virtual atom cluster is devised providing the response of the intact material at the coarse scale. A molecular statics model is employed in the fine scale where crack propagation is modeled by naturally breaking the bonds. The fine scale and coarse scale models are coupled by enforcing the displacement boundary conditions on the ghost atoms. An energy criterion is used to detect the crack tip location. Adaptive refinement and coarsening schemes are developed and implemented during the crack propagation. The results were observed to be in excellent agreement with the pure atomistic simulations. The developed multiscale method is one of the first adaptive multiscale method for fracture. A robust and simple three dimensional coarse graining technique to convert a given atomistic region into an equivalent coarse region, in the context of multiscale fracture has been developed. The developed method is the first of its kind. The developed coarse graining technique can be applied to identify and upscale the defects like: cracks, dislocations and shear bands. The current method has been applied to estimate the equivalent coarse scale models of several complex fracture patterns arrived from the pure atomistic simulations. The upscaled fracture pattern agree well with the actual fracture pattern. The error in the potential energy of the pure atomistic and the coarse grained model was observed to be acceptable. A first novel meshless adaptive multiscale method for fracture has been developed. The phantom node method is replaced by a meshless differential reproducing kernel particle method. The differential reproducing kernel particle method is comparatively more expensive but allows for a more "natural" coupling between the two scales due to the meshless interpolation functions. The higher order continuity is also beneficial. The centro symmetry parameter is used to detect the crack tip location. The developed multiscale method is employed to study the complex crack propagation. Results based on the meshless adaptive multiscale method were observed to be in excellent agreement with the pure atomistic simulations. The developed multiscale methods are applied to study the fracture in practical materials like Graphene and Graphene on Silicon surface. The bond stretching and the bond reorientation were observed to be the net mechanisms of the crack growth in Graphene. The influence of time step on the crack propagation was studied using two different time steps. Pure atomistic simulations of fracture in Graphene on Silicon surface are presented. Details of the three dimensional multiscale method to study the fracture in Graphene on Silicon surface are discussed.}, subject = {Material}, language = {en} } @phdthesis{Eckardt2009, author = {Eckardt, Stefan}, title = {Adaptive heterogeneous multiscale models for the nonlinear simulation of concrete}, doi = {10.25643/bauhaus-universitaet.1416}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20100317-15023}, school = {Bauhaus-Universit{\"a}t Weimar}, year = {2009}, abstract = {The nonlinear behavior of concrete can be attributed to the propagation of microcracks within the heterogeneous internal material structure. In this thesis, a mesoscale model is developed which allows for the explicit simulation of these microcracks. Consequently, the actual physical phenomena causing the complex nonlinear macroscopic behavior of concrete can be represented using rather simple material formulations. On the mesoscale, the numerical model explicitly resolves the components of the internal material structure. For concrete, a three-phase model consisting of aggregates, mortar matrix and interfacial transition zone is proposed. Based on prescribed grading curves, an efficient algorithm for the generation of three-dimensional aggregate distributions using ellipsoids is presented. In the numerical model, tensile failure of the mortar matrix is described using a continuum damage approach. In order to reduce spurious mesh sensitivities, introduced by the softening behavior of the matrix material, nonlocal integral-type material formulations are applied. The propagation of cracks at the interface between aggregates and mortar matrix is represented in a discrete way using a cohesive crack approach. The iterative solution procedure is stabilized using a new path following constraint within the framework of load-displacement-constraint methods which allows for an efficient representation of snap-back phenomena. In several examples, the influence of the randomly generated heterogeneous material structure on the stochastic scatter of the results is analyzed. Furthermore, the ability of mesoscale models to represent size effects is investigated. Mesoscale simulations require the discretization of the internal material structure. Compared to simulations on the macroscale, the numerical effort and the memory demand increases dramatically. Due to the complexity of the numerical model, mesoscale simulations are, in general, limited to small specimens. In this thesis, an adaptive heterogeneous multiscale approach is presented which allows for the incorporation of mesoscale models within nonlinear simulations of concrete structures. In heterogeneous multiscale models, only critical regions, i.e. regions in which damage develops, are resolved on the mesoscale, whereas undamaged or sparsely damage regions are modeled on the macroscale. A crucial point in simulations with heterogeneous multiscale models is the coupling of sub-domains discretized on different length scales. The sub-domains differ not only in the size of the finite elements but also in the constitutive description. In this thesis, different methods for the coupling of non-matching discretizations - constraint equations, the mortar method and the arlequin method - are investigated and the application to heterogeneous multiscale models is presented. Another important point is the detection of critical regions. An adaptive solution procedure allowing the transfer of macroscale sub-domains to the mesoscale is proposed. In this context, several indicators which trigger the model adaptation are introduced. Finally, the application of the proposed adaptive heterogeneous multiscale approach in nonlinear simulations of concrete structures is presented.}, subject = {Beton}, language = {en} } @phdthesis{Luther2010, author = {Luther, Torsten}, title = {Adaptation of atomistic and continuum methods for multiscale simulation of quasi-brittle intergranular damage}, doi = {10.25643/bauhaus-universitaet.1436}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20101101-15245}, school = {Bauhaus-Universit{\"a}t Weimar}, year = {2010}, abstract = {The numerical simulation of damage using phenomenological models on the macroscale was state of the art for many decades. However, such models are not able to capture the complex nature of damage, which simultaneously proceeds on multiple length scales. Furthermore, these phenomenological models usually contain damage parameters, which are physically not interpretable. Consequently, a reasonable experimental determination of these parameters is often impossible. In the last twenty years, the ongoing advance in computational capacities provided new opportunities for more and more detailed studies of the microstructural damage behavior. Today, multiphase models with several million degrees of freedom enable for the numerical simulation of micro-damage phenomena in naturally heterogeneous materials. Therewith, the application of multiscale concepts for the numerical investigation of the complex nature of damage can be realized. The presented thesis contributes to a hierarchical multiscale strategy for the simulation of brittle intergranular damage in polycrystalline materials, for example aluminum. The numerical investigation of physical damage phenomena on an atomistic microscale and the integration of these physically based information into damage models on the continuum meso- and macroscale is intended. Therefore, numerical methods for the damage analysis on the micro- and mesoscale including the scale transfer are presented and the transition to the macroscale is discussed. The investigation of brittle intergranular damage on the microscale is realized by the application of the nonlocal Quasicontinuum method, which fully describes the material behavior by atomistic potential functions, but reduces the number of atomic degrees of freedom by introducing kinematic couplings. Since this promising method is applied only by a limited group of researchers for special problems, necessary improvements have been realized in an own parallelized implementation of the 3D nonlocal Quasicontinuum method. The aim of this implementation was to develop and combine robust and efficient algorithms for a general use of the Quasicontinuum method, and therewith to allow for the atomistic damage analysis in arbitrary grain boundary configurations. The implementation is applied in analyses of brittle intergranular damage in ideal and nonideal grain boundary models of FCC aluminum, considering arbitrary misorientations. From the microscale simulations traction separation laws are derived, which describe grain boundary decohesion on the mesoscale. Traction separation laws are part of cohesive zone models to simulate the brittle interface decohesion in heterogeneous polycrystal structures. 2D and 3D mesoscale models are presented, which are able to reproduce crack initiation and propagation along cohesive interfaces in polycrystals. An improved Voronoi algorithm is developed in 2D to generate polycrystal material structures based on arbitrary distribution functions of grain size. The new model is more flexible in representing realistic grain size distributions. Further improvements of the 2D model are realized by the implementation and application of an orthotropic material model with Hill plasticity criterion to grains. The 2D and 3D polycrystal models are applied to analyze crack initiation and propagation in statically loaded samples of aluminum on the mesoscale without the necessity of initial damage definition.}, subject = {Mechanik}, language = {en} } @phdthesis{Winkel, author = {Winkel, Benjamin}, title = {A three-dimensional model of skeletal muscle for physiological, pathological and experimental mechanical simulations}, doi = {10.25643/bauhaus-universitaet.4300}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20201211-43002}, school = {Bauhaus-Universit{\"a}t Weimar}, abstract = {In recent decades, a multitude of concepts and models were developed to understand, assess and predict muscular mechanics in the context of physiological and pathological events. Most of these models are highly specialized and designed to selectively address fields in, e.g., medicine, sports science, forensics, product design or CGI; their data are often not transferable to other ranges of application. A single universal model, which covers the details of biochemical and neural processes, as well as the development of internal and external force and motion patterns and appearance could not be practical with regard to the diversity of the questions to be investigated and the task to find answers efficiently. With reasonable limitations though, a generalized approach is feasible. The objective of the work at hand was to develop a model for muscle simulation which covers the phenomenological aspects, and thus is universally applicable in domains where up until now specialized models were utilized. This includes investigations on active and passive motion, structural interaction of muscles within the body and with external elements, for example in crash scenarios, but also research topics like the verification of in vivo experiments and parameter identification. For this purpose, elements for the simulation of incompressible deformations were studied, adapted and implemented into the finite element code SLang. Various anisotropic, visco-elastic muscle models were developed or enhanced. The applicability was demonstrated on the base of several examples, and a general base for the implementation of further material models was developed and elaborated.}, subject = {Biomechanik}, language = {en} } @phdthesis{Zhang, author = {Zhang, Yongzheng}, title = {A Nonlocal Operator Method for Quasi-static and Dynamic Fracture Modeling}, doi = {10.25643/bauhaus-universitaet.4732}, url = {http://nbn-resolving.de/urn:nbn:de:gbv:wim2-20221026-47321}, school = {Bauhaus-Universit{\"a}t Weimar}, abstract = {Material failure can be tackled by so-called nonlocal models, which introduce an intrinsic length scale into the formulation and, in the case of material failure, restore the well-posedness of the underlying boundary value problem or initial boundary value problem. Among nonlocal models, peridynamics (PD) has attracted a lot of attention as it allows the natural transition from continuum to discontinue and thus allows modeling of discrete cracks without the need to describe and track the crack topology, which has been a major obstacle in traditional discrete crack approaches. This is achieved by replacing the divergence of the Cauchy stress tensor through an integral over so-called bond forces, which account for the interaction of particles. A quasi-continuum approach is then used to calibrate the material parameters of the bond forces, i.e., equating the PD energy with the energy of a continuum. One major issue for the application of PD to general complex problems is that they are limited to fairly simple material behavior and pure mechanical problems based on explicit time integration. PD has been extended to other applications but losing simultaneously its simplicity and ease in modeling material failure. Furthermore, conventional PD suffers from instability and hourglass modes that require stabilization. It also requires the use of constant horizon sizes, which drastically reduces its computational efficiency. The latter issue was resolved by the so-called dual-horizon peridynamics (DH-PD) formulation and the introduction of the duality of horizons. Within the nonlocal operator method (NOM), the concept of nonlocality is further extended and can be considered a generalization of DH-PD. Combined with the energy functionals of various physical models, the nonlocal forms based on the dual-support concept can be derived. In addition, the variation of the energy functional allows implicit formulations of the nonlocal theory. While traditional integral equations are formulated in an integral domain, the dual-support approaches are based on dual integral domains. One prominent feature of NOM is its compatibility with variational and weighted residual methods. The NOM yields a direct numerical implementation based on the weighted residual method for many physical problems without the need for shape functions. Only the definition of the energy or boundary value problem is needed to drastically facilitate the implementation. The nonlocal operator plays an equivalent role to the derivatives of the shape functions in meshless methods and finite element methods (FEM). Based on the variational principle, the residual and the tangent stiffness matrix can be obtained with ease by a series of matrix multiplications. In addition, NOM can be used to derive many nonlocal models in strong form. The principal contributions of this dissertation are the implementation and application of NOM, and also the development of approaches for dealing with fractures within the NOM, mostly for dynamic fractures. The primary coverage and results of the dissertation are as follows: -The first/higher-order implicit NOM and explicit NOM, including a detailed description of the implementation, are presented. The NOM is based on so-called support, dual-support, nonlocal operators, and an operate energy functional ensuring stability. The nonlocal operator is a generalization of the conventional differential operators. Combining with the method of weighted residuals and variational principles, NOM establishes the residual and tangent stiffness matrix of operate energy functional through some simple matrix without the need of shape functions as in other classical computational methods such as FEM. NOM only requires the definition of the energy drastically simplifying its implementation. For the sake of conciseness, the implementation in this chapter is focused on linear elastic solids only, though the NOM can handle more complex nonlinear problems. An explicit nonlocal operator method for the dynamic analysis of elasticity solid problems is also presented. The explicit NOM avoids the calculation of the tangent stiffness matrix as in the implicit NOM model. The explicit scheme comprises the Verlet-velocity algorithm. The NOM can be very flexible and efficient for solving partial differential equations (PDEs). It's also quite easy for readers to use the NOM and extend it to solve other complicated physical phenomena described by one or a set of PDEs. Several numerical examples are presented to show the capabilities of this method. -A nonlocal operator method for the dynamic analysis of (thin) Kirchhoff plates is proposed. The nonlocal Hessian operator is derived from a second-order Taylor series expansion. NOM is higher-order continuous, which is exploited for thin plate analysis that requires \$C^1\$ continuity. The nonlocal dynamic governing formulation and operator energy functional for Kirchhoff plates are derived from a variational principle. The Verlet-velocity algorithm is used for time discretization. After confirming the accuracy of the nonlocal Hessian operator, several numerical examples are simulated by the nonlocal dynamic Kirchhoff plate formulation. -A nonlocal fracture modeling is developed and applied to the simulation of quasi-static and dynamic fractures using the NOM. The phase field's nonlocal weak and associated strong forms are derived from a variational principle. The NOM requires only the definition of energy. We present both a nonlocal implicit phase field model and a nonlocal explicit phase field model for fracture; the first approach is better suited for quasi-static fracture problems, while the key application of the latter one is dynamic fracture. To demonstrate the performance of the underlying approach, several benchmark examples for quasi-static and dynamic fracture are solved.}, subject = {Variationsprinzip}, language = {en} }