Refine
Document Type
- Article (57)
- Doctoral Thesis (26)
- Part of a Book (16)
- Master's Thesis (7)
- Book (4)
- Preprint (3)
- Conference Proceeding (2)
- Habilitation (2)
- Report (2)
- Bachelor Thesis (1)
Institute
- Institut für Strukturmechanik (ISM) (42)
- Junior-Professur Bildtheorie (17)
- Professur Bauphysik (8)
- Professur Sozialwissenschaftliche Stadtforschung (7)
- Junior-Professur Organisation und vernetzte Medien (6)
- Institut für Europäische Urbanistik (5)
- Professur Bauchemie und Polymere Werkstoffe (3)
- Professur Denkmalpflege und Baugeschichte (3)
- Professur Modellierung und Simulation - Konstruktion (3)
- Bauhaus-Institut für zukunftsweisende Infrastruktursysteme (b.is) (2)
Keywords
- OA-Publikationsfonds2020 (27)
- Maschinelles Lernen (17)
- Machine learning (12)
- Künstlerische Forschung (10)
- Erdbeben (7)
- Deep learning (5)
- Theater (5)
- big data (5)
- Medien (4)
- Raumklima (4)
Year of publication
- 2020 (123) (remove)
This study aims to evaluate a new approach in modeling gully erosion susceptibility (GES) based on a deep learning neural network (DLNN) model and an ensemble particle swarm optimization (PSO) algorithm with DLNN (PSO-DLNN), comparing these approaches with common artificial neural network (ANN) and support vector machine (SVM) models in Shirahan watershed, Iran. For this purpose, 13 independent variables affecting GES in the study area, namely, altitude, slope, aspect, plan curvature, profile curvature, drainage density, distance from a river, land use, soil, lithology, rainfall, stream power index (SPI), and topographic wetness index (TWI), were prepared. A total of 132 gully erosion locations were identified during field visits. To implement the proposed model, the dataset was divided into the two categories of training (70%) and testing (30%). The results indicate that the area under the curve (AUC) value from receiver operating characteristic (ROC) considering the testing datasets of PSO-DLNN is 0.89, which indicates superb accuracy. The rest of the models are associated with optimal accuracy and have similar results to the PSO-DLNN model; the AUC values from ROC of DLNN, SVM, and ANN for the testing datasets are 0.87, 0.85, and 0.84, respectively. The efficiency of the proposed model in terms of prediction of GES was increased. Therefore, it can be concluded that the DLNN model and its ensemble with the PSO algorithm can be used as a novel and practical method to predict gully erosion susceptibility, which can help planners and managers to manage and reduce the risk of this phenomenon.
When it comes to monitoring of huge structures, main issues are limited time, high costs and how to deal with the big amount of data. In order to reduce and manage them, respectively, methods from the field of optimal design of experiments are useful and supportive. Having optimal experimental designs at hand before conducting any measurements is leading to a highly informative measurement concept, where the sensor positions are optimized according to minimal errors in the structures’ models. For the reduction of computational time a combined approach using Fisher Information Matrix and mean-squared error in a two-step procedure is proposed under the consideration of different error types. The error descriptions contain random/aleatoric and systematic/epistemic portions. Applying this combined approach on a finite element model using artificial acceleration time measurement data with artificially added errors leads to the optimized sensor positions. These findings are compared to results from laboratory experiments on the modeled structure, which is a tower-like structure represented by a hollow pipe as the cantilever beam. Conclusively, the combined approach is leading to a sound experimental design that leads to a good estimate of the structure’s behavior and model parameters without the need of preliminary measurements for model updating.
The amount of adsorbed styrene acrylate copolymer (SA) particles on cementitious surfaces at the early stage of hydration was quantitatively determined using three different methodological approaches: the depletion method, the visible spectrophotometry (VIS) and the thermo-gravimetry coupled with mass spectrometry (TG–MS). Considering the advantages and disadvantages of each method, including the respectively required sample preparation, the results for four polymer-modified cement pastes, varying in polymer content and cement fineness, were evaluated.
To some extent, significant discrepancies in the adsorption degrees were observed. There is a tendency that significantly lower amounts of adsorbed polymers were identified using TG-MS compared to values determined with the depletion method. Spectrophotometrically generated values were lying in between these extremes. This tendency was found for three of the four cement pastes examined and is originated in sample preparation and methodical limitations.
The main influencing factor is the falsification of the polymer concentration in the liquid phase during centrifugation. Interactions in the interface between sediment and supernatant are the cause. The newly developed method, using TG–MS for the quantification of SA particles, proved to be suitable for dealing with these revealed issues. Here, instead of the fluid phase, the sediment is examined with regard to the polymer content, on which the influence of centrifugation is considerably lower.
Acoustic travel-time TOMography (ATOM) allows the measurement and reconstruction of air temperature distributions. Due to limiting factors, such as the challenge of travel-time estimation of the early reflections in the room impulse response, which heavily depends on the position of transducers inside the measurement area, ATOM is applied mainly outdoors. To apply ATOM in buildings, this paper presents a numerical solution to optimize the positions of transducers. This optimization avoids reflection overlaps, leading to distinguishable travel-times in the impulse response reflectogram. To increase the accuracy of the measured temperature within tomographic voxels, an additional function is employed to the proposed numerical method to minimize the number of sound-path-free voxels, ensuring the best sound-ray coverage of the room. Subsequently, an experimental set-up has been performed to verify the proposed numerical method. The results indicate the positive impact of the optimal positions of transducers on the distribution of ATOM-temperatures.
Discrete function theory in higher-dimensional setting has been in active development since many years. However, available results focus on studying discrete setting for such canonical domains as half-space, while the case of bounded domains generally remained unconsidered. Therefore, this paper presents the extension of the higher-dimensional function theory to the case of arbitrary bounded domains in Rn. On this way, discrete Stokes’ formula, discrete Borel–Pompeiu formula, as well as discrete Hardy spaces for general bounded domains are constructed. Finally, several discrete Hilbert problems are considered.
This article focuses on further developments of the background-oriented schlieren (BOS) technique to visualize convective indoor air flow, which is usually defined by very small density gradients. Since the light rays deflect when passing through fluids with different densities, BOS can detect the resulting refractive index gradients as integration along a line of sight. In this paper, the BOS technique is used to yield a two-dimensional visualization of small density gradients. The novelty of the described method is the implementation of a highly sensitive BOS setup to visualize the ascending thermal plume from a heated thermal manikin with temperature differences of minimum 1 K. To guarantee steady boundary conditions, the thermal manikin was seated in a climate laboratory. For the experimental investigations, a high-resolution DLSR camera was used capturing a large field of view with sufficient detail accuracy. Several parameters such as various backgrounds, focal lengths, room air temperatures, and distances between the object of investigation, camera, and structured background were tested to find the most suitable parameters to visualize convective indoor air flow. Besides these measurements, this paper presents the analyzing method using cross-correlation algorithms and finally the results of visualizing the convective indoor air flow with BOS. The highly sensitive BOS setup presented in this article complements the commonly used invasive methods that highly influence weak air flows.
The evolution of urbanism under dictatorship forms the core of the current research. This thesis is part of a research network at Bauhaus-Universität Weimar, which studies the 20th century's urbanism under different dictatorships. The network has provided a cross-cultural and cross-border environment and has enabled the author to communicate with other like-minded researchers. The 2015 published book of this group 'Urbanism and Dictatorship: A European Perspective' strengthens the foundation of this research's theoretical and methodological framework.
This thesis investigates urban policies and plans leading to the advancement of urbanization and the transformation of urban space in Iran during the second Pahlavi (1941-1979) when the country faced a milestone in its history: Nationalization of the Iranian oil industry. By reflecting the influence of economic and socio‐political determinants of the time on urbanism and the urbanization process, this work intends to critically trace the effect of dictatorship on evolved urbanism before and after the oil nationalization in 1951.
The research on the second Pahlavi's urbanism has been limitedly addressed and has only recently expanded. Most of the conducted studies date back to less than a decade ago and could not incorporate all the episodes of the second Pahlavi urbanism. These works have often investigated urbanism and architecture by focusing merely on the physical features and urban products in different years regardless of the importance of urbanism as a tool in the service of hegemony. In other words, the majority of the available literature does not intend to address the socio-economic and political roots of urban transformations and by questioning 'what has been built?' investigates the individual urban projects and plans designed by individual designers without interlinking these projects to the state's urban planning program and tracing the beneficiaries of those projects or questioning 'built for whom?'
Moreover, some chapters of this modern urbanism have rarely been investigated. For instance, scant research has looked into the works of foreign designers and consultants involved in the projects such as Peter Georg Ahrens or Constantinos A. Doxiadis. Similarly, the urbanism of the first decade of the second Pahlavi, including the government of Mossadegh, has mainly been overlooked.
Therefore, by critically analyzing the state's urban planning program and the process of urbanization in Iran during the second Pahlavi, this research aims to bridge the literature gap and to unravel the effect of the power structure on urban planning and products while seeking to find a pattern behind the regime's policies.
The main body of this work is concentrated on studying the history of urbanism in Iran, of which collecting data and descriptions played a crucial role. To prevent the limitations associated with singular methods, this research's methodology is based on methodological triangulation (Denzin, 2017). With the triangulation scheme, the data is gathered by combining different qualitative and quantitative methods such as the library, archival and media research, online resources, non-participatory observation, and photography. For the empirical part, the city of Tehran is selected as the case study. Moreover, individual non-structured interviews with the locals were conducted to gain more insights regarding urban projects.
The contribution explores the migratory situation on the Balkans and more specifically in the so-called Refugee District in Belgrade from a spatial perspective. By visualizing the areas of tensions in the Refugee District, the city of Belgrade, Serbia and Europe it aims to disentangle the political and socio-spatial levels that lead to the stuck situation of in-betweenness at the gates of the European Union.
das Theater der sorge ist konzipiert als 3-Phasen-modell:
in der 1. Phase geht es um die hersTellung einer laborsituation, in der präfigurative lebens-, denk- und arbeitsformen in diskursiver und spielerischer weise präsentisch erprobt und in konstituierende Prozesse übertragen werden können;
in der 2. Phase geht es um die gemeinschaftliche Konzeption, Vorbereitung und durchführung einer das labor abschließenden öffentlichen auFsTellung, die auf den kollektiven erfahrungen und ergebnissen der
labore basiert. dadurch können gesellschaftliche resonanzen erzeugt und die konstituierenden Prozesse vorläufig instituiert werden.
in der 3.Phase geht es dann darum, Vorbereitung: alle inhaltlichen erkenntnisse und ästhetischen Versuche in eine wiederholbare Theaterinszenierung / Performance zu überführen. die nötigen Voraussetzungen und bedingungen für diese von uns Vorstellungen genannten Formate sind für die hier verhandelte Fragestellung nicht von bedeutung – respektive würden sie den vorhandenen rahmen überschreiten – daher sparen wir diesen Komplex an dieser stelle aus.
Neuartige Sanitärsysteme zielen auf eine ressourcenorientierte Verwertung von Abwasser ab. Erreicht werden soll dies durch die separate Erfassung von Abwasserteilströmen. In den Fachöffentlichkeiten der Wasserwirtschaft und Raumplanung werden neuartige Sanitärsysteme als ein geeigneter Ansatz für die zukünftige
Sicherung der Abwasserentsorgung in ländlichen Räumen betrachtet. Die Praxistauglichkeit dieser Systeme wurde zwar in Forschungsprojekten nachgewiesen, bisher erschweren jedoch für Abwasserentsorger vielfältige Risiken die Einführung einer ressourcenorientierten Abwasserbewirtschaftung. Ausgehend von einer Untersuchung der Kontexte bei der Umsetzung eines neuartigen Sanitärsystems im ländlichen Raum Thüringens wird in diesem Beitrag der Frage nachgegangen, wie auf Landesebene mit dem abwasserwirtschaftlichen Instrumentarium die Einführung von ressourcenorientierten Systemansätzen unterstützt werden kann. Zentrale Elemente des Beitrags sind die Darstellung der wesentlichen Transformationsrisiken in Bezug auf die Einführung innovativer Lösungsansätze, eine Erläuterung der spezifischen abwasserwirtschaftlichen Instrumente sowie die Darlegung von Steuerungsansätzen,mit denen die Einführung von neuartigen Sanitärsystemen gefördert werden kann. Im Ergebnis wird die Realisierbarkeit von neuartigen Sanitärsystemen durch den strategischen Einsatz des Instrumentariums deutlich, gleichwohl die Wasserwirtschaft durch die Erweiterung der bisherigen Systemgrenzen auf die Kooperation mit anderen Bereichen der Daseinsvorsorge angewiesen ist.
A Machine Learning Framework for Assessing Seismic Hazard Safety of Reinforced Concrete Buildings
(2020)
Although averting a seismic disturbance and its physical, social, and economic disruption is practically impossible, using the advancements in computational science and numerical modeling shall equip humanity to predict its severity, understand the outcomes, and equip for post-disaster management. Many buildings exist amidst the developed metropolitan areas, which are senile and still in service. These buildings were also designed before establishing national seismic codes or without the introduction of construction regulations. In that case, risk reduction is significant for developing alternatives and designing suitable models to enhance the existing structure’s performance. Such models will be able to classify risks and casualties related to possible earthquakes through emergency preparation. Thus, it is crucial to recognize structures that are susceptible to earthquake vibrations and need to be prioritized for retrofitting. However, each building’s behavior under seismic actions cannot be studied through performing structural analysis, as it might be unrealistic because of the rigorous computations, long period, and substantial expenditure. Therefore, it calls for a simple, reliable, and accurate process known as Rapid Visual Screening (RVS), which serves as a primary screening platform, including an optimum number of seismic parameters and predetermined performance damage conditions for structures. In this study, the damage classification technique was studied, and the efficacy of the Machine Learning (ML) method in damage prediction via a Support Vector Machine (SVM) model was explored. The ML model is trained and tested separately on damage data from four different earthquakes, namely Ecuador, Haiti, Nepal, and South Korea. Each dataset consists of varying numbers of input data and eight performance modifiers. Based on the study and the results, the ML model using SVM classifies the given input data into the belonging classes and accomplishes the performance on hazard safety evaluation of buildings.
Personalisierte Lüftung (PL) kann die thermische Behaglichkeit sowie die Qualität der eingeatmeten Atemluft verbessern, in dem jedem Arbeitsplatz Frischluft separat zugeführt wird. In diesem Beitrag wird die Wirkung der PL auf die thermische Behaglichkeit der Nutzer unter sommerlichen Randbedingungen untersucht. Hierfür wurden zwei Ansätze zur Bewertung des Kühlungseffekts der PL untersucht: basierend auf (1) der äquivalenten Temperatur und (2) dem thermischen Empfinden. Grundlage der Auswertung sind in einer Klimakammer gemessene sowie numerisch simulierte Daten. Vor der Durchführung der Simulationen wurde das numerische Modell zunächst anhand der gemessenen Daten validiert. Die Ergebnisse zeigen, dass der Ansatz basierend auf dem thermischen Empfinden zur Evaluierung des Kühlungseffekts der PL sinnvoller sein kann, da bei diesem die komplexen physiologischen Faktoren besser berücksichtigt werden.
Radikale Planung
(2020)
Smart Cities and Mobility Stations: Lessons learned from the Smarter Together in Vienna and Munich
(2020)
With an increasing urban population and urban problems arising from this unplanned growth, several projects aimed at promoting sustainable urban development have emerged. Smart mobility strategies, such as shared mobility and mobility stations, represent some of the solutions to promote changes in travel behavior. Despite its beneficial impacts, however, the implementation of such infrastructure is criticized for not contributing to current urban issues, as well as often disregarding knowledge about urban space and its functioning.
In this context, the Smarter Together, a joint research and innovation project funded through the European Union program H2020, was implemented. The project selected three lighthouse cities to test and upscale innovative solutions: Vienna, Munich, and Lyon.
This master thesis presents the main characteristics of the mobility stations systems implemented in Vienna and Munich in the scope of the project Smarter Together. Its main goal is to share what can be learned from their experiences while approaching critically the concept of smart cities. This master thesis identifies important aspects to take into account when planning, implementing, and operating mobility stations, and provides an understanding of smart cities and smart mobility that goes beyond the adoption of technology. Several methods were combined for the development of this master thesis, such as quantitative secondary data, observational studies, application of survey forms, explorative expert interviews, and literature review.
This work has demonstrated that the Smarter Together has a cutting-edge scope and contributed greatly to research and innovation, by creating living laboratories to test the application of technology in the urban environment. However, from the perspective of the mobility stations assessment, many caveats were made. In short, many lessons could be learned and are presented throughout this work aiming at contributing to the improvement of the mobility stations implemented in the project areas in Munich and Vienna, as well as for inspiring other cities in Europe and worldwide.
Unmanned aircraft systems (UAS) show large potential for the construction industry. Their use in condition assessment has increased significantly, due to technological and computational progress. UAS play a crucial role in developing a digital maintenance strategy for infrastructure, saving cost and effort, while increasing safety and reliability. Part of that strategy are automated visual UAS inspections of the building’s condition. The resulting images can automatically be analyzed to identify and localize damages to the structure that have to be monitored. Further interest in parts of a structure can arise from events like accidents or collisions. Areas of low interest exist, where low resolution monitoring is sufficient.
From different requirements for resolution, different levels of detail can be derived. They require special image acquisition parameters that differ mainly in the distance between camera and structure. Areas with a higher level of detail require a smaller distance to the object, producing more images. This work proposes a multi-scale flight path planning procedure, enabling higher resolution requirements for areas of special interest, while reducing the number of required images to a minimum. Careful selection of the camera positions maintains the complete coverage of the structure, while achieving the required resolution in all areas. The result is an efficient UAS inspection, reducing effort for the maintenance of infrastructure.
The latest earthquakes have proven that several existing buildings, particularly in developing countries, are not secured from damages of earthquake. A variety of statistical and machine-learning approaches have been proposed to identify vulnerable buildings for the prioritization of retrofitting. The present work aims to investigate earthquake susceptibility through the combination of six building performance variables that can be used to obtain an optimal prediction of the damage state of reinforced concrete buildings using artificial neural network (ANN). In this regard, a multi-layer perceptron network is trained and optimized using a database of 484 damaged buildings from the Düzce earthquake in Turkey. The results demonstrate the feasibility and effectiveness of the selected ANN approach to classify concrete structural damage that can be used as a preliminary assessment technique to identify vulnerable buildings in disaster risk-management programs.
Die im Jahr 2020 in Deutschland praktizierte Siedlungs- und Wohnungspolitik erhält in Anbetracht ihrer Auswirkungen auf die soziale und ökologische Lage einen bitteren Beigeschmack. Arm und Reich triften weiter auseinander und einer zielgerichteten ökologischen Transformation der Art und Weise, wie Stadtentwicklung und Wohnungspolitik gestaltet werden,stehen noch immer historisch und systemisch bedingte Pfadabhängigkeiten im Weg. Diese werden nur durch eine integrierte Betrachtung sozialer und ökonomischer Aspekte sichtbar und deuten auf eine der ursprünglichen Fragen linker Gesellschaftsforschung hin: Die Auseinandersetzung mit dem Verhältnis von Eigentum und Gerechtigkeit.
Im Ergebnis stehen drei wesentliche Befunde: Der Diskurs zum Schutz des Klimas und der Biodiversität berührt direkt die Parameter Dichte, Nutzungsmischung und Flächeninanspruchnahme; zweitens steigt letztere relativ mit erhöhtem, individuell verfügbaren Kapital und insbesondere im selbstgenutztem Eigentum gegenüber Mietwohnungen; und drittens wächst der Eigentumsanteil mit fortschreitender Finanzialisierung des Wohnungsmarktes, sodass das Risiko sozialer und ökologischer Krisen sich verschärft.
Marine Makroalgen besitzen vielversprechende Eigenschaften und Inhaltsstoffe für die Verwendung als Energieträger, Nahrungsmittel oder als Ausgangsstoff für Pharmazeutika. Dass die Quantität und Qualität der in natürlicher Umgebung wachsenden Makroalgen schwankt, reduziert jedoch deren Verwertbarkeit und erschwert die Erschließung hochpreisiger Marktsegmente. Zudem ist eine Ausweitung der Zucht in marinen und küstennahen Aquakulturen in Europa gegenwärtig wenig aussichtsreich, da vielversprechende Areale bereits zum Fischfang oder als Erholungs- bzw. Naturschutzgebiete ausgewiesen sind. Im Rahmen dieser Arbeit wird demzufolge ein geschlossenes Photobioreaktorsystem zur Makroalgenkultivierung entwickelt, welches eine umfassende Kontrolle der abiotischen Kultivierungsparameter und eine effektive Aufbereitung des Kulturmediums vorsieht, um eine standortunabhängige Algenproduktion zu ermöglichen. Zur Bilanzierung des Gesamtkonzeptes einer Kultivierung und Verwertung (stofflich oder energetisch) werden die spezifischen Wachstumsraten und Methanbildungspotentiale der Algenarten Ulva intestinalis, Fucus vesiculosus und Palmaria palmata in praktischen Versuchen ermittelt.
Im Ergebnis wird für den gegenwärtigen Entwicklungsstand der Kultivierungsanlage eine positive Bilanz für die stoffliche Verwertung der Algenart Ulva intestinalis und eine negative Bilanz für die energetische Verwertung aller untersuchten Algenarten erzielt. Wird ein Optimalszenario betrachtet, indem die Besatzdichten und Wachstumsraten der Algen in der Zucht erhöht werden, bleibt die Energiebilanz negativ. Allerdings summieren sich die finanzielle Einnahmen durch einen Verkauf der Algen als Produkt auf jährlich 460.869€ für Ulva intestinalis, 4.010€ für Fucus vesiculosus und 16.913€ für Palmaria palmata. Im Ergebnis ist insbesondere eine stoffliche Verwertung der gezüchteten Grünalge Ulva intestinalis anzustreben und die Produktivität der Zuchtanlage im Sinne des Optimalszenarios zu steigern.
This study permits a reliability analysis to solve the mechanical behaviour issues existing in the current structural design of fabric structures. Purely predictive material models are highly desirable to facilitate an optimized design scheme and to significantly reduce time and cost at the design stage, such as experimental characterization.
The present study examined the role of three major tasks; a) single-objective optimization, b) sensitivity analyses and c) multi-objective optimization on proposed weave structures for woven fabric composites. For single-objective optimization task, the first goal is to optimize the elastic properties of proposed complex weave structure under unit cells basis based on periodic boundary conditions.
We predict the geometric characteristics towards skewness of woven fabric composites via Evolutionary Algorithm (EA) and a parametric study. We also demonstrate the effect of complex weave structures on the fray tendency in woven fabric composites via tightness evaluation. We utilize a procedure which does not require a numerical averaging process for evaluating the elastic properties of woven fabric composites. The fray tendency and skewness of woven fabrics depends upon the behaviour of the floats which is related to the factor of weave. Results of this study may suggest a broader view for further research into the effects of complex weave structures or may provide an alternative to the fray and skewness problems of current weave structure in woven fabric composites.
A comprehensive study is developed on the complex weave structure model which adopts the dry woven fabric of the most potential pattern in singleobjective optimization incorporating the uncertainties parameters of woven fabric composites. The comprehensive study covers the regression-based and variance-based sensitivity analyses. The second task goal is to introduce the fabric uncertainties parameters and elaborate how they can be incorporated into finite element models on macroscopic material parameters such as elastic modulus and shear modulus of dry woven fabric subjected to uni-axial and biaxial deformations. Significant correlations in the study, would indicate the need for a thorough investigation of woven fabric composites under uncertainties parameters. The study describes here could serve as an alternative to identify effective material properties without prolonged time consumption and expensive experimental tests.
The last part focuses on a hierarchical stochastic multi-scale optimization approach (fine-scale and coarse-scale optimizations) under geometrical uncertainties parameters for hybrid composites considering complex weave structure. The fine-scale optimization is to determine the best lamina pattern that maximizes its macroscopic elastic properties, conducted by EA under the following uncertain mesoscopic parameters: yarn spacing, yarn height, yarn width and misalignment of yarn angle. The coarse-scale optimization has been carried out to optimize the stacking sequences of symmetric hybrid laminated composite plate with uncertain mesoscopic parameters by employing the Ant Colony Algorithm (ACO). The objective functions of the coarse-scale optimization are to minimize the cost (C) and weight (W) of the hybrid laminated composite plate considering the fundamental frequency and the buckling load factor as the design constraints.
Based on the uncertainty criteria of the design parameters, the appropriate variation required for the structural design standards can be evaluated using the reliability tool, and then an optimized design decision in consideration of cost can be subsequently determined.
Space is a social product and a social producer. The main aim of this thesis is to reveal ‘the process of totalitarian city making in Pyongyang’, especially in the light of the interaction between the power and urban space.
The totalitarian city of Pyongyang was born out of modernization in the process of masses formation. During the growth of colonial capitalism and Christian liberal ideas, Pyongyang was modernized and displayed the characteristics of a modern city with industrialization and urbanization. During the introduction of Japanese colonial capitalism, peasants, women, and slaves became the first masses and urban poor, and they later transformed into the mob; their violence was finally demonstrated during the Anti-Chinese Riot.
After the 1945 independence, Kim’s regime formed the one-party state with a cry for revolution. They produced an atmosphere of imminent war to instill fear and hatred into the psyche of Pyongyang citizens. The regime eliminated all political opponents in 1967 and finally declared the totalitarian ideology in 1974. During this process, Pyongyang demonstrated two main characteristics of a totalitarian city: the space of terror and of ideology. The space of terror produces the fear of death and the space of ideology controls the thought and life of citizens.
After entry to the market, to keep Kim’s controlling power, the regime used the strategy of location exchange. The camp, market, and Foreign Currency Shop were effective tools to prepare for executives’ gifts. However, the market also produces the desire not only for consumption but also for freedom and truth; it is tearing down the foundation of the totalitarian city of Pyongyang.
This research focuses primarily on the interaction between political power and urban space. In the process of making a totalitarian city, the power produced urban space and it influenced the psyche of Pyongyang citizens. Even though this spatial transition has created the totalitarian city and helped maintain political power, it also led and produced intended or unintended social variation in Pyongyang society.
In this paper, an artificial neural network is implemented for the sake of predicting the thermal conductivity ratio of TiO2-Al2O3/water nanofluid. TiO2-Al2O3/water in the role of an innovative type of nanofluid was synthesized by the sol–gel method. The results indicated that 1.5 vol.% of nanofluids enhanced the thermal conductivity by up to 25%. It was shown that the heat transfer coefficient was linearly augmented with increasing nanoparticle concentration, but its variation with temperature was nonlinear. It should be noted that the increase in concentration may cause the particles to agglomerate, and then the thermal conductivity is reduced. The increase in temperature also increases the thermal conductivity, due to an increase in the Brownian motion and collision of particles. In this research, for the sake of predicting the thermal conductivity of TiO2-Al2O3/water nanofluid based on volumetric concentration and temperature functions, an artificial neural network is implemented. In this way, for predicting thermal conductivity, SOM (self-organizing map) and BP-LM (Back Propagation-Levenberq-Marquardt) algorithms were used. Based on the results obtained, these algorithms can be considered as an exceptional tool for predicting thermal conductivity. Additionally, the correlation coefficient values were equal to 0.938 and 0.98 when implementing the SOM and BP-LM algorithms, respectively, which is highly acceptable. View Full-Text
The performance of ductless personalized ventilation (DPV) was compared to the performance of a typical desk fan since they are both stand-alone systems that allow the users to personalize their indoor environment. The two systems were evaluated using a validated computational fluid dynamics (CFD) model of an office room occupied by two users. To investigate the impact of DPV and the fan on the inhaled air quality, two types of contamination sources were modelled in the domain: an active source and a passive source. Additionally, the influence of the compared systems on thermal comfort was assessed using the coupling of CFD with the comfort model developed by the University of California, Berkeley (UCB model). Results indicated that DPV performed generally better than the desk fan. It provided better thermal comfort and showed a superior performance in removing the exhaled contaminants. However, the desk fan performed better in removing the contaminants emitted from a passive source near the floor level. This indicates that the performance of DPV and desk fans depends highly on the location of the contamination source. Moreover, the simulations showed that both systems increased the spread of exhaled contamination when used by the source occupant.
The K-nearest neighbors (KNN) machine learning algorithm is a well-known non-parametric classification method. However, like other traditional data mining methods, applying it on big data comes with computational challenges. Indeed, KNN determines the class of a new sample based on the class of its nearest neighbors; however, identifying the neighbors in a large amount of data imposes a large computational cost so that it is no longer applicable by a single computing machine. One of the proposed techniques to make classification methods applicable on large datasets is pruning. LC-KNN is an improved KNN method which first clusters the data into some smaller partitions using the K-means clustering method; and then applies the KNN for each new sample on the partition which its center is the nearest one. However, because the clusters have different shapes and densities, selection of the appropriate cluster is a challenge. In this paper, an approach has been proposed to improve the pruning phase of the LC-KNN method by taking into account these factors. The proposed approach helps to choose a more appropriate cluster of data for looking for the neighbors, thus, increasing the classification accuracy. The performance of the proposed approach is evaluated on different real datasets. The experimental results show the effectiveness of the proposed approach and its higher classification accuracy and lower time cost in comparison to other recent relevant methods.
Rechargeable lithium ion batteries (LIBs) play a very significant role in power supply and storage. In recent decades, LIBs have caught tremendous attention in mobile communication, portable electronics, and electric vehicles. Furthermore, global warming has become a worldwide issue due to the ongoing production of greenhouse gases. It motivates solutions such as renewable sources of energy. Solar and wind energies are the most important ones in renewable energy sources. By technology progress, they will definitely require batteries to store the produced power to make a balance between power generation and consumption. Nowadays,rechargeable batteries such as LIBs are considered as one of the best solutions. They provide high specific energy and high rate performance while their rate of self-discharge is low.
Performance of LIBs can be improved through the modification of battery characteristics. The size of solid particles in electrodes can impact the specific energy and the cyclability of batteries. It can improve the amount of lithium content in the electrode which is a vital parameter in capacity and capability of a battery. There exist diferent sources of heat generation in LIBs such as heat produced during electrochemical reactions, internal resistance in battery. The size of electrode's electroactive particles can directly affect the produced heat in battery. It will be shown that the smaller size of solid particle enhance the thermal characteristics of LIBs.
Thermal issues such as overheating, temperature maldistribution in the battery, and thermal runaway have confined applications of LIBs. Such thermal challenges reduce the Life cycle of LIBs. As well, they may lead to dangerous conditions such as fire or even explosion in batteries. However, recent advances in fabrication of advanced materials such as graphene and carbon nanotubes with extraordinary thermal conductivity and electrical properties propose new opportunities to enhance their performance. Since experimental works are expensive, our objective is to use computational methods to investigate the thermal issues in LIBS. Dissipation of the heat produced in the battery can improve the cyclability and specific capacity of LIBs. In real applications, packs of LIB consist several battery cells that are used as the power source. Therefore, it is worth to investigate thermal characteristic of battery packs under their cycles of charging/discharging operations at different applied current rates. To remove the produced heat in batteries, they can be surrounded by materials with high thermal conductivity. Parafin wax absorbs high energy since it has a high latent heat. Absorption high amounts of energy occurs at constant temperature without phase change. As well, thermal conductivity of parafin can be magnified with nano-materials such as graphene, CNT, and fullerene to form a nano-composite medium. Improving the thermal conductivity of LIBs increase the heat dissipation from batteries which is a vital issue in systems of battery thermal management. The application of two-dimensional (2D) materials has been on the rise since exfoliation the graphene from bulk graphite. 2D materials are single-layered in an order of nanosizes which show superior thermal, mechanical, and optoelectronic properties. They are potential candidates for energy storage and supply, particularly in lithium ion batteries as electrode material. The high thermal conductivity of graphene and graphene-like materials can play a significant role in thermal management of batteries. However, defects always exist in nano-materials since there is no ideal fabrication process. One of the most important defects in materials are nano-crack which can dramatically weaken the mechanical properties of the materials. Newly synthesized crystalline carbon nitride with the stoichiometry of C3N have attracted many attentions due to its extraordinary mechanical and thermal properties. The other nano-material is phagraphene which shows anisotropic mechanical characteristics which is ideal in production of nanocomposite.
It shows ductile fracture behavior when subjected under uniaxial loadings. It is worth to investigate their thermo-mechanical properties in its pristine and defective states. We hope that the findings of our work not only be useful for both experimental and theoretical researches but also help to design advanced electrodes for LIBs.
Vor dem Hintergrund einer stetig wachsenden Nachfrage an Beton wie auch ambitionierter Reduktionsziele beim in der Zementproduktion anfallenden CO2 gelten calcinierte Tone als derzeit aussichtsreichste technische Neuerung im Bereich nachhaltiger Bindemittelkonzepte. Unter Ausnutzung ihrer Puzzolanität soll ein erheblicher Teil der Klinkerkomponente im Zement ersetzt werden, wobei der zu ihrer Aktivierung notwendige Energiebedarf vergleichsweise niedrig ist. Wesentliche Vorteile der Tone sind ihre weltweit nahezu unbegrenzte Verfügbarkeit sowie der äußerst geringe rohstoffbedingte CO2-Ausstoß während der Calcinierung. Schwierigkeiten auf dem Weg der Umsetzung bestehen allerdings in der Vielseitigkeit des Systems, welches durch eine hohe Varietät der Rohtone und des daraus folgenden thermischen Verhaltens gekennzeichnet ist. Entsprechend schwierig ist die Übertragbarkeit von Erfahrungen mit bereits etablierten calcinierten Tonen wie dem Metakaolin, der sich durch hohe Reinheit, einen aufwendigen Aufbereitungsprozess und eine entsprechend hohe Reaktivität auszeichnet. Ziel der Arbeit ist es daher, den bereits erlangten Kenntnisstand auf andere, wirtschaftlich relevante Tone zu erweitern und deren Eignung für die Anwendung im Beton herauszuarbeiten.
In einem mehrstufigen Arbeitsprogramm wurde untersucht, inwieweit großtechnisch nutzbare Tone aktivierbar sind und welche Eigenschaften sich daraus für Zement und Beton ergeben. Die dabei festgestellte Reihenfolge Kaolinit > Montmorillonit > Illit beschreibt sowohl die Reaktivität der Brennprodukte als auch umgekehrt die Höhe der optimalen Calciniertemperatur. Auch wandelt sich der Charakter der entstandenen Metaphasen in dieser Abfolge von röntgenamorph und hochreaktiv zu glasig und reaktionsträge. Trotz dieser Einordnung konnte selbst mit dem Illit eine mit Steinkohlenflugasche vergleichbare Puzzolanität festgestellt werden. Dies bestätigte sich anschließend in Parameterversuchen, bei denen die Einflüsse von Rohstoffqualität, Calcinierung, Aufbereitung und Zement hinsichtlich der Reaktivitätsausbeute bewertet wurden. Die Bandbreite der erzielbaren Qualitäten ist dabei immens und gipfelt nicht zuletzt in stark unterschiedlichen Wirkungen auf die Festbetoneigenschaften. Hier machte sich vor allem die für Puzzolane typische Porenverfeinerung bemerkbar, sodass viele von Transportvorgängen abhängige Schadmechanismen unterdrückt wurden. Andere Schadex-positionen wie der Frostangriff ließen sich durch Zusatzmaßnahmen wie dem Eintrag von Luftporen beherrschen. Zu bemängeln sind vor allem die schlechte Verarbeitbarkeit kaolinitischer Metatone wie auch die für Puzzolane stark ausgeprägte Carbonatisierungsneigung.
Wesentliches Ergebnis der Arbeit ist, dass auch Tone, die bisher als geringwertig bezüglich des Aktivierungspotentials galten, nutzbare puzzolanische Eigenschaften entwickeln können. So kann selbst ein stark verunreinigter Illit-Ton die Qualität von Flugasche erreichen. Mit stei-gendem Tonmineralgehalt sowie bei Präsens thermisch instabilerer Tonminerale wie Mont-morillonit und Kaolinit erweitert sich das Spektrum nutzbarer Puzzolanitäten bis hin zur hochreaktiven Metakaolin-Qualität. Damit lassen sich gute bis sehr gute Betoneigenschaften erzielen, sodass die Leistungsfähigkeit etablierter Kompositmaterialien erreicht wird. Somit sind die Voraussetzungen für eine umfangreiche Nutzung der erheblichen Tonmengen im Zement und Beton gegeben. Entsprechend können Tone einen effektiven Beitrag zu einer gesteigerten Nachhaltigkeit in der Baustoffproduktion weltweit leisten.
Synergistic Framework for Analysis and Model Assessment in Bridge Aerodynamics and Aeroelasticity
(2020)
Wind-induced vibrations often represent a major design criterion for long-span bridges. This work deals with the assessment and development of models for aerodynamic and aeroelastic analyses of long-span bridges.
Computational Fluid Dynamics (CFD) and semi-analytical aerodynamic models are employed to compute the bridge response due to both turbulent and laminar free-stream. For the assessment of these models, a comparative methodology is developed that consists of two steps, a qualitative and a quantitative one. The first, qualitative, step involves an extension
of an existing approach based on Category Theory and its application to the field of bridge aerodynamics. Initially, the approach is extended to consider model comparability and completeness. Then, the complexity of the CFD and twelve semi-analytical models are evaluated based on their mathematical constructions, yielding a diagrammatic representation of model quality.
In the second, quantitative, step of the comparative methodology, the discrepancy of a system response quantity for time-dependent aerodynamic models is quantified using comparison metrics for time-histories. Nine metrics are established on a uniform basis to quantify the discrepancies in local and global signal features that are of interest in bridge aerodynamics. These signal features involve quantities such as phase, time-varying frequency and magnitude content, probability density, non-stationarity, and nonlinearity.
The two-dimensional (2D) Vortex Particle Method is used for the discretization of the Navier-Stokes equations including a Pseudo-three dimensional (Pseudo-3D) extension within an existing CFD solver. The Pseudo-3D Vortex Method considers the 3D structural behavior for aeroelastic analyses by positioning 2D fluid strips along a line-like structure. A novel turbulent Pseudo-3D Vortex Method is developed by combining the laminar Pseudo-3D VPM and a previously developed 2D method for the generation of free-stream turbulence. Using analytical derivations, it is shown that the fluid velocity correlation is maintained between the CFD strips.
Furthermore, a new method is presented for the determination of the complex aerodynamic admittance under deterministic sinusoidal gusts using the Vortex Particle Method. The sinusoidal gusts are simulated by modeling the wakes of flapping airfoils in the CFD domain with inflow vortex particles. Positioning a section downstream yields sinusoidal forces that are used for determining all six components of the complex aerodynamic admittance. A closed-form analytical relation is derived, based on an existing analytical model. With this relation, the inflow particles’ strength can be related with the target gust amplitudes a priori.
The developed methodologies are combined in a synergistic framework, which is applied to both fundamental examples and practical case studies. Where possible, the results are verified and validated. The outcome of this work is intended to shed some light on the complex wind–bridge interaction and suggest appropriate modeling strategies for an enhanced design.
Diese Dissertation beschäftigt sich mit Kunstwerken, die das alltägliche Ding in den Blick nehmen. Nährboden dieser Kunstform sind die soziokulturellen Entwicklungen des 20. Jahrhunderts, mit denen wesentliche Veränderungen hinsichtlich des Verhältnisses von Mensch und Ding einhergingen.
Daraus resultierte eine allgemeine künstlerische Zuwendung zu den Dingen und eine einzigartige Kulmination aus verschiedenartigen Auseinandersetzungen mit ihnen als kunstfähige Gegenstände, über die sich die neue Dingwelt erschlossen wurde und deren Kunstwerke einen Spiegel dieser Entwicklungen darstellen.
Die Dissertation stellt ebenfalls die Dinge selbst in den Fokus. Vier Aspekte von Dingen (Materialität, Funktionalität, Repräsentationalität und Relationalität) werden gesondert ins Auge gefasst und in den theoretischen Diskurs des 20. Jahrhunderts eingeordnet, um sie als Teil der gelebten Realität besser zu verstehen, von der sich der ästhetische Blick nicht trennen lässt. Anhand der künstlerischen Positionen von Robert Rauschenberg, Christo und Jeanne-Claude, Daniel Spoerri und Arman sowie Claes Oldenburg werden die verschiedenen Aspekte der Dinge näher betrachtet und analysiert, wie diese speziell in den Kunstwerken thematisiert werden und welche Relevanz sie für deren Rezeptionserfahrung haben.
Die Korrelation dieser beiden Ebenen - die Dinge als konstitutiver Bestandteil im sozialen Raum und die Dinge als Elemente in Kunstwerken -, die im Fokus der vorliegenden Untersuchung steht, ermöglicht es, die künstlerische Zuwendung zu den Dingen in den 1960er-Jahren neu einzuordnen. Darüber hinaus wird dadurch ein differenziertes Bild von der Kunst dieser Zeit sowie den Dingen in der Kunst im Allgemeinen gezeichnet.
Städte ohne Wachstum - eine bislang kaum vorstellbare Vision. Doch Klimawandel, Ressourcenverschwendung, wachsende soziale Ungleichheiten und viele andere Zukunftsgefahren stellen das bisherige Allheilmittel Wachstum grundsätzlich infrage. Wie wollen wir heute und morgen zusammenleben? Wie gestalten wir ein gutes Leben für alle in der Stadt? Während in einzelnen Nischen diese Fragen bereits ansatzweise beantwortet werden, fehlt es noch immer an umfassenden Entwürfen und Transformationsansätzen, die eine fundamental andere, solidarische Stadt konturieren. Diesen Versuch wagt das Projekt Postwachstumsstadt.
In diesem Buch werden konzeptionelle und pragmatische Aspekte aus verschiedenen Bereichen der Stadtpolitik zusammengebracht, die neue Pfade aufzeigen und verknüpfen. Die Beiträge diskutieren städtische Wachstumskrisen, transformative Planung und Konflikte um Gestaltungsmacht. Nicht zuletzt wird dabei auch die Frage nach der Rolle von Stadtutopien neu gestellt. Dadurch soll eine längst fällige Debatte darüber angestoßen werden, wie sich notwendige städtische Wenden durch eine sozialökologische Neuorientierung vor Ort verwirklichen lassen.
In this study, machine learning methods of artificial neural networks (ANNs), least squares support vector machines (LSSVM), and neuro-fuzzy are used for advancing prediction models for thermal performance of a photovoltaic-thermal solar collector (PV/T). In the proposed models, the inlet temperature, flow rate, heat, solar radiation, and the sun heat have been considered as the input variables. Data set has been extracted through experimental measurements from a novel solar collector system. Different analyses are performed to examine the credibility of the introduced models and evaluate their performances. The proposed LSSVM model outperformed the ANFIS and ANNs models. LSSVM model is reported suitable when the laboratory measurements are costly and time-consuming, or achieving such values requires sophisticated interpretations.
A novel combination of the ant colony optimization algorithm (ACO)and computational fluid dynamics (CFD) data is proposed for modeling the multiphase chemical reactors. The proposed intelligent model presents a probabilistic computational strategy for predicting various levels of three-dimensional bubble column reactor (BCR) flow. The results prove an enhanced communication between ant colony prediction and CFD data in different sections of the BCR.
Recently, the demand for residence and usage of urban infrastructure has been increased, thereby resulting in the elevation of risk levels of human lives over natural calamities. The occupancy demand has rapidly increased the construction rate, whereas the inadequate design of structures prone to more vulnerability. Buildings constructed before the development of seismic codes have an additional susceptibility to earthquake vibrations. The structural collapse causes an economic loss as well as setbacks for human lives. An application of different theoretical methods to analyze the structural behavior is expensive and time-consuming. Therefore, introducing a rapid vulnerability assessment method to check structural performances is necessary for future developments. The process, as mentioned earlier, is known as Rapid Visual Screening (RVS). This technique has been generated to identify, inventory, and screen structures that are potentially hazardous. Sometimes, poor construction quality does not provide some of the required parameters; in this case, the RVS process turns into a tedious scenario. Hence, to tackle such a situation, multiple-criteria decision-making (MCDM) methods for the seismic vulnerability assessment opens a new gateway. The different parameters required by RVS can be taken in MCDM. MCDM evaluates multiple conflicting criteria in decision making in several fields. This paper has aimed to bridge the gap between RVS and MCDM. Furthermore, to define the correlation between these techniques, implementation of the methodologies from Indian, Turkish, and Federal Emergency Management Agency (FEMA) codes has been done. The effects of seismic vulnerability of structures have been observed and compared.
Experimente lernen, Techniken tauschen. Ein spekulatives Handbuch
Das spekulative Handbuch bietet vielfältige Techniken für ein radikales Lernen und Vermitteln. Es umfasst konkrete Anleitungen, Erfahrungen und theoretische Überlegungen. Die Texte beteiligen sich an der Konzeption einer Vermittlung, die das gemeinsame Experimentieren (wieder) einführt.
Im Seminarraum, in Workshops, auf Festivals, in Fluren, Parks und der Stadt finden Lernen und Verlernen statt. Texte und Anleitungen u. a. zu: Filmessays, Collagen, Banküberfällen, der Universität der Toten, wildem Schreiben, konzeptuellem speed Dating, neurodiversem Lernen, Format-Denken, dem Theater der Sorge, dem Schreiblabor, dem Körperstreik.
Experimente lernen, Techniken tauschen
Ein spekulatives Handbuch
Das spekulative Handbuch bietet vielfältige Techniken für ein radikales Lernen und Vermitteln. Es umfasst konkrete Anleitungen, Erfahrungen und theoretische Überlegungen. Die Texte beteiligen sich an der Konzeption einer Vermittlung, die das gemeinsame Experimentieren (wieder) einführt.
Im Seminarraum, in Workshops, auf Festivals, in Fluren, Parks und der Stadt finden Lernen und Verlernen statt. Texte und Anleitungen u. a. zu: Filmessays, Collagen, Banküberfällen, der Universität der Toten, wildem Schreiben, konzeptuellem speed Dating, neurodiversem Lernen, Format-Denken, dem Theater der Sorge, dem Schreiblabor, dem Körperstreik.
Im Workshop des Sinnlichen
(2020)
Folgende fiktive Situation soll ein Problem markieren, das in diesem Beitrag diskutiert wird und nach Ansicht seines Autors ein an Kunsthochschulen weit verbreitetes Phänomen darstellt. Im Rahmen eines Mentoring-Workshops stellt eine Gruppe Studierender Arbeitsmaterial ihres aktuellen Projekts vor, um es anschließend in der Gruppe zu besprechen. Ziel des Veranstaltungsformats ist es, die Studierenden während der Entwicklung ihrer künstlerischen Praktiken zu begleiten und diese nicht anhand handwerklicher Kriterien überzudeterminieren, sondern ihrer Eigenlogik zu folgen, den ihnen inhärenten ästhetischen Potentialen nachzugehen und ein Bewusstsein für die Kontexte und Diskurse zu schaffen, in denen sie verortet sind. Die Studierenden, die heute ihr Projekt vorstellen, haben ein Photoalbum mitgebracht, in das sie, der Anordnungslogik von Urlaubsaufnahmen folgend, eine Reihe analoger Photographien geklebt haben, die Dutzende Schnappschüsse einer entfernten Insel zeigen, die sich hinter der den Vordergrund des Bildes einnehmenden Meeresoberfläche abzeichnet.
Why isn't Google welcome in Kreuzberg? Social movement and the effects of Internet on urban space
(2020)
Advances in information and communication technologies such as the Internet have driven a great transformation in the interactions between individuals and the urban environment. As the use of the Internet in cities becomes more intense and diverse, there is also a restructuring of urban space, which is experienced by groups in society in various ways, according to the specificity of each context. Accordingly, large Internet companies have emerged as new players in the processes of urbanization, either through partnerships with the public administration or through various services offered directly to urban residents. Once these corporations are key actors in the digitalization of urban services, their operations can affect the patterns of urban inequality and generate a series of new struggles over the production of space. Interested in analyzing this phenomena from the perspective of civil society, the present Master Thesis examined a social movement that prevented Google to settle a new startup campus in the district of Kreuzberg, in Berlin. By asking why Google was not welcome in that context, this study also sought to understand how internet, as well as its main operators, has affected everyday life in the city. Thus, besides analyzing the movement, I investigated the particularities of the urban context where it arose and the elements that distinguish the mobilization’s opponent. In pursuit of an interdisciplinary approach, I analyzed and discussed the results of empirical research in dialogue with critical theories in the fields of urban studies and the Internet, with emphasis on Castells' definitions of urban social movements and network society (1983, 2009, 2015), Couldry's and Mejias' (2019) idea of data colonialism, Lefèbvre's (1991, 1996) concepts of abstract space and the right to the city, as well as Zuboff's (2019) theory of surveillance capitalism. The case at hand has exposed that Google plays a prominent role in the way the Internet has been developed and deployed in cities. From the perspective accessed, the current appropriation of Internet technologies has been detrimental to individual autonomy and has contributed to intensifying existing inequalities in the city. The alternative vision to this relies mainly on the promotion of decentralized solidarity networks.
This research aims to model soil temperature (ST) using machine learning models of multilayer perceptron (MLP) algorithm and support vector machine (SVM) in hybrid form with the Firefly optimization algorithm, i.e. MLP-FFA and SVM-FFA. In the current study, measured ST and meteorological parameters of Tabriz and Ahar weather stations in a period of 2013–2015 are used for training and testing of the studied models with one and two days as a delay. To ascertain conclusive results for validation of the proposed hybrid models, the error metrics are benchmarked in an independent testing period. Moreover, Taylor diagrams utilized for that purpose. Obtained results showed that, in a case of one day delay, except in predicting ST at 5 cm below the soil surface (ST5cm) at Tabriz station, MLP-FFA produced superior results compared with MLP, SVM, and SVM-FFA models. However, for two days delay, MLP-FFA indicated increased accuracy in predicting ST5cm and ST 20cm of Tabriz station and ST10cm of Ahar station in comparison with SVM-FFA. Additionally, for all of the prescribed models, the performance of the MLP-FFA and SVM-FFA hybrid models in the testing phase was found to be meaningfully superior to the classical MLP and SVM models.
Wie können journalistische Angebote nachhaltig finanziert werden? Dies bleibt die Kernfrage für Medienhäuser und journalistische Neugründungen bei der Entwicklung und beim Aufbau tragfähiger digitaler Geschäftsmodelle.
Die Autoren des vorliegenden Bandes vermitteln einen breiten Überblick über den Wissensstand zum Thema Paid Content, Plattformen und Zahlungsbereitschaft im Journalismus und eröffnen innovative Blickwinkel auf neuartige Plattformmodelle ebenso wie auf Motive und Bedürfnisse der Nutzerinnen und Nutzer digitaljournalistischer Inhalte. Auf Grundlage empirischer Forschung werden Handlungsempfehlungen für die nutzerzentrierte Ausgestaltung von Paid-Content-Angeboten sowie neue Perspektiven auf Zahlungsbereitschaft im digitalen Journalismus erschlossen – relevant sowohl für die Wissenschaft wie auch für die Medienpraxis.
Das spekulative Handbuch bietet vielfältige Techniken für ein radikales Lernen und Vermitteln. Es umfasst konkrete Anleitungen, Erfahrungen und theoretische Überlegungen. Die Texte beteiligen sich an der Konzeption einer Vermittlung, die das gemeinsame Experimentieren (wieder) einführt.
Im Seminarraum, in Workshops, auf Festivals, in Fluren, Parks und der Stadt finden Lernen und Verlernen statt. Texte und Anleitungen u. a. zu: Filmessays, Collagen, Banküberfällen, der Universität der Toten, wildem Schreiben, konzeptuellem speed Dating, neurodiversem Lernen, Format-Denken, dem Theater der Sorge, dem Schreiblabor, dem Körperstreik.
In einer systematischen Interpretation von Vilém Flussers Werk schlägt die Arbeit vor, Flussers Ansatz als einen medienphilosophischen zu verstehen, insofern er das „wie“ der medienphilosophischen Fragestellung in den Mittelpunkt rückt. Medien werden nicht erst dann zu einem wesentlichen Bestandteil von Flussers Philosophie, wenn er sie explizit zum Gegenstand seiner Untersuchungen der gegenwärtigen Kultur und Gesellschaft oder historischer Rückblicke macht; Denken vollzieht sich immer in Medien oder medialen Praktiken, es wird nicht nur von ihnen (mit) geprägt – ohne Medien gäbe es kein Denken und umgekehrt verändert sich Philosophie mit den (jeweils) neuen Medien. Ausgehend von Begriffen oder eher Denkfiguren, die neben dem „was“ des jeweils verhandelten Themas auch das „wie“ der Reflexion selbst adressieren, wird der „Umbruch in der Struktur des Denkens“ zugleich als Beschreibung von Medienumbrüchen verstanden – mit dem Fluchtpunkt des Sprungs in das Universum der Komputation – und als Vollzug der gegenwärtigen Veränderung der „Methode des Denkens“. Flussers (Ver)Suche einer Reflexion, die nicht mehr durch das Medium Schrift strukturiert ist, sondern sowohl alten Medien wie dem Bild – bzw. Praktiken des Abbildens, Darstellens, Einbildens usw. – als auch neuen Medien – dem Komputieren – Geltung verschafft, laufen auf eine widersprüchliche Diagnose des neuen Universums der Komputation (anders: der technischen Bilder) hinaus : eine kybermetisch inspirierte Vision der frei modellierbaren Wirklichkeit(en) einerseits und die Dystopie einer Welt, in der Apparaten Denken, Wahrnehmen und Handeln beherrschen andererseits. Die Arbeit zeigt auf, wie Flusser zu dieser Aporie der Medienreflexion – die weit über Flussers Werk hinaus virulent bleibt – gelangt und wie sie, ausgehend von seiner Figur der Geste, im Sinne einer performativen Medienreflexion gelöst werden könnte.
Energy‐Efficient Method for Wireless Sensor Networks Low‐Power Radio Operation in Internet of Things
(2020)
The radio operation in wireless sensor networks (WSN) in Internet of Things (IoT)applications is the most common source for power consumption. Consequently, recognizing and controlling the factors affecting radio operation can be valuable for managing the node power consumption. Among essential factors affecting radio operation, the time spent for checking the radio is of utmost importance for monitoring power consumption. It can lead to false WakeUp or idle listening in radio duty cycles and ContikiMAC. ContikiMAC is a low‐power radio duty‐cycle protocol in Contiki OS used in WakeUp mode, as a clear channel assessment (CCA) for checking radio status periodically. This paper presents a detailed analysis of radio WakeUp time factors of ContikiMAC. Furthermore, we propose a lightweight CCA (LW‐CCA) as an extension to ContikiMAC to reduce the Radio Duty‐Cycles in false WakeUps and idle listening though using dynamic received signal strength indicator (RSSI) status check time. The simulation results in the Cooja simulator show that LW‐CCA reduces about 8% energy consumption in nodes while maintaining up to 99% of the packet delivery rate (PDR).
The classical Internet of things routing and wireless sensor networks can provide more precise monitoring of the covered area due to the higher number of utilized nodes. Because of the limitations in shared transfer media, many nodes in the network are prone to the collision in simultaneous transmissions. Medium access control protocols are usually more practical in networks with low traffic, which are not subjected to external noise from adjacent frequencies. There are preventive, detection and control solutions to congestion management in the network which are all the focus of this study. In the congestion prevention phase, the proposed method chooses the next step of the path using the Fuzzy decision-making system to distribute network traffic via optimal paths. In the congestion detection phase, a dynamic approach to queue management was designed to detect congestion in the least amount of time and prevent the collision. In the congestion control phase, the back-pressure method was used based on the quality of the queue to decrease the probability of linking in the pathway from the pre-congested node. The main goals of this study are to balance energy consumption in network nodes, reducing the rate of lost packets and increasing quality of service in routing. Simulation results proved the proposed Congestion Control Fuzzy Decision Making (CCFDM) method was more capable in improving routing parameters as compared to recent algorithms.
Coronary Artery Disease Diagnosis: Ranking the Significant Features Using a Random Trees Model
(2020)
Heart disease is one of the most common diseases in middle-aged citizens. Among the vast number of heart diseases, coronary artery disease (CAD) is considered as a common cardiovascular disease with a high death rate. The most popular tool for diagnosing CAD is the use of medical imaging, e.g., angiography. However, angiography is known for being costly and also associated with a number of side effects. Hence, the purpose of this study is to increase the accuracy of coronary heart disease diagnosis through selecting significant predictive features in order of their ranking. In this study, we propose an integrated method using machine learning. The machine learning methods of random trees (RTs), decision tree of C5.0, support vector machine (SVM), and decision tree of Chi-squared automatic interaction detection (CHAID) are used in this study. The proposed method shows promising results and the study confirms that the RTs model outperforms other models.
Das Hauptziel der vorliegenden Arbeit war es, eine stetige Kopplung zwischen der ananlytischen und numerischen Lösung von Randwertaufgaben mit Singularitäten zu realisieren. Durch die inter-polationsbasierte gekoppelte Methode kann eine globale C0 Stetigkeit erzielt werden. Für diesen Zweck wird ein spezielle finite Element (Kopplungselement) verwendet, das die Stetigkeit der Lösung sowohl mit dem analytischen Element als auch mit den normalen CST Elementen gewährleistet.
Die interpolationsbasierte gekoppelte Methode ist zwar für beliebige Knotenanzahl auf dem Interface ΓAD anwendbar, aber es konnte durch die Untersuchung von der Interpolationsmatrix und numerische Simulationen festgestellt werden, dass sie schlecht konditioniert ist. Um das Problem mit den numerischen Instabilitäten zu bewältigen, wurde eine approximationsbasierte Kopplungsmethode entwickelt und untersucht. Die Stabilität dieser Methode wurde anschließend anhand der Untersuchung von der Gramschen Matrix des verwendeten Basissystems auf zwei Intervallen [−π,π] und [−2π,2π] beurteilt. Die Gramsche Matrix auf dem Intervall [−2π,2π] hat einen günstigeren Konditionszahl in der Abhängigkeit von der Anzahl der Kopplungsknoten auf dem Interface aufgewiesen. Um die dazu gehörigen numerischen Instabilitäten ausschließen zu können wird das Basissystem mit Hilfe vom Gram-Schmidtschen Orthogonalisierungsverfahren auf beiden Intervallen orthogonalisiert. Das orthogonale Basissystem lässt sich auf dem Intervall [−2π,2π] mit expliziten Formeln schreiben. Die Methode des konsistentes Sampling, die häufig in der Nachrichtentechnik verwendet wird, wurde zur Realisierung von der approximationsbasierten Kopplung herangezogen. Eine Beschränkung dieser Methode ist es, dass die Anzahl der Sampling-Basisfunktionen muss gleich der Anzahl der Wiederherstellungsbasisfunktionen sein. Das hat dazu geführt, dass das eingeführt Basissys-tem (mit 2 n Basisfunktionen) nur mit n Basisfunktion verwendet werden kann.
Zur Lösung diese Problems wurde ein alternatives Basissystems (Variante 2) vorgestellt. Für die Verwendung dieses Basissystems ist aber eine Transformationsmatrix M nötig und bei der Orthogonalisierung des Basissystems auf dem Intervall [−π,π] kann die Herleitung von dieser Matrix kompliziert und aufwendig sein. Die Formfunktionen wurden anschließend für die beiden Varianten hergeleitet und grafisch (für n = 5) dargestellt und wurde gezeigt, dass diese Funktionen die Anforderungen an den Formfunktionen erfüllen und können somit für die FE- Approximation verwendet werden.
Anhand numerischer Simulationen, die mit der Variante 1 (mit Orthogonalisierung auf dem Intervall [−2π,2π]) durchgeführt wurden, wurden die grundlegenden Fragen (Beispielsweise: Stetigkeit der Verformungen auf dem Interface ΓAD, Spannungen auf dem analytischen Gebiet) über-
prüft.
This thesis explores how cultural heritage plays a role in the development of urban identity by engaging both actively and passively with memory, i.e. remembering and forgetting. I argue that architectural heritage is a medium where specific cultural and social decisions form its way of presentation, and it reflects the values and interests of the period. By the process of remembering and forgetting, the meanings between inhabitant and object in urban environment are practiced, and the meanings are created.
To enable the research in narrative observation, cultural tourism management is chosen as the main research object, which reflects the alteration of interaction between the architectural heritage and urban identity. Identifying the role of heritage management, the definition of social resilience and the prospects of cultural heritage as a means of social resilience are addressed. Case region of the research is East Ger- many, thereby, the study examines the distinct approaches and objectives regarding heritage management under the different political systems along the German reunification process.
The framework is based on various theoretical paradigms to investigate the broad research questions: 1) What is the role of historic urban quarters in the revitalisation of East German towns? 2) How was the transition processed by cultural heritage management? 3) How did policy affect residents’ lives?
The case study is applied to macro level (city level: Gotha and Eisenach) and micro level study (object level: specific heritage sites), to analyse the performance of selective remembering and making tourist destination through giving significance to specific heritage. By means of site observations, archival research, qualitative inter- views, photographs, and discourse analysis on printed tourism materials, the study demonstrates that certain sites and characteristics of the city enable creating and focusing messages, which aids the social resilience.
Combining theory and empirical studies this thesis attempts to widen the academic discussion regarding the practice of remembering and forgetting driven by cultural heritage. The thesis argues for cultural heritage tourism as an element of social resilience and one that embraces the historic and cultural identity of the inhabitants.
The longitudinal dispersion coefficient (LDC) plays an important role in modeling the transport of pollutants and sediment in natural rivers. As a result of transportation processes, the concentration of pollutants changes along the river. Various studies have been conducted to provide simple equations for estimating LDC. In this study, machine learning methods, namely support vector regression, Gaussian process regression, M5 model tree (M5P) and random forest, and multiple linear regression were examined in predicting the LDC in natural streams. Data sets from 60 rivers around the world with different hydraulic and geometric features were gathered to develop models for LDC estimation. Statistical criteria, including correlation coefficient (CC), root mean squared error (RMSE) and mean absolute error (MAE), were used to scrutinize the models. The LDC values estimated by these models were compared with the corresponding results of common empirical models. The Taylor chart was used to evaluate the models and the results showed that among the machine learning models, M5P had superior performance, with CC of 0.823, RMSE of 454.9 and MAE of 380.9. The model of Sahay and Dutta, with CC of 0.795, RMSE of 460.7 and MAE of 306.1, gave more precise results than the other empirical models. The main advantage of M5P models is their ability to provide practical formulae. In conclusion, the results proved that the developed M5P model with simple formulations was superior to other machine learning models and empirical models; therefore, it can be used as a proper tool for estimating the LDC in rivers.
Temporary changes in precipitation may lead to sustained and severe drought or massive floods in different parts of the world. Knowing the variation in precipitation can effectively help the water resources decision-makers in water resources management. Large-scale circulation drivers have a considerable impact on precipitation in different parts of the world. In this research, the impact of El Niño-Southern Oscillation (ENSO), Pacific Decadal Oscillation (PDO), and North Atlantic Oscillation (NAO) on seasonal precipitation over Iran was investigated. For this purpose, 103 synoptic stations with at least 30 years of data were utilized. The Spearman correlation coefficient between the indices in the previous 12 months with seasonal precipitation was calculated, and the meaningful correlations were extracted. Then, the month in which each of these indices has the highest correlation with seasonal precipitation was determined. Finally, the overall amount of increase or decrease in seasonal precipitation due to each of these indices was calculated. Results indicate the Southern Oscillation Index (SOI), NAO, and PDO have the most impact on seasonal precipitation, respectively. Additionally, these indices have the highest impact on the precipitation in winter, autumn, spring, and summer, respectively. SOI has a diverse impact on winter precipitation compared to the PDO and NAO, while in the other seasons, each index has its special impact on seasonal precipitation. Generally, all indices in different phases may decrease the seasonal precipitation up to 100%. However, the seasonal precipitation may increase more than 100% in different seasons due to the impact of these indices. The results of this study can be used effectively in water resources management and especially in dam operation.
Hydrological drought forecasting plays a substantial role in water resources management. Hydrological drought highly affects the water allocation and hydropower generation. In this research, short term hydrological drought forecasted based on the hybridized of novel nature-inspired optimization algorithms and Artificial Neural Networks (ANN). For this purpose, the Standardized Hydrological Drought Index (SHDI) and the Standardized Precipitation Index (SPI) were calculated in one, three, and six aggregated months. Then, three states where proposed for SHDI forecasting, and 36 input-output combinations were extracted based on the cross-correlation analysis. In the next step, newly proposed optimization algorithms, including Grasshopper Optimization Algorithm (GOA), Salp Swarm algorithm (SSA), Biogeography-based optimization (BBO), and Particle Swarm Optimization (PSO) hybridized with the ANN were utilized for SHDI forecasting and the results compared to the conventional ANN. Results indicated that the hybridized model outperformed compared to the conventional ANN. PSO performed better than the other optimization algorithms. The best models forecasted SHDI1 with R2 = 0.68 and RMSE = 0.58, SHDI3 with R 2 = 0.81 and RMSE = 0.45 and SHDI6 with R 2 = 0.82 and RMSE = 0.40.