### Refine

#### Year of publication

- 2013 (62) (remove)

#### Document Type

- Doctoral Thesis (36)
- Preprint (18)
- Report (2)
- Article (1)
- Conference Proceeding (1)
- Habilitation (1)
- Master's Thesis (1)
- Periodical Part (1)
- Working Paper (1)

#### Language

- English (62) (remove)

#### Keywords

- Chisel (2)
- Field-programmable gate array (FPGA) (2)
- IR-MPD (2)
- cobalt (2)
- dipeptide (2)
- metal (2)
- AXI4-Stream (1)
- Ahr Knockout Model (1)
- Analytical method (1)
- Aryl hydrocarbon Receptor (1)

#### Faculty / Organisational entity

The main purpose of the study was to improve the physical properties of the modelling of compressed materials, especially fibrous materials. Fibrous materials are finding increasing application in the industries. And most of the materials are compressed for different applications. For such situation, we are interested in how the fibre arranged, e.g. with which distribution. For given materials it is possible to obtain a three-dimensional image via micro computed tomography. Since some physical parameters, e.g. the fibre lengths or the directions for points in the fibre, can be checked under some other methods from image, it is beneficial to improve the physical properties by changing the parameters in the image.
In this thesis, we present a new maximum-likelihood approach for the estimation of parameters of a parametric distribution on the unit sphere, which is various as some well known distributions, e.g. the von-Mises Fisher distribution or the Watson distribution, and for some models better fit. The consistency and asymptotic normality of the maximum-likelihood estimator are proven. As the second main part of this thesis, a general model of mixtures of these distributions on a hypersphere is discussed. We derive numerical approximations of the parameters in an Expectation Maximization setting. Furthermore we introduce a non-parametric estimation of the EM algorithm for the mixture model. Finally, we present some applications to the statistical analysis of fibre composites.

Conditional Compilation (CC) is frequently used as a variation mechanism in software product lines (SPLs). However, as a SPL evolves the variable code realized by CC erodes in the sense that it becomes overly complex and difficult to understand and maintain. As a result, the SPL productivity goes down and puts expected advantages more and more at risk. To investigate the variability erosion and keep the productivity above a sufficiently good level, in this paper we 1) investigate several erosion symptoms in an industrial SPL; 2) present a variability improvement process that includes two major improvement strategies. While one strategy is to optimize variable code within the scope of CC, the other strategy is to transition CC to a new variation mechanism called Parameterized Inclusion. Both of these two improvement strategies can be conducted automatically, and the result of CC optimization is provided. Related issues such as applicability and cost of the improvement are also discussed.

As a Software Product Line (SPL) evolves with increasing number of features and feature values, the feature correlations become extremely intricate, and the specifications of these correlations tend to be either incomplete or inconsistent with their realizations, causing misconfigurations in practice. In order to guide product configuration processes, we present a solution framework to recover complex feature correlations from existing product configurations. These correlations are further pruned automatically and validated by domain experts. During implementation, we use association mining techniques to automatically extract strong association rules as potential feature correlations. This approach is evaluated using a large-scale industrial SPL in the embedded system domain, and finally we identify a large number of complex feature correlations.

The noise issue in manufacturing system is widely discussed from legal and health aspects. Regarding the existing laws and guidelines, various investigation methods are implemented in industry. The sound pressure level can be measured and reduced by using established approaches in reality. However, a straightforward and low cost approach to study noise issue using existing digital factory models is not found.
This thesis attempts to develop a novel concept for sound pressure level investigation in a virtual environment. With this, the factory planners are able to investigate the noise issue during factory design and layout planning phase.
Two computer aided tools are used in this approach: acoustic simulation and virtual reality (VR). The former enables the planner to simulate the sound pressure level by given factory layout and facility sound features. And the latter provides a visualization environment to view and explore the simulation results. The combination of these two powerful tools provides the planners a new possibility to analyze the noise in a factory.
To validate the simulations, the acoustic measurements are implemented in a real factory. Sound pressure level and sound intensity are determined respectively. Furthermore, a software tool is implemented using the introduced concept and approach. With this software, the simulation results are represented in a Cave Automatic Virtual Environment (CAVE).
This thesis describes the development of the approach, the measurement of sound features, the design of visualization framework, and the implementation of VR software. Based on this know-how, the industry users are able to design their own method and software for noise investigation and analysis.

Backward compatibility of class libraries ensures that an old implementation of a library can safely be replaced by a new implementation without breaking existing clients.
Formal reasoning about backward compatibility requires an adequate semantic model to compare the behavior of two library implementations.
In the object-oriented setting with inheritance and callbacks, finding such models is difficult as the interface between library implementations and clients are complex.
Furthermore, handling these models in a way to support practical reasoning requires appropriate verification tools.
This thesis proposes a formal model for library implementations and a reasoning approach for backward compatibility that is implemented using an automatic verifier. The first part of the thesis develops a fully abstract trace-based semantics for class libraries of a core sequential object-oriented language. Traces abstract from the control flow (stack) and data representation (heap) of the library implementations. The construction of a most general context is given that abstracts exactly from all possible clients of the library implementation.
Soundness and completeness of the trace semantics as well as the most general context are proven using specialized simulation relations on the operational semantics. The simulation relations also provide a proof method for reasoning about backward compatibility.
The second part of the thesis presents the implementation of the simulation-based proof method for an automatic verifier to check backward compatibility of class libraries written in Java. The approach works for complex library implementations, with recursion and loops, in the setting of unknown program contexts. The verification process relies on a coupling invariant that describes a relation between programs that use the old library implementation and programs that use the new library implementation. The thesis presents a specification language to formulate such coupling invariants. Finally, an application of the developed theory and tool to typical examples from the literature validates the reasoning and verification approach.

In this paper we analyze the vibrations of nonlinear structures by means of the novel approach of isogeometric finite elements. The fundamental idea of isogeometric finite elements is to apply the same functions, namely B-Splines and NURBS (Non-Uniform Rational B-Splines), for describing the geometry and for representing the numerical solution. In case of linear vibrational analysis, this approach has already been shown to possess substantial advantages over classical finite elements, and we extend it here to a nonlinear framework based on the harmonic balance principle.
As application, the straight nonlinear Euler-Bernoulli beam is used, and overall, it is demonstrated that isogeometric finite elements with B-Splines in combination with the harmonic balance method are a powerful means for the analysis of nonlinear structural vibrations. In particular, the smoother k-method provides higher accuracy than the p-method for isogeometric nonlinear vibration analysis.

Due to tremendous improvements of high-performance computing resources as well
as numerical advances computational simulations became a common tool for modern
engineers. Nowadays, simulation of complex physics is more and more substituting a
large amount of physical experiments. While the vast compute power of large-scale
high-performance systems enabled for simulating more complex numerical equations,
handling the ever increasing amount of data with spatial and temporal resolution
burdens new challenges to scientists. Huge hardware and energy costs desire for
ecient utilization of high-performance systems. However, increasing complexity of
simulations raises the risk of failing simulations resulting in a single simulation to be
restarted multiple times. Computational Steering is a promising approach to interact
with running simulations which could prevent simulation crashes. The large amount
of data expands gaps in the amount of data that can be calculated and the amount of
data that can be processed. Extreme-scale simulations produce more data that can
even be stored. In this thesis, I propose several methods that enhance the process
of steering, exploring, visualizing, and analyzing ongoing numerical simulations.

Hardware prototyping is an essential part in the hardware design flow. Furthermore, hardware prototyping usually relies on system-level design and hardware-in-the-loop simulations in order to develop, test and evaluate intellectual property cores. One common task in this process consist on interfacing cores with different port specifications. Data width conversion is used to overcome this issue. This work presents two open source hardware cores compliant with AXI4-Stream bus protocol, where each core performs upsizing/downsizing data width conversion.

It is well known that the greedy algorithm solves matroid base problems for all linear cost functions and is, in fact, correct if and only if the underlying combinatorial structure of the problem is a matroid. Moreover, the algorithm can be applied to problems with sum, bottleneck, algebraic sum or \(k\)-sum objective functions.

In this paper we consider a multivariate switching model, with constant states means
and covariances. In this model, the switching mechanism between the basic states of
the observed time series is controlled by a hidden Markov chain. As illustration, under
Gaussian assumption on the innovations and some rather simple conditions, we prove
the consistency and asymptotic normality of the maximum likelihood estimates of the model parameters.

Investigate the hardware description language Chisel - A case study implementing the Heston model
(2013)

This paper presents a case study comparing the hardware description language „Constructing Hardware in a Scala Embedded Language“(Chisel) to VHDL. For a thorough comparison the Heston Model was implemented, a stochastic model used in financial mathematics to calculate option prices. Metrics like hardware utilization and maximum clock rate were extracted from both resulting designs and compared to each other. The results showed a 30% reduction in code size compared to VHDL, while the resulting circuits had about the same hardware utilization. Using Chisel however proofed to be difficult because of a few features that were not available for this case study.

We prove the global existence, along with some basic boundedness properties, of weak solutions to a PDE-ODE system modeling the multiscale invasion of tumor cells through the surrounding tissue matrix. The model has been proposed in [22] and accounts on the macroscopic level for the evolution of cell and tissue densities, along with the concentration of a chemoattractant, while on the subcellular level it involves the binding of integrins to soluble and insoluble components of the peritumoral region. The connection between the two scales is realized with the aid of a contractivity function characterizing the ability of the tumor cells to adapt their motility behavior
to their subcellular dynamics.
The resulting system, consisting of three partial and three ordinary differential equations including a temporal delay, in particular involves chemotactic and haptotactic cross-diffusion. In order to overcome technical obstacles stemming from the corresponding highest-order interaction terms, we base our analysis on a certain functional, inter alia involving the cell and tissue densities in the diffusion and haptotaxis terms respectively, which is shown to enjoy a quasi-dissipative property. This will be used as a starting point for the derivation of a series of integral estimates finally allowing for the construction of a generalized solution as the limit of solutions to suitably regularized problems.

In most cases in a safety analysis the influences of security problems are omitted or even forgotten. Because more and more systems are accessible from outside the system via maintenance interfaces, this missing security analysis is becoming a problem. This is why we propose an approach on how to extend the safety analysis by security aspects. Such a more comprehensive analysis should lead to systems that react in less catastrophic ways to attacks.

This thesis deals with generalized inverses, multivariate polynomial interpolation and approximation of scattered data. Moreover, it covers the lifting scheme, which basically links the aforementioned topics. For instance, determining filters for the lifting scheme is connected to multivariate polynomial interpolation. More precisely, sets of interpolation sites are required that can be interpolated by a unique polynomial of a certain degree. In this thesis a new class of such sets is introduced and elements from this class are used to construct new and computationally more efficient filters for the lifting scheme.
Furthermore, a method to approximate multidimensional scattered data is introduced which is based on the lifting scheme. A major task in this method is to solve an ordinary linear least squares problem which possesses a special structure. Exploiting this structure yields better approximations and therefore this particular least squares problem is analyzed in detail. This leads to a characterization of special generalized inverses with partially prescribed image spaces.

Palladium-Catalyzed C–C Bond Formations via Activation of Carboxylic Acids and Their Derivatives
(2013)

Applications of carboxylic acids and their derivatives in transition metal-catalyzed cross-coupling reactions regio-selectively forming Csp3-Csp2, and Csp2-Csp2 bonds were explored in this thesis. Several important organic building blocks such as aryl acetates, diaryl acetates, imines, ketones, biaryls, styrenes and polysubstituted alkenes were successfully accessed from carboxylic acids and their derivatives by the means of C–H activation and decarboxylative cross-couplings.
An efficient and practical protocol for the synthesis of biologically important ethyl 2-arylacates through the dealkoxycarbonlative cross-coupling reaction between aryl halides and malonates was developed. Activation of the alpha-proton of alkyl esters by a copper catalyst allowed the deprotonation of esters even in the presence of mild bases, leading to a straightforward and efficient approach to alkyl alpha-diarylacetate from simple alkyl acetates and aryl halides.
The addition of a primary amine into the coupling reaction of alpha-oxocarboxylic acids and aryl halides led to an unprecedented low-temperature redox-neutral decarboxylative coupling process, providing a green and efficient method for the preparation of azomethines, in which all the three substituents can be independently varied. A minor modification of this protocol allowed us to easily access the corresponding ketones.
The decarboxylative coupling of robust aryl mesylates as well as polysubstituted alkenyl mesylates using our customized imidazolyl phosphine ligands was realized, further expanding the scope of carbon electrophiles in decarboxylative coupling reactions. Variation of the ligands led to two complementary protocols, providing the corresponding biaryls and polysubstituted olefins in high yields.
The use of a new class of pyrimidinyl phosphine ligands dramatically reduced the reaction temperatures of decarboxylative cross-coupling reactions between aromatic carboxylic acids and aryl or alkenyl triflates. The new catalyst system for the first time allowed the efficient decarboxylative biaryls synthesis at only 100 °C, representing a significant achievement in redox-neutral decarboxylative coupling reactions.

This thesis is divided into two parts. Both cope with multi-class image segmentation and utilize
non-smooth optimization algorithms.
The topic of the first part, namely unsupervised segmentation, is the application of clustering
to image pixels. Therefore, we start with an introduction of the biconvex center-based clustering
algorithms c-means and fuzzy c-means, where c denotes the number of classes. We show that
fuzzy c-means can be seen as an approximation of c-means in terms of power means.
Since noise is omnipresent in our image data, these simple clustering models are not suitable
for its segmentation. To this end, we introduce a general and finite dimensional segmentation
model that consists of a data term stemming from the aforementioned clustering models plus a
continuous regularization term. We tackle this optimization model via an alternating minimiza-
tion approach called regularized c-centers (RcC). Thereby, we fix the centers and optimize the
segment membership of the pixels and vice versa. In this general setting, we prove convergence
in the sense of set-valued algorithms using Zangwill’s Theory [172].
Further, we present a segmentation model with a total variation regularizer. While updating
the cluster centers is straightforward for fixed segment memberships of the pixels, updating the
segment membership can be solved iteratively via non-smooth, convex optimization. Thereby,
we do not iterate a convex optimization algorithm until convergence. Instead, we stop as soon as
we have a certain amount of decrease in the objective functional to increase the efficiency. This
algorithm is a particular implementation of RcC providing also the corresponding convergence
theory. Moreover, we show the good performance of our method in various examples such as
simulated 2d images of brain tissue and 3d volumes of two materials, namely a multi-filament
composite superconductor and a carbon fiber reinforced silicon carbide ceramics. Thereby, we
exploit the property of the latter material that two components have no common boundary in
our adapted model.
The second part of the thesis is concerned with supervised segmentation. We leave the area
of center based models and investigate convex approaches related to graph p-Laplacians and
reproducing kernel Hilbert spaces (RKHSs). We study the effect of different weights used to
construct the graph. In practical experiments we show on the one hand image types that
are better segmented by the p-Laplacian model and on the other hand images that are better
segmented by the RKHS-based approach. This is due to the fact that the p-Laplacian approach
provides smoother results, while the RKHS approach provides often more accurate and detailed
segmentations. Finally, we propose a novel combination of both approaches to benefit from the
advantages of both models and study the performance on challenging medical image data.

In recent years, recommender systems have been widely used for a variety of different kinds of items such as books, movies, and music. However, current recommendation approaches have often been criticized to suffer from overspecialization thus not enough considering a user’s diverse topics of interest. In this thesis we present a novel approach to extracting contextualized user profiles which enable recommendations taking into account a user’s full range of interests. The method applies algorithms from the domain of topic detection and tracking to automatically identify diverse user interests and to represent them with descriptive labels. That way manual annotations of interest topics by the users, e. g., from a predefined domain taxonomy, are no longer required. The approach has been tested in two scenarios: First, we implemented a content-based recommender system for an Enterprise 2.0 resource sharing platform where the contextualized user interest profiles have been used to generate recommendations with a high degree of inter-topic diversity. In an effort to harness the collective intelligence of the users, the resources in the system were described by making use of user-generated metadata. The evaluation experiments show that our approach is likely to capture a multitude of diverse interest topics per user. The labels extracted are specific for these topics and can be used to retrieve relevant on-topic resources. Second, a slightly adapted variation of the algorithm has been used to target music recommendations based on the user’s current mood. In this scenario music artists are described by using freely available Semantic Web data from the Linked Open Data cloud thus not requiring expensive metadata annotations by experts. The evaluation experiments conducted show that many users have a multitude of different preferred music styles. However a correlation between these music styles and music mood categories could not be observed. An integration of our proposed user profiles with existing user model ontologies seems promising for enabling context-sensitive recommendations.

There is a growing trend for ever larger wireless sensor networks (WSNs) consisting of thousands or tens of thousands of sensor nodes (e.g., [91, 79]). We believe this trend will continue and thus scalability plays a crucial role in all protocols and mechanisms for WSNs. Another trend in many modern WSN applications is the time sensitivity to information from sensors to sinks. In particular, WSNs are a central part of the vision of cyber-physical systems and as these are basically closed-loop systems many WSN applications will have to operate under stringent timing requirements. Hence, it is crucial to develop algorithms that minimize the worst-case delay in WSNs. In addition, almost all WSNs consist of battery-powered nodes, and thus energy-efficiency clearly remains another premier goal in order to keep network lifetime high. This dissertation presents and evaluates designs for WSNs using multiple sinks to achieve high lifetime and low delay. Firstly, we investigate random and deterministic node placement strategies for large-scale and time-sensitive WSNs. In particular, we focus on tiling-based deterministic node placement strategies and analyze their effects on coverage, lifetime, and delay performance under both exact placement and stochastically disturbed placement. Next, we present sink placement strategies, which constitutes the main contributions of this dissertation. Static sinks will be placed and mobile sinks will be given a trajectory. A proper sink placement strategy can improve the performance of a WSN significantly. In general, the optimal sink placement with lifetime maximization is an NP-hard problem. The problem is even harder if delay is taken into account. In order to achieve both lifetime and delay goals, we focus on the problem of placing multiple (static) sinks such that the maximum worst-case delay is minimized while keeping the energy consumption as low as possible. Different target networks may need a corresponding sink placement strategy under differing levels of apriori assumptions. Therefore, we first develop an algorithm based on the Genetic Algorithm (GA) paradigm for known sensor nodes' locations. For a network where global information is not feasible we introduce a self-organized sink placement (SOSP) strategy. While GA-based sink placement achieves a near-optimal solution, SOSP provides a good sink placement strategy with a lower communication overhead. How to plan the trajectories of many mobile sinks in very large WSNs in order to simultaneously achieve lifetime and delay goals had not been treated so far in the literature. Therefore, we delve into this difficult problem and propose a heuristic framework using multiple orbits for the sinks' trajectories. The framework is designed based on geometric arguments to achieve both, high lifetime and low delay. In simulations, we compare two different instances of our framework, one conceived based on a load-balancing argument and one based on a distance minimization argument, with a set of different competitors spanning from statically placed sinks to battery-state aware strategies. We find our heuristics outperform the competitors in both, lifetime and delay. Furthermore, and probably even more important, the heuristic, while keeping its good delay and lifetime performance, scales well with an increasing number of sinks. In brief, the goal of this dissertation is to show that placing nodes and sinks in conventional WSNs as well as planning trajectories in mobility enabled WSNs carefully really pays off for large-scale and time-sensitive WSNs.

Data usage control is a concept that extends access control to also protect data after it
has been released. Usage control enforcement relies on available information about the
distribution of data in the monitored system. In this thesis we introduce an information
flow tracking approach for JavaScript in order to enable usage control for dynamic content
in web browsers. The proposed model is implemented as a prototype in the JavaScript
engine V8 of the Chromium browser to evaluate the feasibility of the chosen approach.

Recent progresses and advances in the field of consumer electronics, driven by display
technologies and also the sector of mobile, hand-held devices, enable new ways in
presenting information to users, as well as new ways of user interaction, therefore
providing a basis for user-centered applications and work environments.
My thesis focuses on how arbitrary display environments can be utilized to improve
both the user experience, regarding perception of information, and also to provide
intuitive interaction possibilities. On the one hand advances in display technologies
provide the basis for new ways of visualizing content and collaborative work, on the
other hand forward-pressing developments in the consumer market, especially the
market of smart phones, offer potential to enhance usability in terms of interaction
and therefore can provide additional benefit for users.
Tiled display setups, combining both large screen real estate and high resolution,
provide new possibilities and chances to visualize large datasets and to facilitate col-
laboration in front of a large screen area. Furthermore these display setups present
several advantages over the traditional single-user-workspace environments: con-
trary to single-user-workspaces, multiple users are able to explore a dataset displayed
on a tiled display system, at the same time, thus allowing new forms of collabora-
tive work. Based on that, face-to-face discussions are enabled, an additional value
is added. Large displays also allow the utilization of the user’s spatial memory, al-
lowing physical navigation without the need of switching between different windows
to explore information.
With Tiled++ I contributed a versatile approach to address the bezel problem. The
bezel problem is one of the Top Ten research challenges in the research field of LCD-
based tiled wall setups. By applying the Tiled++ approach a large high resolution
Focus & Context screen is created, combining high resolution focus areas with low
resolution context information, projected onto the bezel area.
Additionally the field of user interaction poses an important challenge, especially
regarding the utilization of large tiled displays, since traditional keyboard & mouse
interaction devices reached their limits. My focus in this thesis is on Mobile HCI.Devices like mobile phones are utilized to interact with large displays, since they
feature various interaction modalities and preserve user mobility.
Large public displays, as a modernized form of traditional bulletin boards, also en-
able new ways of handling information, displaying content, and user interaction.
Utilized in hot spots, Digital Interactive Public Pinboards can provide an adequate
answer to questions like how to approach pressing issues like disaster and crisis man-
agement for both responders as well as citizens and also new ways of how to handle
information flow (contribution & distribution & accession). My contribution to the
research field of public display environments was the conception and implementa-
tion of an easy-to-use and easy-to-set-up architecture to overcome shortcomings of
current approaches and to cover the needs of aid personnel.
Although being a niche, Virtual Reality (VR) environments can provide additional
value for visualizing specific content. Disciplines like earth sciences & geology, me-
chanical engineering, design, and architecture can benefit from VR environments. In
order to consider the variety of users, I introduce a more intuitive and user friendly
interaction metaphor, the ARC metaphor.
Visualization challenges base on being able to cope with more and more complex
datasets and to bridge the gap between comprehensibility and loss of information.
Furthermore the visualization approach has to be reasonable, which is a crucial
factor when working in interdisciplinary teams, where the standard of knowledge
is diverse. Users have to be able to conceive the visualized content in a fast and
reliable way. My contribution are visualization approaches in the field of supportive
visualization.
Finally, my work illuminates how the synthesis of visualization, interaction and dis-
play technologies enhance the user experience. I promote a holistic view. The user
is brought back into the focus of attention, provided with a tool-set to support him,
without overextending the abilities of, for example, non-expert users, a crucial factor
in the more and more interdisciplinary field of computer science.

This thesis is concerned with tropical moduli spaces, which are an important tool in tropical enumerative geometry. The main result is a construction of tropical moduli spaces of rational tropical covers of smooth tropical curves and of tropical lines in smooth tropical surfaces. The construction of a moduli space of tropical curves in a smooth tropical variety is reduced to the case of smooth fans. Furthermore, we point out relations to intersection theory on suitable moduli spaces on algebraic curves.

An extension of the finite element method–flux corrected transport stabilization (FEM-FCT) for hyperbolic problems in the context of partial differential-
algebraic equations (PDAEs) is proposed. Given a local extremum diminishing
property of the spatial discretization, the positivity preservation of the one-step
θ−scheme when applied to the time integration of the resulting differential-
algebraic equation (DAE) is shown, under a mild restriction on the time step-
size. As crucial tool in the analysis, the Drazin inverse and the corresponding
Drazin ODE are explicitly derived. Numerical results are presented for non-
constant and time-dependent boundary conditions in one space dimension and
for a two-dimensional advection problem where the advection proceeds skew to
the mesh.

The automatic analysis and retrieval of technical line drawings is hindered by many challenges such as: the large amount of contextual clutter around the symbols within the drawings, degradation, transformations on the symbols in drawings, large databases of drawings
and large alphabets of symbols. The core tasks required for the analysis of technical line
drawings are: symbol recognition, spotting and retrieval. The current systems for performing these tasks have poor performance due to the mentioned challenges. This dissertation
presents a number of methods that address these challenges. These methods achieve both
accurate and efficient symbol spotting and retrieval in technical line drawings, and perform
significantly better than state-of-the-art methods on the same problems. An overview of
the key contributions of this dissertation is given in the following.
First, this dissertation presents a geometric matching-based method for symbol recognition
and spotting. The method performs recognition in the presence of large amounts of contextual clutter, and provides precise localization of the recognized symbols. On standard
databases such as GREC-2005 and GREC-2011, the method achieves up to 10% higher
recall and up to 28% higher precision than state-of-the-art methods on the spotting task,
and achieves up to 7% higher recognition accuracy on the isolated recognition task. The
method is based on a geometric matching approach, which is flexible enough to incorporate
improvements on the matching strategy, feature types and information on the features. The
method also includes an adaptive preprocessing algorithm that deals with a wide variety
of noise types.
In order to improve the performance of the spotting method when dealing with degraded
drawings, two novel methods are presented in this dissertation. Both methods are based on
combining geometric matching with machine learning techniques. The geometric matching
is used to automatically generate training data that contain information on how well the
features of the queries are matched in both the true and the false matches found by the
spotting method. The first method learns the feature weights of the different query symbols
by linear discriminant analysis (LDA). The weighted query features are used in the spotting
method and result in 27% higher average precision than the original method, with a speedup
factor of 2. The second method uses SVM classification as a post-spotting step to distinguish
the true from the false matches in the spotting method. The use of the classification step
further improves the average precision of the spotting method by 20.6%.
This dissertation also presents methods for content analysis of line drawings. First, a
method for accurate and consistent detection (95.8%) of regions of interest (ROIs) is presented. The method is based on statistical feature grouping. The ROI-finding method is
identified as an important part of a symbol retrieval system: the better the detected ROIs,the higher the performance of a retrieval system. The ROI-finding method is also used to
improve the performance of the geometric-based spotting system.
Second, a symbol clustering method for building a compact and accurate representation of
a large database of technical drawings is presented. This method uses the output from the
ROI-finding method as input, and uses geometric matching as a similarity measure. The
method achieves high accuracy (90.1% recall, 94.3% precision) in forming clusters of symbols. The representatives of the clusters (34 symbols) are used as key entries to a symbol
index, which is identified as the outcome of an off-line stage of a symbol retrieval system.
Finally, an efficient and high performing large scale symbol retrieval system is presented
in this dissertation. The system follows the bag of visual words (BoVW) model, but with
using methods that are suitable to line drawings. The system uses the symbol index to
represent a database of drawings. During the on-line query retrieval stage, the query is
analyzed by the ROI-finding method, matched with the key entries of the symbol index via
geometric matching, and finally, a spatial verification step is performed on the retrieved
matches. The system achieves a query lookup time that is independent of the size of the
database, and is instead dependent on the size of the symbol index. The system achieves up
to 10% higher recall and up to 28% higher precision than state-of-the-art spotting systems
on similar databases.
Overall, these contributions are major advancements in the research of graphics recognition.
The hope is that, such contributions provide the basis for the development of reliable and
accurate performing applications for browsing, querying or classification of line drawings
for the benefit of end users.

Most innovation in the automotive industry is driven by embedded systems. They make usage of dynamic adaption to environmental changes or component/subsystem failures for remaining safe. Following this evolution, fault tree analysis techniques have been extended with concept for dynamic adaptation but resulting techniques like state event fault tree analysis, are not widely used in practice.
In this report we present the results of a controlled experiment that analyze these two techniques (State Events Fault Trees and Faul trees combined with markov chains) with regard to their applicability and efficiency in modeling dynamic behavior of dynamic embedded systems.
The experiment was conducted with students of the TU Kaiserslautern to modeli different safety aspects of an ambient assisted living system.
The main results of the experiment show that SEFTs where more easy and effective to use.

Most of the evolution in ambient assisted living is due to embedded
systems that dynamically adapt themself to react to environmental
changes or component/subsystem failures to maintain a certain level of
safety. Following this evolution fault tree analysis techniques have been
extended with concept for dynamic adaptation but resulting techniques
such as dynamic fault trees or state event fault trees analysis are not
widely used as expected.
In this report we describe a controlled experiment to analyze these two
techniques with regard to their applicability and efficiency in modeling
dynamic behavior of ambient assisted living systems.
Results of the experiment show that Dynamic Fault Trees are easier and more effective
to use, although they produce better results (models) with State Events Fault Trees.

Cancer cell migration is an essential feature in the process of tumor spread and establishing of metastasis. It characterizes the invasion observed on the level of the cell population, but it is also tightly connected to the events taking place on the subcellular level. These are conditioning the motile and proliferative behavior of the cells, but are also influenced by it. In this work we propose a multiscale model linking these two levels and aiming to assess their interdependence. On the subcellular, microscopic scale it accounts for integrin binding to soluble and insoluble components present in the peritumoral environment, which is seen as the onset of biochemical events leading to changes in the cell's ability to contract and modify its shape. On the macroscale of the cell population this leads to modifications in the diffusion and haptotaxis performed by the tumor cells and implicitly to changes in the tumor environment. We prove the (local) well posedness of our model and perform numerical simulations in order to illustrate the model predictions.

This thesis reports on investigations on the structure and reactivity of dipeptide-alkali metal complexes, a series of ruthenium bearing catalysts, dysprosium based single molecule magnets and organometallic di-cobalt complexes. A variety of experimental and theoretical methods was used dependent on the problem: collision induced dissociation, hydrogen/deuterium exchange reactions, gas phase reactions with \(D_2\), infrared multiple-photon dissociation and the determination of minimum energy structures, IR absorption spectra, transition states and electronic transitions based on density functional theory.
A case study was carried out to explore the influence of alkali metal ions on the gas phase structure of the dipeptide Carnosine. CID experiments on protonated Carnosine and its alkali metal complexes in an ion trap resulted in different fragment pathways dependent on the size of the alkali metal. The complexation of small ions (\(Li^+\) and \(Na^+\)) promoted the cleavage of bonds in the molecules backbone under CID, while \(Rb^+\)- and \(Cs^+\)-Carnosine complexes underwent the exclusive loss of the alkali metal. CID breakdown curves reflected the different binding behavior of the alkali ions to Carnosine. Gas phase H/D exchange reactions with \(D_2O\) resulted in the exchange of several protons of the protonated dipeptide, while its alkali metal complexes underwent no exchange reactions. DFT derived energetical minimum isomers exhibited only charge solvated tridentate structures, whereas salt bridge as well as charge solvated binding motives are reported in literature on complexes of alkali metal ions and oligopeptides. This study was published in a similar version as a paper in Zeitschrift für Physikalische Chemie.
A combination of the four dipeptides Carnosine, Anserine, GlyHis and HisGly with alkali metal ions was investigated with the help of CID, IR-MPD spectroscopy and H/D exchange reactions with \(ND_3\). The aim of the survey was to elucidate the influence of the methyl-group at the histidine ring, of the peptide sequence and chain length on the binding motives of the alkali ions. The experimental results were compared to DFT derived minimum energetical isomers. A moderate accordance was found for DFT predicted IR absorptions to IR-MPD spectra. A systematic nomenclature was developed reflecting all binding motives of the four dipeptides to alkali ions. Carnosine complexes all alkali metal ions in an uniform motive. DFT derived energetical minimum isomers of the three other dipeptides showed strong conformational changes with increasing size of the alkali ion. The most favored binding motive of all peptides was the tridentate complexation of the alkali ion by a carboxylic and an amidic oxygen atom, while the electron donating nitrogen atom either belongs to the Histidine ring or the amine group. The ability to form hydrogen bonds in a certain binding motive is essential for the preference of the Histidine or amine nitrogen atom as an electron donor. The charge solvated binding motive is the most common within all found isomers. Several structures exhibited hydrogen bonded protons. Those can be interpretated as intermediates between the charge solvated and the salt bridge binding motive. CID breakdown curves of the cationic complexes of the dipeptides with \(K^+\), \(Rb^+\) and \(Cs^+\) resulted in a fair agreement of \(E^{50\%}_{com}\) values with DFT derived Gibbs free binding energies. CID led to multiple fragments of the \(Li^+\) and \(Na^+\) dipeptide complexes and to an insufficient correlation between the \(E^{50\%}_{com}\) values and metal-dipeptide free binding enthalpies. Gas phase H/D exchange reactions of the protonated dipeptides with \(ND_3\) resulted in the exchange of all labile protons with comparable relative partial rate constants. The assumption of coexisting single and double exchange reactions per single collision led to an enhancement in quality of the pseudo first order kinetic fits of the experimental derived data. The \(Li^+\), \(Na^+\) and \(K^+\) complexes of the dipeptides exhibited a reduction in the number of exchanged protons, significantly lower rate constants for H/D exchange and only single exchange reactions.
The complexation of the doubly charged cationic transition metal \(Zn^{2+}\) by deprotonated Carnosine led to crucial conformational changes with respect to the alkali metal complexes. Former DFT calculations on the gas phase structure of \([Carn-H,Zn^{II}]^+\) were now compared to IR-MPD spectra. IR-MPD spectra exhibited several of the DFT predicted IR absorptions while the overall agreement in the position of bands is only partially satisfactory. The complex \([Carn-H,Zn^{II}]^+\) was furthermore used in order to study the band dependent enhancement of fragmentation efficiency by application of a resonant 2-color IR-MPD pump/probe scheme. In literature, it is assumed that the slopes of linear fits to the log-log scale of experimental data (fragmentation efficiency vs. laser pulse energy) correlate to the number of photons needed for fragmentation. No reasonable number of photons for the fragmentation of the molecule was derived with this approach. However, it could be shown that the number of photons of the pump laser needed for fragmentation is reduced by the use of a second IR color. The change of the delay between the pump and probe laser pulse had an influence on the shape of the absorption bands. Irradiation with the probe laser pulse before the pump laser caused a heating of the molecule which resulted in a broadening of bands. No broadening was observed when the probe laser was applied simultaneously or after the pump laser. CID and IR-MPD fragmentation channels differed in their relative abundance. Furthermore, relative abundancies of fragments were specific to the excited vibrational motions. This study provides essential approaches for the further study of the mechanism of resonant 2-color IR-MPD spectroscopy.
Several ruthenium catalysts for transfer hydrogenation reactions were synthesized by L. Ghoochany (research group W. Thiel, TU Kaiserlautern). CID measurements on isotopic labeled species led to the following conclusion about the activation process of the catalyst: a nitrogen-ruthenium bond is broken, the pyrimidine ring of the substituted 2-R-4-(2-pyridinyl)pyrimidine ligand rotates about 160° and a carbon-ruthenium bond is formed under subsequent loss of a HCl (or DCl) molecule. The mass spectrometers CID amplitude was calibrated with a set of “thermometer ions”. CID breakdown curves were used for determination of \(E^{50\%}_{com}\) values of three differently substituted catalysts. Finally, activation energies were estimated by means of the calibration. The resulting activation energies showed a qualitative correlation to DFT derived activation energies. These results are part of a manuscript which was submitted to Chemistry – A European Journal and is currently in the review process. Further studies on this series of transition metal complexes included CID on ligand exchanged species, 1- and 2-color IR-MPD spectroscopy, gas phase reactions with \(D_2\) and DFT based modeling of the reaction coordinate of the \(D_2\) insertion. The exchange of the anionic chlorido ligand in solution led to three complexes with different fragmentation thresholds. CID derived activation amplitudes correspond well to the order predicted by the hard/soft acids/bases (HSAB) concept. 1-color IR-MPD experiments on two complexes showed only a few bands. Resonant 2-color IR-MPD increased the overall fragmentation efficiency and uncovered several dark bands. DFT derived IR absorption spectra correlate well to IR-MPD spectra while some bands are still not observable. Gas phase reactions with \(D_2\) showed an increase of the mass of the activated complex of +4 m/z. This was interpreted in terms of an incorporation of a \(D_2\) molecule under heterolytical cleavage of the \(D_2\) molecule and can be compared to a back reaction of the activation. The reaction coordinate of the \(D_2\) incorporation was modeled with DFT at the B3LYP/cc-pVTZ level of theory and different activation energies were derived dependent on the substituent. Reactions of three differently substituted complexes with \(D_2\) resulted in different relative partial rate constants. The comparison to rate constants derived from transition state theory showed a qualitative but not quantitative correlation to the experimental results. This study contributes to our ongoing work on the assignment and isolation of reaction intermediates in the gas phase.
A series of dysprosium based complexes was synthesized by A. Bhunia (research group P. W. Roesky, KIT) and studied within the collaborative research center SFB/TRR 88 “3MET”. We contributed to this work with ESI-MS, CID and experiments on H/D exchange reactions with \(ND_3\) in the gas phase. Those complexes consist of a central triple-charged dysprosium cation and two identical salen-type ligands which allow for a complexation of up to two transition metals. The monometallic dysprosium complex shows single molecule magnet (SMM) behavior in SQUID measurements, while the incorporation of two double-charged manganese cations leads to ferromagnetic behavior. The interaction of terminal amine groups with the manganese ions caused a hinderance of the exchange H/D exchange reaction with \(ND_3\) in the gas phase. Alternatively, the terminal amine groups of the monometallic dysprosium complex allow for the bond of two \(Ni^{2+}(tren)\) complexes. ESI-MS studies showed anionic as well as cationic complexes due to deprotonation or protonation in solution. CID studies led to fragmentation schemes which correlate quite well to the predicted structures of the complexes. These results are part of two publications in Inorganic Chemistry and Dalton Transactions. Further studies on this series of mono-, di- and trimetallic complexes are reported in this thesis. H/D exchange reactions with \(D_2O\) in solution yielded in an exchange of all labile protons for the cationic complexes. Anionic complexes underwent a partial or a complete exchange of labile protons. A comparison of 1- and 2-color IR-MPD spectra of anionic and cationic complexes as well as H/D exchanged species allowed for the assignment of vibrational bands. Furthermore, preferred protonation sites were derived by comparing the results of IR-MPD experiments and H/D exchange reactions in solution and in the gas phase. This study contributes to our ongoing work on the determination of magnetic properties of isolated ions in the gas phase at the Helmholtz-Zentrum Berlin.
The complex \([(^4CpCo)_2(\mu-C_2Ph_2)]\) (\(^4Cp\) = tetraisopropyl-cyclopentadiene) was synthesized by J. Becker (research group H. Sitzmann, TU Kaiserslautern). The cationic complex and several reaction products were characterized by ESI-MS. Some of the experimental data contributed to the diploma thesis of J. Becker. The cationic reaction products and the complex itself were subject of IR spectroscopic characterization. IR-MPD efficiency changed crucially with modification of the complex, yielding \([(^4CpCo)_2(\mu-C_2Ph_2)X]^+ (X=H, (H+CH_3CN), Cl, O)\). The contribution of various fragmentation channels to the overall fragmentation efficiency was studied in detail. An increase of photon flux resulted in a saturation of preferred \(C_2Ph_2\) loss, additional alkyl fragments out of the \(^4Cp\) rings arising. Several absorption bands were found in the mid- and near-IR region. A model system from literature was used to identify seemingly levels of DFT theory by reference to X-ray crystal structure data. The B3LYP and the B97D functional with cc-pVDZ and Stuttgart 1997 ECP basis sets were identified for calculations of the complex \([(^4CpCo)_2(\mu-C_2Ph_2)]^+\) and of its reaction products. An elongation of the Co-Co bond distance was observed for the cationic reaction products with \(Cl^-\) and \(O^{2-}\). Calculations with B3LYP and B97D resulted in different electronic ground states. We did not obtain a good agreement of calculated vibrational modes and recorded IR-MPD spectra. DFT predicted more absorption bands than observed, especially those corresponding to aliphatic symmetric \(CH_n (n=2, 3)\) and aromatic CH stretch motions. Future 2-color IR-MPD experiments might resolve currently prevailing discrepancies. TD-DFT calculations yielded several electronic transitions that do not correspond to the IR-MPD spectra. The chosen levels of theory for DFT and TD-DFT calculations does not seem to be appropriate. IR-MPD spectra have to be remeasured in order to normalize the spectra to photon flux. Furthermore, a different strategy has to be developed for ab initio calculations on the complexes under study.
A combination of various methods applied to isolated ions in the gas phase and in solution allowed for the study of their structure, binding energies and reactivity. 1- and 2-color IR-MPD spectroscopy combined with DFT predicted absorption spectra of different isomers enabled an assignment of vibrational bands and binding motives of the molecules. The derived results are important for further studies on the binding behavior of peptides and the reaction behavior of metal complexes.

There is growing international concern about the necessity to re-think the university so that it might remain relevant in a modern society. In the traditional task division at universities, knowledge is the main resource. Universities make use of both the cognitive and the informational approach. It was expected that universities use each approach to improve overall university performance. To effectively use the informational approach, universities should apply the tools from knowledge management. To effectively use the cognitive approach, universities must update their teaching-learning strategies to incorporate some of the recent advances in neuroscience and biology of knowledge, specifically from neurobiology and autopoiesis. With this frame, the main contribution of this work is the result of merging pedagogy and biology, towards an ideal future university. This goal was achieved through an exploratory study conducted to identify opportunities and difficulties in improving the teaching-learning process for the future of higher education in Honduras. The Delphi Study was used as a predictive method. Nineteen Honduran experts participated in this study, and two rounds were necessary to achieve consensus.
The multi-disciplinary approach of this research addresses three different fields whose core element is knowledge. First, input from the present field of higher education is used to speak about the future. Second, input is taken from the biology of knowledge, and its contributions from neurobiology and autopoiesis that allow modifying and completing the already existing learning theories with a biological basis. Third, input is taken from the knowledge process, which is traditionally used as an organizational tool and know is translated to the individual level. The exploration shows that experts are concerned about all the missions and responsibilities of universities, but they agree that changes should primarily take place in the teaching dimension. Even though they are not aware of the possible contributions of biology, they suggest new forms of teaching that more favor skills development, promotes values, pertinent knowledge, and personal development over short-term contents. The resulting BRAIN Model encompasses the ideal future of higher education regarding teaching and learning, according to experts’ answers. It provides a useful guide that any reform in teaching should take into account for a holistic, integral, and therefore more efficient learning task.

The aim is to prove global existence and uniqueness of square integrable solutions to a class of multiscale models for tumour
cell migration involving chemotaxis, haptotaxis, and subcellular dynamics. This approach allows the tissue
fibre and cell densities as well as concentrations of chemotactic signals to be less regular and the conditions sufficient for well-posedness of the multiscale model to be less restrictive than in previous settings.

Polychlorinated dibenzo-p-dioxins, dibenzofurans, and polychlorinated biphenyls are persistent environmental pollutants which ubiquitously occur as complex mixtures and accumulate in the food and feed chain due to their high lipophilic properties. Of the 419 possible congeners, only 29 share a common mechanism of action and cause similar effects, the so called dioxin-like compounds. Dioxin-like compounds evoke a broad spectrum of biochemical and toxic responses, i.e. enzyme induction, dermal toxicity, hepatotoxicity, immunotoxicity, carcinogenicity as well as adverse effects on reproduction, development, and the endocrine system in laboratory animals and in humans. Most, if not all, of the aforementioned responses, are mediated by the aryl hydrocarbon receptor. In the present work, the elicited biochemical effects of a selection of dioxin-like compounds and the non dioxin-like PCB 153 were examined in mouse (in vivo) and in human liver cell models (in vitro). Emphasis was given to the main contributors to the total toxic equivalents in human blood and tissues TCDD, 1-PnCDD, 4-PnCDF, PCB 118, PCB 126, and PCB 156, which likewise contribute about 90 % to the dioxin-like activity in the human food chain.
Three mouse in vivo studies were carried out aiming to characterize the alterations in hepatic gene expression as well as the induction of hepatic xenobiotic metabolizing enzymes after single oral dose. Based on the results obtained from mouse 3-day and 14-day studies, the seven test compounds can be categorized into three classes; the ones which are 'pure' AhR ligands (TCDD, 1-PnCDD, 4-PnCDF, and PCB 126) or solely CAR inducers (PCB 153), and the ones which are AhR/CAR mixed-type inducers (PCB 118, PCB 156). Moreover, the analysis of hepatic gene expression patterns after a single oral dose of either TCDD or PCB 153 revealed that the altered genes fundamentally differed. Profiling of significantly altered genes led to the conclusion that changes in gene expression were associated with different signalling pathways, in fact by AhR and CAR.
For investigating the role of the AhR in mediating biological responses, several experimental approaches were carried out, such as the analysis of blood plasma metabolites in Ahr knockout and wild-type mice. Genotype specifics and similarities were determined by HPLC-MS/MS analysis. Several plasma metabolites could be identified in both genotypes, but also differences were detected. Furthermore, an in vivo experiment was performed aiming to characterize AhR-dependent and -independent effects in female Ahr knockout and wild-type mice. For this purpose, mice received a single oral dose of TCDD and were killed 96 h later. Microarray analysis of mouse livers revealed that although the Ahr gene was knocked out in Ahr-/- mice, the quantity of affected genes were in the same order of magnitude as for Ahr+/+ mice, but the pattern of altered genes distinctly differed. In addition, the relative liver weights of TCDD-treated Ahr+/+ mice were significantly increased which led to the conclusion, that TCDD induced the development of hepatic steatosis in female Ahr wild-type.
The performed in vitro experiments aimed to characterize the effects elicited by selected DLCs and PCB 153 in human liver cell models by the use of HepG2 cells and primary human hepatocytes. In general, primary human hepatocytes were less responsive than HepG2 cells. This was not only observed in EC values derived from EROD assay, but also regarding microarray analysis in terms of differently regulated genes. In vitro REPs gained from both liver cell models widely confirmed the current TEFs, but some deviations occurred. The comparison of the TCDD-altered genes in both human cell types revealed that only a considerably small number of genes was in common up regulated by both human liver cell models, such as the established AhR-regulated highly inducible cytochrome P450s 1A1, 1A2, and 1B1 as well as other AhR target genes. Although the overlap was rather small, the TCDD-induced genes could be consistently associated with the broad spectrum of established dioxin-related biological responses. The gene expression pattern in primary human hepatocytes after treatment with selected DLCs (TCDD, 1-PnCDD, 4-PnCDF, and PCB 126) and PCB 153 was additionally characterized by microarray analysis. The highest response in terms of significantly altered genes was determined for TCDD, followed by 4-PnCDF, 1-PnCDD, and PCB 126, whereas exposure to PCB 153 did not evoke any significant changes in gene expression. The pattern of significantly altered genes was very homogenous among the four congeners. Genes associated with well-established DLC-related biological responses as well as novel dioxin-inducible target genes were identified, whereby an extensive overlap in terms of up regulated genes by all four DLCs occurred. In conclusion, the results from the in vitro experiments performed in primary human hepatocytes provided fundamental insight into the congeners' potencies and caused alterations in gene expression patterns. The obtained findings implicate that although the extent of enzyme inducibilities varied, the gene expression patterns are coincidental. Microarray analysis identified species-specific (mouse vs. human) as well as model-specific (in vitro vs. in vivo and transformed cells vs. untransformed cells) differences. In order to identify novel biomarkers for AhR activation due to treatment with dioxin-like compounds, five candidates were selected based on the microarray results i.e. ALDH3A1, TIPARP, HSD17B2, CD36, and AhRR. Eventually, ALDH3A1 turned out to be the most reliable and suitable marker for exposure to DLCs in both human liver cell models eliciting the highest mRNA inducibility among the five chosen candidates. In which way these species- and cell type-specific markers are involved in the dioxin-elicited toxic responses should be further characterized in vivo and in vitro.

Tire-soil interaction is important for the performance of off-road vehicles and the soil compaction in the agricultural field. With an analytical model, which is integrated in multibody-simulation software, and a Finite Element model, the forces and moments generated on the tire-soil contact patch were studied to analyze the tire performance. Simulations with these two models for different tire operating conditions were performed to evaluate the mechanical behaviors of an excavator tire. For the FE model validation a single wheel tester connected to an excavator arm was designed. Field tests were carried out to examine the tire vertical stiffness, the contact pressure on the tire – hard ground interface, the longitudinal/vertical force and the compaction of the sandy clay from the test field under specified operating conditions. The simulation and experimental results were compared to evaluate the model quality. The Magic Formula was used to fit the curves of longitudinal and lateral forces. A simplified tire-soil interaction model based on the fitted Magic Formula could be established and further applied to the simulation of vehicle-soil interaction.

Factorization of multivariate polynomials is a cornerstone of many applications in computer algebra. To compute it, one uses an algorithm by Zassenhaus who used it in 1969 to factorize univariate polynomials over \(\mathbb{Z}\). Later Musser generalized it to the multivariate case. Subsequently, the algorithm was refined and improved.
In this work every step of the algorithm is described as well as the problems that arise in these steps.
In doing so, we restrict to the coefficient domains \(\mathbb{F}_{q}\), \(\mathbb{Z}\), and \(\mathbb{Q}(\alpha)\) while focussing on a fast implementation. The author has implemented almost all algorithms mentioned in this work in the C++ library factory which is part of the computer algebra system Singular.
Besides, a new bound on the coefficients of a factor of a multivariate polynomial over \(\mathbb{Q}(\alpha)\) is proven which does not require \(\alpha\) to be an algebraic integer. This bound is used to compute Hensel lifting and recombination of factors in a modular fashion. Furthermore, several sub-steps are improved.
Finally, an overview on the capability of the implementation is given which includes benchmark examples as well as random generated input which is supposed to give an impression of the average performance.

This thesis is concerned with a phase field model for brittle fracture.
The high potential of phase field modeling in computational fracture mechanics lies in the generality of the approach and the straightforward numerical implementation, combined with a good accuracy of the results in the sense of continuum fracture mechanics.
However, despite the convenient numerical application of phase field fracture models, a detailed understanding of the physical properties is crucial for a correct interpretation of the numerical results. Therefore, the driving mechanisms of crack propagation and nucleation in the proposed phase field fracture model are explored by a thorough numerical and analytical investigation in this work.

Many real life problems have multiple spatial scales. In addition to the multiscale nature one has to take uncertainty into account. In this work we consider multiscale problems with stochastic coefficients.
We combine multiscale methods, e.g., mixed multiscale finite elements or homogenization, which are used for deterministic problems with stochastic methods, such as multi-level Monte Carlo or polynomial chaos methods.
The work is divided into three parts.
In the first two parts we study homogenization with different stochastic methods. Therefore we consider elliptic stationary diffusion equations with stochastic coefficients.
The last part is devoted to the study of mixed multiscale finite elements in combination with multi-level Monte Carlo methods. In the third part we consider multi-phase flow and transport equations.

Constructing accurate earth models from seismic data is a challenging task. Traditional methods rely on ray based approximations of the wave equation and reach their limit in geologically complex areas. Full waveform inversion (FWI) on the other side seeks to minimize the misﬁt between modeled and observed data without such approximation.
While superior in accuracy, FWI uses a gradient based iterative scheme that makes it also very computationally expensive. In this thesis we analyse and test an Alternating Direction Implicit (ADI) scheme in order to reduce the costs of the two dimensional time domain algorithm for solving the acoustic wave equation. The ADI scheme can be seen as an intermediate between explicit and implicit ﬁnite diﬀerence modeling schemes. Compared to full implicit schemes the ADI scheme only requires the solution of much smaller matrices and is thus less computationally demanding. Using ADI we can handle coarser discretization compared to an explicit method. Although order of convergence and CFL conditions for the examined explicit method and ADI scheme are comparable, we observe that the ADI scheme is less prone to dispersion. Furhter, our algorithm is eﬃciently parallelized with vectorization and threading techniques. In a numerical comparison, we can demonstrate a runtime advantage of the ADI scheme over an explicit method of the same accuracy.
With the modeling in place, we test and compare several inverse schemes in the second part of the thesis. With the goal of avoiding local minima and improving speed of convergence, we use diﬀerent minimization functions and hierarchical approaches. In several tests, we demonstrate superior results of the L1 norm compared to the L2 norm – especially in the presence of noise. Furthermore we show positive eﬀects for applying three diﬀerent multiscale approaches to the inverse problem. These methods focus on low frequency, early recording, or far oﬀset during early iterations of the minimization and then proceed iteratively towards the full problem. We achieve best results with the frequency based multiscale scheme, for which we also provide a heuristical method of choosing iteratively increasing frequency bands.
Finally, we demonstrate the eﬀectiveness of the diﬀerent methods ﬁrst on the Marmousi model and then on an extract of the 2004 BP model, where we are able to recover both high contrast top salt structures and lower contrast inclusions accurately.

Non-smooth contact dynamics provides an increasingly popular simulation framework for granular material. In contrast to classical discrete element methods, this approach is stable for arbitrary time steps and produces visually acceptable results in very short computing time. Yet when it comes to the prediction of draft forces, non-smooth contact dynamics is typically not accurate enough. We therefore propose to combine the method class with an interior point algorithm for higher accuracy. Our specific algorithm is based on so-called Jordan algebras and exploits the relation to symmetric cones in order to tackle the conical constraints that are intrinsic to frictional contact problems. In every interior point iteration a linear system has to be solved. We analyze how the interior point method behaves when it is combined with Krylov subspace solvers and incomplete factorizations. We show that efficient preconditioners and efficient linear solvers are essential for the method to be applicable to large-scale problems. Using BiCGstab as a linear solver and incomplete Cholesky factorizations, we substantially improve the accuracy in comparison to the projected Gauss-Jacobi solver.

In this paper we introduce a binary autoregressive model. In contrast to the typical autoregression framework, we allow the conditional distribution of the observed process to depend on past values of the time series and some exogenous variables. Such processes have
potential applications in econometrics, medicine and environmental sciences. In this
paper, we establish stationarity and geometric ergodicity of these
processes under suitable conditions on the parameters of the model. Such properties are
important for understanding the stability properties of the model as well as for deriving the
asymptotic behavior of the parameter estimators.

Data integration aims at providing uniform access to heterogeneous data, managed by distributed source systems. Data sources can range from legacy systems, databases, and enterprise applications to web-scale data management systems. The materialized approach to data integration, extracts data from the sources, transforms and consolidates the data, and loads it into an integration system, where it is persistently stored and can be queried and analyzed.
To support materialized data integration, so called Extract-Transform-Load (ETL) systems have been built and are widely used to populate data warehouses today. While ETL is considered state-of-the-art in enterprise data warehousing, a new paradigm known as MapReduce has recently gained popularity for web-scale data transformations, such as web indexing or page rank computation.
The input data of both, ETL and MapReduce programs keeps changing over time, while business transactions are processed or the web is crawled, for instance. Hence, the results of ETL and MapReduce programs get stale and need to be recomputed from time to time. Recurrent computations over changing input data can be performed in two ways. The result may either be recomputed from scratch or recomputed in an incremental fashion. The idea behind the latter approach is to update the existing result in response to incremental changes in the input data. This is typically more efficient than the full recomputation approach, because reprocessing unchanged portions of the input data can often be avoided.
Incremental recomputation techniques have been studied by the database research community mainly in the context of the maintenance of materialized views and have been adopted by all major commercial database systems today. However, neither today's ETL tools nor MapReduce support incremental recomputation techniques. The situation of ETL and MapReduce programmers nowadays is thus much comparable to the situation of database programmers in the early 1990s. This thesis makes an effort to transfer incremental recomputation techniques into the ETL and MapReduce environments. This poses interesting research challenges, because these environments differ fundamentally from the relational world with regard to query and programming models, change data capture, transactional guarantees and consistency models. However, as this thesis will show, incremental recomputations are feasible in ETL and MapReduce and may lead to considerable efficiency improvements.

The use of trading stops is a common practice in financial markets for a variety of reasons: it provides a simple way to control losses on a given trade, while also ensuring that profit-taking is not deferred indefinitely; and it allows opportunities to consider reallocating resources to other investments. In this thesis, it is explained why the use of stops may be desirable in certain cases.
This is done by proposing a simple objective to be optimized. Some simple and commonly-used rules for the placing and use of stops are investigated; consisting of fixed or moving barriers, with fixed transaction costs. It is shown how to identify optimal levels at which to set stops, and the performances of different rules and strategies are compared. Thereby, uncertainty and altering of the drift parameter of the investment are incorporated.

The only quadrature operator of order two on \(L_2 (\mathbb{R}^2)\) which covaries with orthogonal
transforms, in particular rotations is (up to the sign) the Riesz transform. This property
was used for the construction of monogenic wavelets and curvelets. Recently, shearlets
were applied for various signal processing tasks. Unfortunately, the Riesz transform does
not correspond with the shear operation. In this paper we propose a novel quadrature operator called linearized Riesz transform which is related to the shear operator. We prove
properties of this transform and analyze its performance versus the usual Riesz transform numerically. Furthermore, we demonstrate the relation between the corresponding
optical filters. Based on the linearized Riesz transform we introduce finite discrete quasi-monogenic shearlets and prove that they form a tight frame. Numerical experiments show
the good fit of the directional information given by the shearlets and the orientation ob-
tained from the quasi-monogenic shearlet coefficients. Finally we provide experiments on
the directional analysis of textures using our quasi-monogenic shearlets.

Aim of this work was the extension and development of a coupled Computational Fluid Dynamics (CFD) and population balance model (PBM) solver to enable a simulation aided design of stirred liquid-liquid extraction columns. The principle idea is to develop a new design methodology based on a CFD-PBM approach and verify it with existing data and correlations. On this basis, the separation performance in any apparatus geometry should be possible to predict without any experimental input. Reliable “experiments in silico” (computer calculations) should give the engineer a valuable and user-friendly tool for early design studies at minimal costs.
The layout of extraction columns is currently based on experimental investigations from miniplant to pilot plant and a scale-up to the industrial scale. The hydrodynamic properties can be varied by geometrical adjustments of the stirrer diameter, the stirrer height, the free cross sectional area of the stator, the compartment height as well as the positioning and the size of additional baffles. The key parameter for the liquid–liquid extraction is the yield which is mainly determined at the in- and outlets of the column. Local phenomena as the swirl structure are influenced by geometry changes. However, these local phenomena are generally neglected in state-of-the are design methodologies due to the complex required measurement techniques. A geometrical optimization of the column therefore still results in costs for validation experiments as assembly and operation of the column, which can be reduced by numerical investigations. The still mainly in academics used simulation based layout of counter-current extraction columns is based at the beginning of this work on one dimensional simulations of extraction columns and first three dimensional simulations. The one dimensional simulations are based on experimental derived, geometrical dependent correlations for the axial backmixing (axial dispersion), the hold-up, the phase fraction, the droplet sedimentation and the energy dissipation. A combination of these models with droplet population balance modeling resulted in a description of the complex droplet-droplet interactions (droplet size) along the column height. The three dimensional CFD simulations give local information about the flow field (velocity, swirl structure) based on the used numerical mesh corresponding to the real geometry. A coupling of CFD with population balance modeling further provides information about the local droplet size. A backcoupling of the droplet size with the CFD (drag model) results in an enhancement of the local hydrodynamics (e.g. hold-up, dispersed phase velocity). CFD provided local information about the axial dispersion coefficient of simple geometrical design (e.g. Rotating Disc Contactor (RDC) column). First simulations of the RDC column using a two dimensional rotational geometry combined with population balance modeling were performed and gave local information about the droplet size for different boundary conditions (rotational speed, different column sizes).
In this work, two different column types were simulated using an extended OpenSource CFD code. The first was the RDC column, which were mainly used for code development due to its simple geometry. The Kühni DN32 column is equipped with a six-baffled stirring device and flat baffles for disturbing the flow and requires a full three dimensional description. This column type was mainly used for experimental validation of the simulations due to the low required volumetric flow rate. The Kühni DN60 column is similar to the Kühni DN32 column with slight changes to the stirring device (4-baffles) and was used for scale up investigations. For the experimental validation of the hydrodynamics, laser based measurement techniques as Particle Image Velocimetry (PIV) and Laser Induced Fluorescence (LIF) were used. A good agreement between the experimental derived values for velocity, hold-up and energy dissipation, experimentally derived correlations from literature and the simulations with a modified Euler-Euler based OpenSource CFD code could be found. The experimental derived axial dispersion coefficient was further compared to Euler-Lagrange simulations. The experimental derived correlations for the Kühni DN32 in literature fit to the simulated values. Also the axial dispersion coefficient for the dispersed phase satisfied a correlation from literature. However, due to the complexity of the dispersed phase axial dispersion coefficient measurement, the available correlations gave no distinct agreement to each other.
A coupling of the modified Euler-Euler OpenSource CFD code was done with a one group population balance model. The implementation was validated to the analytical solution of the population balance equation for constant breakage and coalescence kernels. A further validation of the population balance transport equation was done by comparing the results of a five compartment section to the results of the commercial CFD code FLUENT using the Quadrature Method of Moments (QMOM).
For the simulation of the droplet-droplet interactions in liquid-liquid extraction columns, several breakage and coalescence models are available in the literature. The models were compared to each other using the one-group population balance model in Matlab which allows the determination of the minimum stable droplet diameter at a certain energy dissipation. Based on this representation, it was possible to determine the parameters for a specific breakage and coalescence model combination which allowed the simulation of a Kühni miniplant column at different rotational speeds. The resulting simulated droplet size was in very good agreement to the experimental derived droplet size from literature. Several column designs of the DN32 were investigated by changing the compartment height and the axial stirrer position. It could be shown that a decrease of the stirrer position increases the phase fraction inside the compartment. At the same time, the droplet size decreases inside the compartment, which allows a higher mass transfer due to a higher available interfacial area. However, the shifting results in an expected earlier flooding of the column due to a compressed flow structure underneath the stirring device. In a next step, the code was further extended by mass transfer equations based on the two-film theory. Mass transfer coefficient models for the dispersed and continuous phase were investigated for the RDC column design.
A first mass transfer simulation of a full miniplant column was done. The change in concentration was accounted by the mixture density, viscosity and interfacial tension in dependence of the concentration, which affects the calculation of the droplet size. The results of the column simulation were compared to own experimental data of the column. It could be shown that the concentration profile along the column height can be predicted by the presented CFD/population balance/mass transfer code. The droplet size decreases corresponding to the interfacial tension along the column height. Compared to the experimental derived droplet size at the outlet, the simulation is in good agreement.
Besides the occurrence of a mono dispersed droplet size, high breakage may lead to the generation of small satellite droplets and coalescence underneath the stator leads to larger droplets inside the column and hence to a change of the hold-up and of the flooding point. A multi-phase code was extended by the Sectional Quadrature Method of Moment (SQMOM) allowing a modeling of the droplet interactions of bimodal droplet interactions or multimodal distributions. The implementations were in good agreement to the analytical solution. In addition, the simulation of an RDC column section showed the different distribution of the smaller droplets and larger droplets. The smaller droplets tend to follow the continuous phase flow structure and show a higher distribution of inside the column. The larger droplets tend to rise directly through the column and show only a low influence to the continuous phase flow.
The current results strengthen the use of CFD for the layout of liquid-liquid extraction columns in future. The coupling of CFD/PBM and mass transfer using an OpenSource CFD code allows the investigation of computational intensive column designs (e.g. pilot plant columns). Furthermore the coupled code enhances the accuracy of the hydrodynamics simulations and leads to a better understanding of counter-current liquid-liquid extraction columns. The gained correlation were finally used as an input for one dimensional mass transfer simulations, where a perfect fit of the concentration profiles at varied boundary conditions could be obtained. By using the multi-scale approach, the computational time for mass transfer simulations could be reduced to minutes. In future, with increasing computational power, a further extend of the multiphase CFD/SQMOM model including mass transfer equation will provide an efficient tool to model multimodal and multivariate systems as bubble column reactors.

Annual Report 2012
(2013)

Annual Report, Jahrbuch AG Magnetismus

In the last few years a lot of work has been done in the investigation of Brownian motion with point interaction(s) in one and higher dimensions. Roughly speaking a Brownian motion with point interaction is nothing else than a Brownian motion whose generator is disturbed by a measure supported in just one point.
The purpose of the present work is the introducing of curve interactions of the two dimensional Brownian motion for a closed curve \(\mathcal{C}\). We will understand a curve interaction as a self-adjoint extension of the restriction of the Laplacian to the set of infinitely often continuously differentiable functions with compact support in \(\mathbb{R}^{2}\) which are constantly 0 at the closed curve. We will give a full description of all these self-adjoint extensions.
In the second chapter we will prove a generalization of Tanaka's formula to \(\mathbb{R}^{2}\). We define \(g\) to be a so-called harmonic single layer with continuous layer function \(\eta\) in \(\mathbb{R}^{2}\). For such a function \(g\) we prove
\begin{align}
g\left(B_{t}\right)=g\left(B_{0}\right)+\int\limits_{0}^{t}{\nabla g\left(B_{s}\right)\mathrm{d}B_{s}}+\int\limits_{0}^{t}\eta\left(B_{s}\right)\mathrm{d}L\left(s,\mathcal{C}\right)
\end{align}
where \(B_{t}\) is just the usual Brownian motion in \(\mathbb{R}^{2}\) and \(L\left(t,\mathcal{C}\right)\) is the connected unique local time process of \(B_{t}\) on the closed curve \(\mathcal{C}\).
We will use the generalized Tanaka formula in the following chapter to construct classes of processes related to curve interactions. In a first step we get the generalization of point interactions in a second step we get processes which behaves like a Brownian motion in the complement of \(\mathcal{C}\) and has an additional movement along the curve in the time- scale of \(L\left(t,\mathcal{C}\right)\). Such processes do not exist in the one point case since there we cannot move when the Brownian motion is in the point.
By establishing an approximation of a curve interaction by operators of the form Laplacian \(+V_{n}\) with "nice" potentials \(V_{n}\) we are able to deduce the existence of superprocesses related to curve interactions.
The last step is to give an approximation of these superprocesses by a sytem of branching particles. This approximation gives a better understanding of the related mass creation.

Chisel (Constructing Hardware in a Scala embedded language) is a new programming language, which embedded in Scala, used for hardware synthesis. It aims to increase productivity when creating hardware by enabling designers to use features present in higher level programming languages to build complex hardware blocks. In this paper, the most advertised features of Chisel are investigated and compared to their VHDL counterparts, if present. Afterwards, the authors’ opinion if a switch to Chisel is worth considering is presented. Additionally, results from a related case study on Chisel are briefly summarized. The author concludes that, while Chisel has promising features, it is not yet ready for use in the industry.

The Bus Evacuation Problem (BEP) is a vehicle routing problem that arises in emergency planning. It models the evacuation of a region from a set of collection points to a set of capacitated shelters with the help of buses, minimizing the time needed to bring the last person out of the endangered region.
In this work, we describe multiple approaches for finding both lower and upper bounds for the BEP, and apply them in a branch and bound framework. Several node pruning techniques and branching rules are discussed. In computational experiments, we show that solution times of our approach are significantly improved compared to a commercial integer programming solver.

We consider the problem of evacuating a region with the help of buses. For a given set of possible collection points where evacuees gather, and possible shelter locations where evacuees are brought to, we need to determine both collection points and shelters we would like to use, and bus routes that evacuate the region in minimum time.
We model this integrated problem using an integer linear program, and present a branch-cut-and-price algorithm that generates bus tours in its pricing step. In computational experiments we show that our approach is able to solve instances of realistic size in sufficient time for practical application, and considerably outperforms the usage of a generic ILP solver.

We consider the problem of scheduling a bus fleet to evacuate persons from an endangered region. As most of the planning data is subject to uncertainty, we develop a two-stage bicriteria robust formulation, which considers both the evacuation time, and the vulnerability of the schedule to changing evacuation circumstances.
As the resulting integer program is too large to solve it directly using an off-the-shelf solver, we develop an iterative algorithm that successively adds new scenarios to the currently considered subproblem. In computational experiments, we show that this approach is fast enough to deal with an instance modeling an evacuation case within the city of Kaiserslautern, Germany.

An huge amount of computational models and programming languages have been proposed
for the description of embedded systems. In contrast to traditional sequential programming
languages, they cope directly with the requirements for embedded systems: direct support for
concurrent computations and periodic interaction with the environment are only some of the
features they offer. Synchronous languages are one class of languages for the development of
embedded systems and they follow the fundamental principle that the execution is divided into
a sequence of logical steps. Thereby, each step follows the simplification that the computation
of the outputs is finished directly when the inputs are available. This rigorous abstraction leads
to well-defined deterministic parallel composition in general, and to deterministic abortion
and suspension in imperative synchronous languages in particular. These key features also
allow to translate programs to hardware and software, and also formal verification techniques
like model checking can be easily applied.
Besides the advantages of imperative synchronous languages, also some drawbacks can
be listed. Over-synchronization is an effect being caused by parallel threads which have to
synchronize for each execution step, even if they do not communicate, since the synchronization
is implicitly forced by the control-flow. This thesis considers the idea of clock refinement to
introduce several abstraction layers for communication and synchronization in addition to the
existing single-clock abstraction. Thereby, clocks can be refined by several independent clocks
so that a controlled amount of asynchrony between subsequent synchronization points can be
exploited by compilers. The declarations of clocks form a tree, and clocks can be defined within
the threads of the parallel statement, which allows one to do independent computations based
on these clocks without synchronizing the threads. However, the synchronous abstraction is
kept at each level of the abstraction.
Clock refinement is introduced in this thesis as an extension to the imperative synchronous
language Quartz. Therefore, new program statements are introduced which allow to define
a new clock as a refinement of an existing one and to finish a step based on a certain clock.
Examples are considered to show the impact of the behavior of the new statements to
the already existing statements, before the semantics of this extension is formally defined.
Furthermore, the thesis presents a compile algorithm to translate programs to an intermediate
format, and to translate the intermediate format to a hardware description. The advantages
obtained by the new modeling feature are finally evaluated based on examples.

The research presented in this PhD thesis is a contribution to the field of anion recognition in competitive aqueous solvent mixtures. Neutral anion receptors having a cage-type architecture have been developed on the basis of triply-linked bis(cyclopeptides) and their binding properties toward various inorganic anions have been studied.
The synthetic approaches chosen to assemble the targeted container molecules rely on dynamic chemistry under the template effects of anions such as sulfate and halides. As reversible reactions metal-ligand exchange and thiol-disulfide exchange were used. Disulfide exchange has previously provided singly- and doubly-linked bis(cyclopeptide) receptors whose anion affinities in 2:1 acetonitrile/water mixtures approached the nanomolar range. Metal-ligand interactions have so far not been used to assemble bis(cyclopeptides) in our group. The cyclopeptide building blocks required for both approaches, namely cyclic hexapeptides containing alternating 6-aminopicolinic acid and either (2S,4S)-4-cyanoproline or (2S,4S)-4-thioproline subunits could be synthesized successfully.
Self-assembly of the bis(cyclopeptide) held together by coordinative interactions has been attempted by treating the cyclopeptide trinitrile with square-planar palladium (II) complexes. The reaction was followed with different NMR spectroscopic techniques. Unfortunately, none of the experiments provides conclusive evidence that the targeted triply-linked cage was indeed formed.
Bis(cyclopeptides) containing three dithiol derived linkers between the cyclopeptide rings could be synthesizes successfully. Two complexes were isolated, albeit in small amounts, one containing linkers derived from 1,2-ethanedithiol and the other one from 1,3-benzenedithiol that contain a sulfate anion incorporated in the cavity between the cyclopeptide rings. Formation of triply-linked bis(cyclopeptides) containing different types of linkers could be achieved by performing the synthesis in the presence of different dithiols. Unfortunately, the two C3 symmetrical bis(cyclopeptides) containing a single linker type could not be isolated in analytically pure form so that only qualitative binding studies could be performed. Investigations in this context indicate extraordinary sulfate affinity for these bis(cyclopeptides). In particular, affinity of the receptor containing the 1,2-ethanedithiol linkers for sulfate anions is so high that is even able to dissolve barium sulfate under appropriate conditions and presumably exceeds the sulfate affinity of the doubly-linked bis(cyclopeptides). The sulfate anion present in the cavity of this bis(cyclopeptide) can be replaced by a large number of other anions, i.e. by selenate, perrhenate, nitrate, tetrafluoroborate, hexafluorophosphate and halides. None of these complexes proved to be as stable as the corresponding sulfate complex. In addition, 1H-NMR spectroscopic investigations provided information about the solution structure of the bis(cyclopeptide) anion complexes. Sulfate release from the cavity of the receptor is a slow process while exchange of other anions is significantly faster. Another interesting feature that has been observed for sulfate and selenate complexes of the 1,2-ethanedithiol-containing bis(cyclopeptide) is the very slow H/D rate with which protons on amide groups located inside the cavity of the cage are replaced by deuterium atoms in protic deuterated solvents. This effect in combination with the observation that the different deuterated bis(cyclopeptide) species exhibit individual amide NH signals in the 1H-NMR spectrum are indicative for well defined complex geometries with strong hydrogen-bonding interactions between the anion and the amide NH groups of the receptor. Following the H/D exchange rate in the presence of various salts indicated that anion exchange proceeds via the dissociated complex and not by direct replacement of one anion by another one.