### Refine

#### Year of publication

- 2012 (61) (remove)

#### Document Type

- Doctoral Thesis (33)
- Report (12)
- Preprint (7)
- Article (4)
- Master's Thesis (2)
- Conference Proceeding (1)
- Periodical Part (1)
- Working Paper (1)

#### Language

- English (61) (remove)

#### Keywords

- Transaction Costs (2)
- Arithmetic data-path (1)
- Bildverarbeitung (1)
- Bioinformatik (1)
- Carbon footprint (1)
- Chlamydomonas reinhardii (1)
- Cloud Computing (1)
- Cohen-Lenstra heuristic (1)
- Computeralgebra (1)
- Consistent Price Processes (1)

#### Faculty / Organisational entity

In urban planning, both measuring and communicating sustainability are among the most recent concerns. Therefore, the primary emphasis of this thesis concerns establishing metrics and visualization techniques in order to deal with indicators of sustainability.
First, this thesis provides a novel approach for measuring and monitoring two indicators of sustainability - urban sprawl and carbon footprints – at the urban neighborhood scale. By designating different sectors of relevant carbon emissions as well as different household categories, this thesis provides detailed information about carbon emissions in order to estimate impacts of daily consumption decisions and travel behavior by household type. Regarding urban sprawl, a novel gridcell-based indicator model is established, based on different dimensions of urban sprawl.
Second, this thesis presents a three-step-based visualization method, addressing predefined requirements for geovisualizations and visualizing those indicator results, introduced above. This surface-visualization combines advantages from both common GIS representation and three-dimensional representation techniques within the field of urban planning, and is assisted by a web-based graphical user interface which allows for accessing the results by the public.
In addition, by focusing on local neighborhoods, this thesis provides an alternative approach in measuring and visualizing both indicators by utilizing a Neighborhood Relation Diagram (NRD), based on weighted Voronoi diagrams. Thus, the user is able to a) utilize original census data, b) compare direct impacts of indicator results on the neighboring cells, and c) compare both indicators of sustainability visually.

The safety of embedded systems is becoming more and more important nowadays. Fault Tree Analysis (FTA) is a widely used technique for analyzing the safety of embedded systems. A standardized tree-like structure called a Fault Tree (FT) models the failures of the systems. The Component Fault Tree (CFT) provides an advanced modeling concept for adapting the traditional FTs to the hierarchical architecture model in system design. Minimal Cut Set (MCS) analysis is a method that works for qualitative analysis based on the FTs. Each MCS represents a minimal combination of component failures of a system called basic events, which may together cause the top-level system failure. The ordinary representations of MCSs consist of plain text and data tables with little additional supporting visual and interactive information. Importance analysis based on FTs or CFTs estimates the contribution of each potential basic event to a top-level system failure. The resulting importance values of basic events are typically represented in summary views, e.g., data tables and histograms. There is little visual integration between these forms and the FT (or CFT) structure. The safety of a system can be improved using an iterative process, called the safety improvement process, based on FTs taking relevant constraints into account, e.g., cost. Typically, relevant data regarding the safety improvement process are presented across multiple views with few interactive associations. In short, the ordinary representation concepts cannot effectively facilitate these analyses.
We propose a set of visualization approaches for addressing the issues above mentioned in order to facilitate those analyses in terms of the representations.
Contribution:
1. To support the MCS analysis, we propose a matrix-based visualization that allows detailed data of the MCSs of interest to be viewed while maintaining a satisfactory overview of a large number of MCSs for effective navigation and pattern analysis. Engineers can also intuitively analyze the influence of MCSs of a CFT.
2. To facilitate the importance analysis based on the CFT, we propose a hybrid visualization approach that combines the icicle-layout-style architectural views with the CFT structure. This approach facilitates to identify the vulnerable components taking the hierarchies of system architecture into account and investigate the logical failure propagation of the important basic events.
3. We propose a visual safety improvement process that integrates an enhanced decision tree with a scatter plot. This approach allows one to visually investigate the detailed data related to individual steps of the process while maintaining the overview of the process. The approach facilitates to construct and analyze improvement solutions of the safety of a system.
Using our visualization approaches, the MCS analysis, the importance analysis, and the safety improvement process based on the CFT can be facilitated.

This thesis deals with the relationship between no-arbitrage and (strictly) consistent price processes for a financial market with proportional transaction costs
in a discrete time model. The exact mathematical statement behind this relationship is formulated in the so-called Fundamental Theorem of Asset Pricing (FTAP). Among the many proofs of the FTAP without transaction costs there
is also an economic intuitive utility-based approach. It relies on the economic
intuitive fact that the investor can maximize his expected utility from terminal
wealth. This approach is rather constructive since the equivalent martingale measure is then given by the marginal utility evaluated at the optimal terminal payoff.
However, in the presence of proportional transaction costs such a utility-based approach for the existence of consistent price processes is missing in the literature. So far, rather deep methods from functional analysis or from the theory of random sets have been used to show the FTAP under proportional transaction costs.
For the sake of existence of a utility-maximizing payoff we first concentrate on a generic single-period model with only one risky asset. The marignal utility evaluated at the optimal terminal payoff yields the first component of a
consistent price process. The second component is given by the bid-ask prices
depending on the investors optimal action. Even more is true: nearby this consistent price process there are many strictly consistent price processes. Their exact structure allows us to apply this utility-maximizing argument in a multi-period model. In a backwards induction we adapt the given bid-ask prices in such a way so that the strictly consistent price processes found from maximizing utility can be extended to terminal time. In addition possible arbitrage opportunities of the 2nd kind vanish which can present for the original bid-ask process. The notion of arbitrage opportunities of the 2nd kind has been so
far investigated only in models with strict costs in every state. In our model
transaction costs need not be present in every state.
For a model with finitely many risky assets a similar idea is applicable. However, in the single-period case we need to develop new methods compared
to the single-period case with only one risky asset. There are mainly two reasons
for that. Firstly, it is not at all obvious how to get a consistent price process
from the utility-maximizing payoff, since the consistent price process has to be
found for all assets simultaneously. Secondly, we need to show directly that the
so-called vector space property for null payoffs implies the robust no-arbitrage condition. Once this step is accomplished we can à priori use prices with a
smaller spread than the original ones so that the consistent price process found
from the utility-maximizing payoff is strictly consistent for the original prices.
To make the results applicable for the multi-period case we assume that the prices are given by compact and convex random sets. Then the multi-period case is similar to the case with only one risky asset but more demanding with regard to technical questions.

This thesis is devoted to furthering the tropical intersection theory as well as to applying the
developed theory to gain new insights about tropical moduli spaces.
We use piecewise polynomials to define tropical cocycles that generalise the notion of tropical Cartier divisors to higher codimensions, introduce an intersection product of cocycles with tropical cycles and use the connection to toric geometry to prove a Poincaré duality for certain cases. Our
main application of this Poincaré duality is the construction of intersection-theoretic fibres under a
large class of tropical morphisms.
We construct an intersection product of cycles on matroid varieties which are a natural
generalisation of tropicalisations of classical linear spaces and the local blocks of smooth tropical
varieties. The key ingredient is the ability to express a matroid variety contained in another matroid variety by a piecewise polynomial that is given in terms of the rank functions of the corresponding
matroids. In particular, this enables us to intersect cycles on the moduli spaces of n-marked abstract
rational curves. We also construct a pull-back of cycles along morphisms of smooth varieties, relate
pull-backs to tropical modifications and show that every cycle on a matroid variety is rationally
equivalent to its recession cycle and can be cut out by a cocycle.
Finally, we define families of smooth rational tropical curves over smooth varieties and construct a tropical fibre product in order to show that every morphism of a smooth variety to the moduli space of abstract rational tropical curves induces a family of curves over the domain of the morphism.
This leads to an alternative, inductive way of constructing moduli spaces of rational curves.

SHIM is a concurrent deterministic programming language for embedded systems built on rendezvous communication. It abstracts away many details to give the developer a high-level view that includes virtual shared variables, threads as orthogonal statements, and deterministic concurrent exceptions.
In this paper, we present a new way to compile a SHIM-like language into a set of asynchronous guarded actions, a well-established intermediate representation for concurrent systems. By doing so, we build a bridge to many other tools, including hardware synthesis and formal verification. We present our translation in detail, illustrate it through examples, and show how the result can be used by various other tools.

By natural or man-made disasters, the evacuation of a whole region or city may become necessary. Apart from private traffic, the evacuation from collection points to secure shelters outside the endangered region will be realized by a bus fleet made available by emergency relief. The arising Bus Evacuation Problem (BEP) is a vehicle scheduling problem, in which a given number of evacuees needs to be transported from a set of collection points to a set of capacitated shelters, minimizing the total evacuation time, i.e., the time needed until the last person is brought to safety.
In this paper we consider an extended version of the BEP, the Robust Bus Evacuation Problem (RBEP), in which the exact numbers of evacuees are not known, but may stem from a set of probable scenarios. However, after a given reckoning time, this uncertainty is eliminated and planners are given exact figures. The problem is to decide for each bus, if it is better to send it right away -- using uncertain numbers of evacuees -- or to wait until the numbers become known.
We present a mixed-integer linear programming formulation for the RBEP and discuss solution approaches; in particular, we present a tabu search framework for finding heuristic solutions of acceptable quality within short computation time. In computational experiments using both randomly generated instances and the real-world scenario of evacuating the city of Kaiserslautern, we compare our solution approaches.

Due to their N-glycosidase activity, ribosome-inactivating proteins (RIPs) are attractive candidates as antitumor and antiviral agents in medical and biological research. In the present study, we have successfully cloned two different truncated gelonins into pET-28a(+) vectors and expressed intact recombinant gelonin (rGel), recombinant C-terminally truncated gelonin (rC3-gelonin) and recombinant N- and C-terminally truncated gelonin (rN34C3-gelonin). Biological experiments showed that all these recombinant gelonins have no inhibiting effect on MCF-7 cell lines. These data suggest that the truncated-gelonins are still having a specific structure that does not allow for internalization into cells. Further, truncation of gelonin leads to partial or complete loss of N-glycosidase as well as DNase activity compared to intact rGel. Our data suggest that C-and N-terminal amino acid residues are involved in the catalytic and cytotoxic activities of rGel. In addition, the intact gelonin should be selected as a toxin in the immunoconjugate rather than truncated gelonin.
In the second part, an immunotoxin composed of gelonin, a basic protein of 30 kDa isolated from the Indian plant Gelonium multiflorum and the cytotoxic drug MTX has been studied as a potential tool of gelonin delivery into the cytoplasm of cells. Results of many experiments showed that, on the average, about 5 molecules of MTX were coupled to one molecule of gelonin. The MTX-gelonin conjugate is able to reduce the viability of MCF-7 cell in a dose-dependent manner (ID50, 10 nM) as shown by MTT assay and significantly induce direct and oxidative DNA damage as shown by the alkaline comet assay. However, in-vitro translation toxicity MTX-gelonin conjugates have IC50, 50.5 ng/ml which is less toxic than that of gelonin alone IC50, 4.6 ng/ml. It can be concluded that the positive charge plays an important role in the N-glycosidase activity of gelonin. Furthermore, conjugation of MTX with gelonin through α- and γ- carboxyl groups leads to the partial loss of its anti-folate activity compared to free MTX. These results, taken together, indicate that conjugation of MTX to gelonin permits delivery of the gelonin into the cytoplasm of cancer cells and exerts a measurable toxic effect.
In the third part, we have isolated and characterized two ribosome-inactivating proteins (RIPs) type I, gelonin and GAP31, from seeds of Gelonium multiflorum. Both proteins exhibit RNA-N-glycosidase activity. The amino acid sequences of gelonin and GAP31 were identified by MALDI and ESI mass spectrometry. Gelonin and GAP31 peptides - obtained by proteolytic digestion (trypsin and Arg-C) - are consistent with the amino acid sequence published by Rosenblum and Huang, respectively. Further structural characterization of gelonin and GAP31 (tryptic and Arg-C peptide mapping) showed that the two RIPs have 96% similarity in their sequence. Thus, these two proteins are most probably isoforms arisen from the same gene by alternative splicing. The ESI-MS analysis of gelonin and GAP31 exhibited at least three different post-translational modified forms. A standard plant paucidomannosidic N-glycosylation pattern (GlcNAc2Man2-5Xyl0-1 and GlcNAc2Man6-12Fuc1-2Xyl0-2) was identified using electrospray ionization MS for gelonin on N196 and GAP31 on N189, respectively. Based on these results, both proteins are located in the vacuoles of Gelonium multiflorum seeds.

We consider a variant of the generalized assignment problem (GAP) where the amount of space used in each bin is restricted to be either zero (if the bin is not opened) or above a given lower bound (a minimum quantity). We provide several complexity results for different versions of the problem and give polynomial time exact algorithms and approximation algorithms for restricted cases.
For the most general version of the problem, we show that it does not admit a polynomial time approximation algorithm (unless P=NP), even for the case of a single bin. This motivates to study dual approximation algorithms that compute solutions violating the bin capacities and minimum quantities by a constant factor. When the number of bins is fixed and the minimum quantity of each bin is at least a factor \(\delta>1\) larger than the largest size of an item in the bin, we show how to obtain a polynomial time dual approximation algorithm that computes a solution violating the minimum quantities and bin capacities by at most a factor \(1-\frac{1}{\delta}\) and \(1+\frac{1}{\delta}\), respectively, and whose profit is at least as large as the profit of the best solution that satisfies the minimum quantities and bin capacities strictly.
In particular, for \(\delta=2\), we obtain a polynomial time (1,2)-approximation algorithm.

This paper considers supervised multi-class image segmentation: from a labeled set of
pixels in one image, we learn the segmentation and apply it to the rest of the image or to other similar images. We study approaches with p-Laplacians, (vector-valued) Reproducing Kernel Hilbert
Spaces (RKHSs) and combinations of both. In all approaches we construct segment membership
vectors. In the p-Laplacian model the segment membership vectors have to fulfill a certain probability simplex constraint. Interestingly, we could prove that this is not really a constraint in the case p=2 but is automatically fulfilled. While the 2-Laplacian model gives a good general segmentation, the case of the 1-Laplacian tends to neglect smaller segments. The RKHS approach has
the benefit of fast computation. This direction is motivated by image colorization, where a given
dab of color is extended to a nearby region of similar features or to another image. The connection
between colorization and multi-class segmentation is explored in this paper with an application to
medical image segmentation. We further consider an improvement using a combined method. Each
model is carefully considered with numerical experiments for validation, followed by medical image
segmentation at the end.

One of the fundamental problems in computational structural biology is the prediction of RNA secondary structures from a single sequence. To solve this problem, mainly two different approaches have been used over the past decades: the free energy minimization (MFE) approach which is still considered the most popular and successful method and the competing stochastic context-free grammar (SCFG) approach. While the accuracy of the MFE based algorithms is limited by the quality of underlying thermodynamic models, the SCFG method abstracts from free energies and instead tries to learn about the structural behavior of the molecules by training the grammars on known real RNA structures, making it highly dependent on the availability of a rich high quality training set. However, due to the respective problems associated with both methods, new statistics based approaches towards RNA structure prediction have become increasingly appreciated. For instance, over the last years, several statistical sampling methods and clustering techniques have been invented that are based on the computation of partition functions (PFs) and base pair probabilities according to thermodynamic models. A corresponding SCFG based statistical sampling algorithm for RNA secondary structures has been studied just recently. Notably, this probabilistic method is capable of producing accurate (prediction) results, where its worst-case time and space requirements are equal to those of common RNA folding algorithms for single sequences.
The aim of this work is to present a comprehensive study on how enriching the underlying SCFG by additional information on the lengths of generated substructures (i.e. by incorporating length-dependencies into the SCFG based sampling algorithm, which is actually possible without significant losses in performance) affects the reliability of the induced RNA model and the accuracy of sampled secondary structures. As we will see, significant differences with respect to the overall quality of generated sample sets and the resulting predictive accuracy are typically implied. In principle, when considering the more specialized length-dependent SCFG model as basis for statistical sampling, a higher accuracy of predicted foldings can be reached at the price of a lower diversity of generated candidate structures (compared to the more general traditional SCFG variant or sampling based on PFs that rely on free energies).

Standard bases are one of the main tools in computational commutative algebra. In 1965
Buchberger presented a criterion for such bases and thus was able to introduce a first approach for their computation. Since the basic version of this algorithm is rather inefficient
due to the fact that it processes lots of useless data during its execution, active research for
improvements of those kind of algorithms is quite important.
In this thesis we introduce the reader to the area of computational commutative algebra with a focus on so-called signature-based standard basis algorithms. We do not only
present the basic version of Buchberger’s algorithm, but give an extensive discussion of different attempts optimizing standard basis computations, from several sorting algorithms
for internal data up to different reduction processes. Afterwards the reader gets a complete
introduction to the origin of signature-based algorithms in general, explaining the under-
lying ideas in detail. Furthermore, we give an extensive discussion in terms of correctness,
termination, and efficiency, presenting various different variants of signature-based standard basis algorithms.
Whereas Buchberger and others found criteria to discard useless computations which
are completely based on the polynomial structure of the elements considered, Faugère presented a first signature-based algorithm in 2002, the F5 Algorithm. This algorithm is famous for generating much less computational overhead during its execution. Within this
thesis we not only present Faugère’s ideas, we also generalize them and end up with several
different, optimized variants of his criteria for detecting redundant data.
Being not completely focussed on theory, we also present information about practical
aspects, comparing the performance of various implementations of those algorithms in the
computer algebra system Singular over a wide range of example sets.
In the end we give a rather extensive overview of recent research in this area of computational commutative algebra.

Predicting secondary structures of RNA molecules is one of the fundamental problems of and thus a challenging task in computational structural biology. Existing prediction methods basically use the dynamic programming principle and are either based on a general thermodynamic model or on a specific probabilistic model, traditionally realized by a stochastic context-free grammar. To date, the applied grammars were rather simple and small and despite the fact that statistical approaches have become increasingly appreciated over the past years, a corresponding sampling algorithm based on a stochastic RNA structure model has not yet been devised. In addition, basically all popular state-of-the-art tools for computational structure prediction have the same worst-case time and space requirements of O(n^3) and O(n^2) for sequence length n, limiting their applicability for practical purposes due to the often quite large sizes of native RNA molecules. Accordingly, the prime demand imposed by biologists on computational prediction procedures is to reach a reduced waiting time for results that are not significantly less accurate.
We here deal with all of these issues, by describing algorithms and performing comprehensive studies that are based on sophisticated stochastic context-free grammars of similar complexity as those underlying thermodynamic prediction approaches, where all of our methods indeed make use of the concept of sampling. We also employ the approximation technique known from theoretical computer science in order to reach a heuristic worst-case speedup for RNA folding.
Particularly, we start by describing a way for deriving a sequence-independent random sampler for an arbitrary class of RNAs by means of (weighted) unranking. The resulting algorithm may generate any secondary structure of a given fixed size n in only O(n·log(n)) time, where the results are observed to be accurate, validating its practical applicability.
With respect to RNA folding, we present a novel probabilistic sampling algorithm that generates statistically representative and reproducible samples of the entire ensemble of feasible structures for a particular input sequence. This method actually samples the possible foldings from a distribution implied by a suitable (traditional or length-dependent) grammar. Notably, we also propose several (new) ways for obtaining predictions from generated samples. Both variants have the same worst-case time and space complexities of O(n^3) and O(n^2) for sequence length n. Nevertheless, evaluations of our sampling methods show that they are actually capable of producing accurate (prediction) results.
In an attempt to resolve the long-standing problem of reducing the time complexity of RNA folding algorithms without sacrificing much of the accuracy of the results, we invented an innovative heuristic statistical sampling method that can be implemented to require only O(n^2) time for generating a fixed-size sample of candidate structures for a given sequence of length n. Since a reasonable prediction can still efficiently be obtained from the generated sample set, this approach finally reduces the worst-case time complexity by a liner factor compared to all existing precise methods. Notably, we also propose a novel (heuristic) sampling strategy as opposed to the common one typically applied for statistical sampling, which may produce more accurate results for particular settings. A validation of our heuristic sampling approach by comparison to several leading RNA secondary structure prediction tools indicates that it is capable of producing competitive predictions, but may require the consideration of large sample sizes.

Worldwide the installed capacity of renewable technologies for electricity production is
rising tremendously. The German market is particularly progressive and its regulatory
rules imply that production from renewables is decoupled from market prices and electricity
demand. Conventional generation technologies are to cover the residual demand
(defined as total demand minus production from renewables) but set the price at the
exchange. Existing electricity price models do not account for the new risks introduced
by the volatile production of renewables and their effects on the conventional demand
curve. A model for residual demand is proposed, which is used as an extension of
supply/demand electricity price models to account for renewable infeed in the market.
Infeed from wind and solar (photovoltaics) is modeled explicitly and withdrawn from
total demand. The methodology separates the impact of weather and capacity. Efficiency
is transformed on the real line using the logit-transformation and modeled as a stochastic process. Installed capacity is assumed a deterministic function of time. In a case study the residual demand model is applied to the German day-ahead market
using a supply/demand model with a deterministic supply-side representation. Price trajectories are simulated and the results are compared to market future and option
prices. The trajectories show typical features seen in market prices in recent years and the model is able to closely reproduce the structure and magnitude of market prices.
Using the simulated prices it is found that renewable infeed increases the volatility of forward prices in times of low demand, but can reduce volatility in peak hours. Prices
for different scenarios of installed wind and solar capacity are compared and the meritorder effect of increased wind and solar capacity is calculated. It is found that wind
has a stronger overall effect than solar, but both are even in peak hours.

The main topic of this thesis is to define and analyze a multilevel Monte Carlo algorithm for path-dependent functionals of the solution of a stochastic differential equation (SDE) which is driven by a square integrable, \(d_X\)-dimensional Lévy process \(X\). We work with standard Lipschitz assumptions and denote by \(Y=(Y_t)_{t\in[0,1]}\) the \(d_Y\)-dimensional strong solution of the SDE.
We investigate the computation of expectations \(S(f) = \mathrm{E}[f(Y)]\) using randomized algorithms \(\widehat S\). Thereby, we are interested in the relation of the error and the computational cost of \(\widehat S\), where \(f:D[0,1] \to \mathbb{R}\) ranges in the class \(F\) of measurable functionals on the space of càdlàg functions on \([0,1]\), that are Lipschitz continuous with respect to the supremum norm.
We consider as error \(e(\widehat S)\) the worst case of the root mean square error over the class of functionals \(F\). The computational cost of an algorithm \(\widehat S\), denoted \(\mathrm{cost}(\widehat S)\), should represent the runtime of the algorithm on a computer. We work in the real number model of computation and further suppose that evaluations of \(f\) are possible for piecewise constant functions in time units according to its number of breakpoints.
We state strong error estimates for an approximate Euler scheme on a random time discretization. With this strong error estimates, the multilevel algorithm leads to upper bounds for the convergence order of the error with respect to the computational cost. The main results can be summarized in terms of the Blumenthal-Getoor index of the driving Lévy process, denoted by \(\beta\in[0,2]\). For \(\beta <1\) and no Brownian component present, we almost reach convergence order \(1/2\), which means, that there exists a sequence of multilevel algorithms \((\widehat S_n)_{n\in \mathbb{N}}\) with \(\mathrm{cost}(\widehat S_n) \leq n\) such that \( e(\widehat S_n) \precsim n^{-1/2}\). Here, by \( \precsim\), we denote a weak asymptotic upper bound, i.e. the inequality holds up to an unspecified positive constant. If \(X\) has a Brownian component, the order has an additional logarithmic term, in which case, we reach \( e(\widehat S_n) \precsim n^{-1/2} \, (\log(n))^{3/2}\).
For the special subclass of $Y$ being the Lévy process itself, we also provide a lower bound, which, up to a logarithmic term, recovers the order \(1/2\), i.e., neglecting logarithmic terms, the multilevel algorithm is order optimal for \( \beta <1\).
An empirical error analysis via numerical experiments matches the theoretical results and completes the analysis.

The scientific aim of this work was to synthesize and characterize new bidentate and tridentate phosphine ligands , their corresponding palladium complexes and to examine their application as homogenous catalysts. Later on, a part of the obtained palladium catalysts was immobilized and used as heterogonous catalyst.
Pyrimidinyl functionalized diphenyl phosphine ligands were synthesized by ring closure of [2-(3-dimethylamino-1-oxoprop-2-en-yl)phenyl]diphenylphosphine with an excess of substituted guanidinium salts. Furthermore to increase the electron density at phosphorous centre the two aryl substituents on the phosphanyl group were exchanged against two alkyl substituents. Electron rich pyrimidinyl functionalized dialkyl phosphine ligands were synthesized from pyrimidinyl functionalized bromobenzene in a process involving lithiation followed by reaction with a chlorodialkylphosphine.
Starting from the new synthesized diaryl phosphine ligands, their corresponding palladium complexes were synthesized. I was able to show that slight changes at the amino group of [(2-aminopyrimidin-4-yl)aryl]phosphines lead to pronounced differences in the stability and catalytic activity of the corresponding palladium(II) complexes. Having a P,C coordination mode, the palladium complex can catalyze rapidly the Suzuki coupling reaction of phenylbronic acid with arylbromides even at room temperature with a low loading.
Using the NH2 group of the aminopyrimidine as a potential site for the introduction of an other substituent, bidentate and tridentate ligands containing phosphorous atoms connected to the aminopyrimidine group and their corresponding palladium complexes were synthesized and characterized.
Two ligands [2- and 4-(4-(2-amino)pyrimidinyl)phenyl]diphenylphosphine (containing NH2 group) functionalized with a ethoxysilane group were synthesized. The palladium complexes based on these ligands were prepared and immobilized on commercial silica and MCM-41. Using elemental analysis, FT-IR, solid state 31P, 13C and 29Si CP–MAS NMR spectroscopy, XRD and N2 adsorption the success of the immobilization was confirmed and the structure of the heterogenized catalyst was investigated.
The resulting heterogeneous catalysts were applied for the Suzuki reaction and exhibited excellent activity, selectivity and reusability.

This thesis generalizes the Cohen-Lenstra heuristic for the class groups of real quadratic
number fields to higher class groups. A "good part" of the second class group is defined.
In general this is a non abelian proper factor group of the second class group. Properties
of those groups are described, a probability distribution on the set of those groups is in-
troduced and proposed as generalization of the Cohen-Lenstra heuristic for real quadratic
number fields. The calculation of number field tables which contain information about
higher class groups is explained and the tables are compared to the heuristic. The agree-
ment is close. A program which can create an internet database for number field tables is
presented.

On Gyroscopic Stabilization
(2012)

This thesis deals with systems of the form
\(
M\ddot x+D\dot x+Kx=0\;, \; x \in \mathbb R^n\;,
\)
with a positive definite mass matrix \(M\), a symmetric damping matrix \(D\) and a positive definite stiffness
matrix \(K\).
If the equilibrium in the system is unstable, a small disturbance is enough to set the system in motion again. The motion of the system sustains itself, an effect which is called self-excitation or self-induced vibration. The reason behind this effect is the presence of negative damping, which results for example from dry friction.
Negative damping implies that the damping matrix \(D\) is indefinite or negative definite. Throughout our work, we assume \(D\) to be indefinite, and that the system possesses both stable and unstable modes and thus is unstable.
It is now the idea of gyroscopic stabilization to mix the modes of a system with indefinite damping such
that the system is stabilized without introducing further
dissipation. This is done by adding gyroscopic forces \(G\dot x\) with a suitable
skew-symmetric matrix \(G\) to the left-hand side. We call \(G=-G^T\in\mathbb R^{n\times n}\) a gyroscopic stabilizer for
the unstable system, if
\(
M\ddot x+(D+ G)\dot x+Kx=0
\)
is asymptotically stable. We show the existence of \(G\) in space dimensions three and four.

The increasing complexity of modern SoC designs makes tasks of SoC formal verification
a lot more complex and challenging. This motivates the research community to develop
more robust approaches that enable efficient formal verification for such designs.
It is a common scenario to apply a correctness by integration strategy while a SoC
design is being verified. This strategy assumes formal verification to be implemented in
two major steps. First of all, each module of a SoC is considered and verified separately
from the other blocks of the system. At the second step – when the functional correctness
is successfully proved for every individual module – the communicational behavior has
to be verified between all the modules of the SoC. In industrial applications, SAT/SMT-based interval property checking(IPC) has become widely adopted for SoC verification. Using IPC approaches, a verification engineer is able to afford solving a wide range of important verification problems and proving functional correctness of diverse complex components in a modern SoC design. However, there exist critical parts of a design where formal methods often lack their robustness. State-of-the-art property checkers fail in proving correctness for a data path of an industrial central processing unit (CPU). In particular, arithmetic circuits of a realistic size (32 bits or 64 bits) – especially implementing multiplication algorithms – are well-known examples when SAT/SMT-based
formal verification may reach its capacity very fast. In cases like this, formal verification
is replaced with simulation-based approaches in practice. Simulation is a good methodology that may assure a high rate of discovered bugs hidden in a SoC design. However, in contrast to formal methods, a simulation-based technique cannot guarantee the absence of errors in a design. Thus, simulation may still miss some so-called corner-case bugs in the design. This may potentially lead to additional and very expensive costs in terms of time, effort, and investments spent for redesigns, refabrications, and reshipments of new chips.
The work of this thesis concentrates on studying and developing robust algorithms
for solving hard arithmetic decision problems. Such decision problems often originate from a task of RTL property checking for data-path designs. Proving properties of those
designs can efficiently be performed by solving SMT decision problems formulated with
the quantifier-free logic over fixed-sized bit vectors (QF-BV).
This thesis, firstly, proposes an effective algebraic approach based on a Gröbner basis theory that allows to efficiently decide arithmetic problems. Secondly, for the case of custom-designed components, this thesis describes a sophisticated modeling technique which is required to restore all the necessary arithmetic description from these components. Further, this thesis, also, explains how methods from computer algebra and the modeling techniques can be integrated into a common SMT solver. Finally, a new QF-BV SMT solver is introduced.

The scientific intention of this work was to synthesize and characterize new bidentate, tridentate and multidentate ligands and to apply them in heterogenous catalysis. For each type of the ligands, new methods of synthesis were developed. Starting from 1,1'-(pyridine-2,6-diyl)diethanone and dimethylpyridine-2,6-dicarboxylate different bispyrazolpyridines were
synthesized and novel ruthenium complexes of the type (L)(NNN)RuCl2 could be obtained. The complexes with L = triphenylphosphine turned out to be highly efficient
catalyst precursors for the transfer hydrogenation of aromatic ketones. Introduction of a butyl group in the 5-positions of the pyrazoles leads to a pronounced increase of catalytic activity.
To find a method for the synthesis of bispyrimidinepyridines, different reactants and condition were applied and it was found that these tridentate ligands can be obtained by mixing and grinding the tetraketone with guanidinium carbonate and silica, which plays the role of a catalyst in this ring closing reaction.
The bidentate 2-amino-4-(2-pyridinyl)pyrimidines were synthesized from different substrates according to the desired substituent on the pyrimidine ring.
Reacting these bidentate ligands with the ruthenium(II) precursor [(η6-cymene)Ru(Cl)(μ
2-Cl)]2 gave cationic ruthenium(II) complexes of the type [(η6-cymene)Ru(Cl)(adpm)]Cl (adpm = chelating 2-amino-4-(2-yridinyl)pyrimidine ligand). Stirring the freshly prepared complexes with either NaBPh4, NaBF4 or KPF6, the chloride anion was exchanged against other coordinating anions (BF4-, PF6-, BPh4-).Some of these ruthenium complexes have shown very special activities in the transfer hydrogenation of ketones by reacting them in the absence of the base. This led to detailed investigations on the mechanism of this reaction. According to the activities and with the help
of ESI-MS experiments and DFT calculations, a mechanism was proposed for the transfer hydrogenation of acetophenone in the absence of the base. It shows that in the absence of the base, a C-H bond activation at the pyrimidine ring should occur to activate the catalyst.
The palladium complexes of bidentate N,N ligands were examined in coupling reactions. As expected, they did not show very special activities.
Multidentate ligands, having pyrimidine groups as relatively soft donors for late transition metals and simultaneously possessing a binding position for a hard Lewis-acid, could be obtained using the new synthesized bidentate and tridentate ligands.

In this study, two outstanding subgroups of organic-inorganic hybrid materials have been investigated. The first part covers the design, synthesis, characterization and application of seven novel Metal Organic Frameworks (MOFs) containing functionalized biphenyl dicarboxylates as linkers. In the second part, the surface modification of the metal oxides ZrO2, TiO2 and Al2O3 using phosphonate derivates is reported.
Firstly three functionalized MOF structures; ZnBrBPDC, ZnNO2BPDC and ZnNH2BPDC were synthesised using 4,4´-biphenyldicarboxylic acid derivatives with different functional groups (-Br, -NO2, -NH2) Powder X-ray diffraction (PXRD) measurements indicated that the synthesised MOFs posses the interpenetrated IRMOF-9 structure with a cubic topology, which was also confirmed with single crystal X-ray measurements. The chemical structure of the MOF materials was further proved by solid state NMR and IR measurements. N2 adsorption measurements showed Type I isotherms for all three structures with large surface areas. TGA measurements of the evacuated samples were in good agreement with the elemental analysis data. The results proved that their thermal stability is between 325 °C - 450 °C.
Adsorption properties of these MOF structures were tested using light alkanes (CH4, C2H6, C3H8, and n-C4H10) at three different temperatures. For all adsorbents, the maximum uptakes were observed at 273 K. When the temperature was increased, the amount of the adsorbed gas decreased. All three MOFs showed strong affinities for n-butane. The lowest uptakes were observed for CH4.
The effect of functional groups on the IRMOF series was also examined by synthesizing amide functionalized biphenyl linkers. For this purpose, four different linkers containing amides with different alkyl chains (C1-C4) were synthesized and used for the synthesis of four new MOF structures ZnAcBPDC, ZnPrBPDC, ZnBuBPDC and ZnPeBPDC.
PXRD measurements of ZnAcBPDC indicated that the structure contains two different phases. PXRD patterns of ZnPrBPDC, ZnBuBPDC and ZnPeBPDC revealed non-interpenetrated structures which were further proved by single crystal X-ray measurements. The chemical structure of the MOF materials was further confirmed by X-ray spectoscopy, solid state NMR and IR measurements.
N2 adsorption measurements of the MOF structures were carried out using different activation methods. For all four MOFs, Type I isotherms were obtained. ZnAcBPDC showed the highest BET surface area. ZnAcBPDC and ZnBuBPDC were tested for their alkane, alkene and CO2 adsorption capacities.
In the second part of the work, the surface modification of three different metal oxides, ZrO2, TiO2 and Al2O3 was performed. For this purpose firstly three different fluorescent phosphonate derivatives containing thiophene units were synthesized from their halo derivatives in a four step synthesis and then used as coupling molecules for the surface modification. Nine different surfaces were obtained (38@TiO2, 39@TiO2, 40@TiO2, 38@Al2O3, 39@Al2O3, 40@Al2O3, 38@ZrO2, 39@ZrO2, 40@ZrO2).
All three modified metal oxide surfaces were characterized using elemental analysis, solid state NMR and IR spectroscopy. The BET surface areas of the materials were determined by N2 adsorption measurements. TGA was used to determine the stability of the surfaces. Maximum loadings were obtained for ZrO2 surfaces.
Due to the strong luminescence of the coupling molecules, the modified surfaces were checked for their light emission. All ZrO2 and Al2O3 surfaces showed fluorescence with exception of 40@Al2O3. On the other hand, for the modified TiO2 surfaces, no fluorescence could be observed.

This thesis addresses challenges faced by small package shipping companies and investigates the integration of 1) service consistency and driver knowledge aspects and 2) the utilization of electric vehicles into the route planning of small package shippers. We use Operations Research models and solution methods to gain insights into the newly arising problems and thus support managerial decisions concerning these issues.

Nanoparticle-Filled Thermoplastics and Thermoplastic Elastomer: Structure-Property Relationships
(2012)

The present work focuses on the structure-property relationships of
particulate-filled thermoplastics and thermoplastic elastomer (TPE). In this work
two thermoplastics and one TPE were used as polymer matrices, i.e. amorphous
bisphenol-A polycarbonate (PC), semi-crystalline isotactic polypropylene (iPP),
and a block copolymer poly(butylene terephthalate)-block-poly(tetramethylene
glycol) TPE(PBT-PTMG). For PC, a selected type of various Aerosil® nano-SiO2
types was used as filler to improve the thermal and mechanical properties by
maintaining the transparency of PC matrix. Different types of SiO2 and TiO2
nanoparticles with different surface polarity were used for iPP. The goal was to
examine the influence of surface polarity and chemical nature of nanoparticles on
the thermal, mechanical and morphological properties of iPP composites. For
TPE(PBT-PTMG), three TiO2 particles were used, i.e. one grade with hydroxyl
groups on the particle surface and the other two grades are surface-modified with
metal and metal oxides, respectively. The influence of primary size and dispersion
quality of TiO2 particles on the properties of TPE(PBT-PTMG)/TiO2 composites
were determined and discussed.
All polymer composites were produced by direct melt blending in a twin-screw
extruder via masterbatch technique. The dispersion of particles was examined by
using scanning electron microscopy (SEM) and micro-computerized tomography
(μCT). The thermal and crystalline properties of polymer composites were characterized by using thermogravimetric analysis (TGA) and differential
scanning calorimetry (DSC). The mechanical and thermomechanical properties
were determined by using mechanical tensile testing, compact tension and
Charpy impact as well as dynamic-mechanical thermal analysis (DMTA).
The SEM results show that the unpolar-surface modified nanoparticles are better
dispersed in polymer matrices as iPP than polar-surface nanoparticles, especially
in case of using Aeroxide® TiO2 nanoparticles. The Aeroxide® TiO2 nanoparticles
with a polar surface due to Ti-OH groups result in a very high degree of
agglomeration in both iPP and TPE matrices because of strong van der Waals
interactions among particles (hydrogen bonding). Compared to unmodified
Aeroxide® TiO2 nanoparticles, the other grades of surface modified TiO2 particles
are very homogenously dispersed in used iPP and TPE(PBT-PTMG). The
incorporation of SiO2 nanoparticles into bisphenol-A PC significantly increases
the mechanical properties of PC/SiO2 nanocomposites, particularly the resistance
against environmental stress crazing (ESC). However, the transparency of
PC/SiO2 nanocomposites decreases with increasing nanoparticle content and
size due to a mismatch of infractive indices of PC and SiO2 particles. The different
surface polarity of nanoparticles in iPP shows evident influence on properties of
iPP composites. Among iPP/SiO2 nanocomposites, the nanocomposite
containing SiO2 nanoparticles with a higher degree of hydrophobicity shows
improved fracture and impact toughness compared to the other iPP/SiO2
composites. The TPE(PBT-PTMG)/TiO2 composites show much better thermal and mechanical properties than neat TPE(PBT-PTMG) due to strong chemical
interactions between polymer matrix and TiO2 particles. In addition, better
dispersion quality of TiO2 particles in used TPE(PBT-PTMG) leads to dramatically
improved mechanical properties of TPE(PBT-PTMG)/TiO2 composites.

In this work we extend the multiscale finite element method (MsFEM)
as formulated by Hou and Wu in [14] to the PDE system of linear elasticity.
The application, motivated from the multiscale analysis of highly heterogeneous
composite materials, is twofold. Resolving the heterogeneities on
the finest scale, we utilize the linear MsFEM basis for the construction of
robust coarse spaces in the context of two-level overlapping Domain Decomposition
preconditioners. We motivate and explain the construction
and present numerical results validating the approach. Under the assumption
that the material jumps are isolated, that is they occur only in the
interior of the coarse grid elements, our experiments show uniform convergence
rates independent of the contrast in the Young's modulus within the
heterogeneous material. Elsewise, if no restrictions on the position of the
high coefficient inclusions are imposed, robustness can not be guaranteed
any more. These results justify expectations to obtain coefficient-explicit
condition number bounds for the PDE system of linear elasticity similar to
existing ones for scalar elliptic PDEs as given in the work of Graham, Lechner
and Scheichl [12]. Furthermore, we numerically observe the properties
of the MsFEM coarse space for linear elasticity in an upscaling framework.
Therefore, we present experimental results showing the approximation errors
of the multiscale coarse space w.r.t. the fine-scale solution.

In this paper, we propose multi-level Monte Carlo(MLMC) methods that use ensemble level mixed multiscale methods in the simulations of multi-phase flow and transport. The main idea of ensemble level multiscale methods is to construct local multiscale basis functions that can be used for any member of the ensemble. We consider two types of ensemble level mixed multiscale finite element methods, (1) the no-local-solve-online ensemble level method (NLSO) and (2) the local-solve-online ensemble level method (LSO). Both mixed multiscale methods use a number of snapshots of the permeability media to generate a multiscale basis.
As a result, in the offline stage, we construct multiple basis functions for
each coarse region where basis functions correspond to different realizations.
In the no-local-solve-online ensemble level method one uses the whole set of pre-computed basis functions to approximate the solution for an arbitrary realization. In the local-solve-online ensemble level method one uses the pre-computed functions to construct a multiscale basis for a particular realization. With this basis the solution corresponding to this
particular realization is approximated in LSO mixed MsFEM. In both approaches
the accuracy of the method is related to the number of snapshots computed based on different realizations that one uses to pre-compute a
multiscale basis. We note that LSO approaches share similarities with reduced basis methods [11, 21, 22].
In multi-level Monte Carlo methods ([14, 13]), more accurate (and expensive) forward simulations are run with fewer samples while less accurate(and inexpensive) forward simulations are run with a larger number of samples. Selecting the number of expensive and inexpensive simulations carefully, one can show that MLMC methods can provide better accuracy
at the same cost as MC methods. In our simulations, our goal is twofold. First, we would like to compare NLSO and LSO mixed MsFEMs. In particular, we show that NLSO
mixed MsFEM is more accurate compared to LSO mixed MsFEM. Further, we use both approaches in the context of MLMC to speed-up MC
calculations. We present basic aspects of the algorithm and numerical
results for coupled flow and transport in heterogeneous porous media.

The development of autonomous vehicle systems demands the increased usage of software based control mechanisms. Generally, this leads to very complex systems, whose proper functioning has to be ensured. In our work we aim at investigating and assessing the potential effects of software issues on the safety, reliability and availability of complex embedded autonomous systems. One of the key aspects of the research concerns the mapping of functional descriptions in form of integrated behavior-based control networks to State-Event Fault Tree models.

This papers deals with the minimization of seminorms \(\|L\cdot\|\) on \(\mathbb R^n\) under the constraint of a bounded I-divergence \(D(b,H\cdot)\). The I-divergence is also known as Kullback-Leibler divergence and appears in many models in imaging science, in particular when dealing with Poisson data. Typically, \(H\) represents here, e.g., a linear blur operator and \(L\) is some discrete derivative operator. Our preference for the constrained approach over
the corresponding penalized version is based on the fact that the I-divergence of data
corrupted, e.g., by Poisson noise or multiplicative Gamma noise can be estimated by statistical methods. Our minimization technique rests upon relations between constrained and penalized convex problems and resembles the idea of Morozov's discrepancy principle.
More precisely, we propose first-order primal-dual algorithms which reduce the problem to the solution of certain proximal minimization problems in each iteration step. The most interesting of these proximal minimization problems is an I-divergence constrained least squares problem. We solve this problem by connecting it to the corresponding I-divergence
penalized least squares problem with an appropriately chosen regularization parameter. Therefore, our algorithm produces not only a sequence of vectors which converges to a minimizer of the constrained problem but also a sequence of parameters which convergences to a regularization parameter so that the penalized problem has the same solution as our constrained one. In other words, the solution of this penalized problem fulfills the I-divergence constraint. We provide the proofs which are necessary to understand
our approach and demonstrate the performance of our algorithms for different
image restoration examples.

Today, polygonal models occur everywhere in graphical applications, since they are easy
to render and to compute and a very huge set of tools are existing for generation and
manipulation of polygonal data. But modern scanning devices that allow a high quality
and large scale acquisition of complex real world models often deliver a large set of
points as resulting data structure of the scanned surface. A direct triangulation of those
point clouds does not always result in good models. They often contain problems like
holes, self-intersections and non manifold structures. Also one often looses important
surface structures like sharp corners and edges during a usual surface reconstruction.
So it is suitable to stay a little longer in the point based world to analyze the point cloud
data with respect to such features and apply a surface reconstruction method afterwards
that is known to construct continuous and smooth surfaces and extend it to reconstruct
sharp features.

In this thesis we consider the problem of maximizing the growth rate with proportional and fixed costs in a framework with one bond and one stock, which is modeled as a jump diffusion with compound Poisson jumps. Following the approach from [1], we prove that in this framework it is optimal for an investor to follow a CB-strategy. The boundaries depend only on the parameters of the underlying stock and bond. Now it is natural to ask for the investor who follows a CB-strategy which is given by the stopping times \((\tau_i)_{i\in\mathbb N}\) and impulses \((\eta_i)_{i\in\mathbb N}\) how often he has to rebalance. In other words we want to obtain the limit of the inter trading times
\[
\lim_{n\rightarrow\infty}\frac{1}{n}\sum_{i=1}^n(\tau_{i+1}-\tau_{i}).
\]
We are able to obtain this limit which is given by the expected first exit time of the risky fraction process from some interval under the invariant measure of the Markov chain \((\eta_i)_{i\in\mathbb N}\) using the Ergodic Theorem from von Neumann and Birkhoff. In general, it is difficult to obtain the expectation of the first exit time for the process with jumps. Because of the jump part, when the process crosses the boundaries of the interval an overshoot may occur which makes it difficult to obtain the distribution. Nevertheless we can obtain the first exit time if the process has only negative jumps using scale functions. The main difficulty of this approach is that the scale functions are known only up to their Laplace transforms. In [2] and [3] the closed-form expression for the scale function of the Levy process with phase-type distributed jumps is obtained. Phase-type distributions build a rich class of positive-valued distributions: the exponential, hyperexponential, Erlang, hyper-Erlang and Coxian distributions. Since the scale function is given as a function in a closed form we can differentiate to obtain the expected first exit time using the fluctuation identities explicitly.
[1] Irle, A. and Sass,J.: Optimal portfolio policies under fixed and proportional transaction costs, Advances in Applied Probability 38, 916-942.
[2] Egami, M., Yamazaki, K.: On scale functions of spectrally negative Levy processes with phase-type jumps, working paper, July 3.
[3]Egami, M., Yamazaki, K.: Precautionary measures for credit risk management in jump models, working paper, June 17.

In this thesis we outline the Kerner's 3-phase traffic flow theory, which states that the flow of vehicular traffic occur in three phases i.e. free flow, synchronized flow and wide moving jam phases.
A macroscopic 3-phase traffic model of the Aw-Rascle type is derived from the microscopic Speed Adaptation 3-phase traffic model
developed by Kerner and Klenov [J. Phys. A: Math. Gen., 39(2006), pp. 1775-1809 ].
We derive the same macroscopic model from the kinetic traffic flow model of Klar and Wegener [SIAM J. Appl. Math., 60(2000), pp. 1749-1766 ] as well as that of Illner, Klar and Materne [Comm. Math. Sci., 1(2003), pp. 1-12 ].
In the above stated derivations, the 3-phase traffic theory is constituted in the macroscopic model through a relaxation term.
This serves as an incentive to modify the relaxation term of the `switching curve' model of Greenberg,
Klar and Rascle [SIAM J. Appl. Math.,63(2003), pp.818-833 ] to obtain another macroscopic 3-phase traffic model, which is still of the Aw-Rascle type.
By specifying the relaxation term differently we obtain three kinds of models, namely the macroscopic Speed Adaptation,
the Switching Curve and the modified Switching Curve models.
To demonstrate the capability of the derived macroscopic traffic models to reproduce the features of 3-phase traffic theory, we simulate a
multi-lane road that has a bottleneck. We consider a stationary and a moving bottleneck.
The results of the simulations for the three models are compared.

Paper production is a problem with significant importance for the society and it is a challenging topic for scientific investigations. This study is concerned with the simulations of the pressing section of a paper machine. We aim at the development of an advanced mathematical model of the pressing section, which is able to recover the behavior of the fluid flow within the paper felt sandwich obtained in laboratory experiments.
From the modeling point of view the pressing of the paper-felt sandwich is a complex process since one has to deal with the two-phase flow in moving and deformable porous media. To account for the solid deformations, we use developments from the PhD thesis by S. Rief where the elasticity model is stated and discussed in detail. The flow model which accounts for the movement of water within the paper-felt sandwich is described with the help of two flow regimes: single-phase water flow and two-phase air-water flow. The model for the saturated flow is presented by the Darcy's law and the mass conservation. The second regime is described by the Richards' approach together with dynamic capillary effects. The model for the dynamic capillary pressure - saturation relation proposed by Hassanizadeh and Gray is adapted for the needs of the paper manufacturing process.
We have started the development of the flow model with the mathematical modeling in one-dimensional case. The one-dimensional flow model is derived from a two-dimensional one by an averaging procedure in vertical direction. The model is numerically studied and verified in comparison with measurements. Some theoretical investigations are performed to prove the convergence of the discrete solution to the continuous one. For completeness of the studies, the models with the static and dynamic capillary pressure–saturation relations are considered. Existence, compactness and convergence results are obtained for both models.
Then, a two-dimensional model is developed, which accounts for a multilayer computational domain and formation of the fully saturated zones. For discretization we use a non-orthogonal grid resolving the layer interfaces and the multipoint flux approximation O-method. The numerical experiments are carried out for parameters which are typical for the production process. The static and dynamic capillary pressure-saturation relations are tested to evaluate the influence of the dynamic capillary effect.
The last part of the thesis is an investigation of the validity range of the Richards’ assumption for the two-dimensional flow model with the static capillary pressure-saturation relation. Numerical experiments show that the Richards’ assumption is not the best choice in simulating processes in the pressing section.

Recently, a new Quicksort variant due to Yaroslavskiy was chosen as standard sorting
method for Oracle's Java 7 runtime library. The decision for the change was based on
empirical studies showing that on average, the new algorithm is faster than the formerly
used classic Quicksort. Surprisingly, the improvement was achieved by using a dual pivot
approach — an idea that was considered not promising by several theoretical studies in the
past. In this thesis, I try to find the reason for this unexpected success.
My focus is on the precise and detailed average case analysis, aiming at the flavor of
Knuth's series “The Art of Computer Programming”. In particular, I go beyond abstract
measures like counting key comparisons, and try to understand the efficiency of the
algorithms at different levels of abstraction. Whenever possible, precise expected values are
preferred to asymptotic approximations. This rigor ensures that (a) the sorting methods
discussed here are actually usable in practice and (b) that the analysis results contribute to
a sound comparison of the Quicksort variants.

An isogeometric Reissner-Mindlin shell derived from the continuum theory is presented. The geometry is described by NURBS surfaces. The kinematic description of the employed shell theory requires the interpolation of the director vector and of a local basis system. Hence, the definition of nodal basis systems at the control points is necessary for the proposed formulation. The control points are in general not located on the shell reference surface and thus, several choices for the nodal values are possible. The proposed new method uses the higher continuity of the geometrical description to calculate nodal basis system and director vectors which lead to geometrical exact interpolated values thereof. Thus, the initial director vector coincides with the normal vector even for the coarsest mesh. In addition to that a more accurate interpolation of the current director and its variation is proposed. Instead of the interpolation of nodal director vectors the new approach interpolates nodal rotations. Account is taken for the discrepancy between interpolated basis systems and the individual nodal basis systems with an additional transformation. The exact evaluation of the initial director vector along with the interpolation of the nodal rotations lead to a shell formulation which yields precise results even for coarse meshes. The convergence behavior is shown to be correct for k-refinement allowing the use of coarse meshes with high orders of NURBS basis functions. This is potentially advantageous for applications with high numerical effort per integration point. The geometrically nonlinear formulation accounts for large rotations. The consistent tangent matrix is derived. Various standard benchmark examples show the superior accuracy of the presented shell formulation. A new benchmark designed to test the convergence behavior for free form surfaces is presented. Despite the higher numerical effort per integration point the improved accuracy yields considerable savings in computation cost for a predefined error bound.

A simple transformation of the Equation of Motion (EoM) allows us to directly integrate nonlinear structural models into the recursive Multibody System (MBS) formalism of SIMPACK. This contribution describes how the integration is performed for a discrete Cosserat rod model which has been developed at the ITWM. As a practical example, the run-up of a simplified three-bladed wind turbine is studied where the dynamic deformations of the three blades are calculated by the Cosserat rod model.

Dealing with information in modern times involves users to cope with hundreds of thousands of documents, such as articles, emails, Web pages, or News feeds.
Above all information sources, the World Wide Web presents information seekers with great challenges.
It offers more text in natural language than one is capable to read.
The key idea for this research intends to provide users with adaptable filtering techniques, supporting them in filtering out the specific information items they need.
Its realization focuses on developing an Information Extraction system,
which adapts to a domain of concern, by interpreting the contained formalized knowledge.
Utilizing the Resource Description Framework (RDF), which is the Semantic Web's formal language for exchanging information,
allows extending information extractors to incorporate the given domain knowledge.
Because of this, formal information items from the RDF source can be recognized in the text.
The application of RDF allows a further investigation of operations on recognized information items, such as disambiguating and rating the relevance of these.
Switching between different RDF sources allows changing the application scope of the Information Extraction system from one domain of concern to another.
An RDF-based Information Extraction system can be triggered to extract specific kinds of information entities by providing it with formal RDF queries in terms of the SPARQL query language.
Representing extracted information in RDF extends the coverage of the Semantic Web's information degree and provides a formal view on a text from the perspective of the RDF source.
In detail, this work presents the extension of existing Information Extraction approaches by incorporating the graph-based nature of RDF.
Hereby, the pre-processing of RDF sources allows extracting statistical information models dedicated to support specific information extractors.
These information extractors refine standard extraction tasks, such as the Named Entity Recognition, by using the information provided by the pre-processed models.
The post-processing of extracted information items enables representing these results in RDF format or lists, which can now be ranked or filtered by relevance.
Post-processing also comprises the enrichment of originating natural language text sources with extracted information items by using annotations in RDFa format.
The results of this research extend the state-of-the-art of the Semantic Web.
This work contributes approaches for computing customizable and adaptable RDF views on the natural language content of Web pages.
Finally, due to the formal nature of RDF, machines can interpret these views allowing developers to process the contained information in a variety of applications.

Mechanisms underlying the biological effects of coffee and its constituents are incompletely understood. Many effects have been attributed solely to caffeine, neglecting that coffee is a mixture of many chemical substances. Some authors suggest that the main mechanism of action of caffeine is to antagonize adenosine receptors (AR); a second effect is the inhibition of phosphodiesterases with the subsequent accumulation of cAMP and an intensification of the effects of catecholamines. Although the inhibition of phosphodiesterases may contribute to the actions of caffeine, there is growing evidence that most pharmacological effects of this xanthine result from antagonism of AR.
One of the main objectives of this work was to investigate whether substances other than caffeine in coffee may influence the homeostasis of intracellular cyclic nucleotides in vitro and in vivo. The influence of selected coffee compounds, extracts and brews on key elements involved in the adenosine receptor-mediated signaling pathway have been investigated.
A further aim of this work was also to determine if coffee or some coffee constituents may have a stimulatory effect on the cellular heme oxygenase activity (HO-activity). Two coffee extracts, a slightly (AB1) and an intensively roasted coffee (AB2), were studied along with selected individual compounds. Caffeine and low substituted pyrazines showed no effect on the HO-activity, while NMP, pyrazines with a greater substitution pattern such as Tetramethylpyrazine (TMP) and 2-Ethyl-3,5(6)-dimethylpyrazine (2-E-3,5-DMP) and both coffee extracts significantly induced the HO-activity in liver hepatocellular carcinoma (HepG2), intestinal colo-rectal adenocarcinoma (Caco-2) and in some instances in monocytic leukemia (MM6) cells.
It was found that caffeine, theophylline, coffee extracts from conventional or functional coffees, pyrazines (2,3-DE-6-MP, 2-Isobutyl-3-methoxyP), 5-CQA and caffeic acid all significantly inhibited the basal cytoplasmatic PDE activity in lysates of lung tumour xenograft cells (LXFL529L) and human platelets. To a somewhat lesser extent, PDE inhibition was also found in experiments performed with paraxanthine and other pyrazines (2-E-3,5-DMP, TMP and 2-E-5-MP). Thus the degree of roasting has a considerable impact on constituents of influence for PDE activity. Caffeine, coffee polyphenols and some pyrazines and further, as yet unknown roasting products appear to represent the main modulating constituents.
In two coffee intervention studies, a short-term (8 weeks) and a long-term study (24 weeks), comprising 8 and 84 healthy volunteers respectively, we examined extracellular key elements of the adenosine pathway including plasma adenosine levels and adenosine deaminase activity. Additionally, we studied the intracellular cAMP concentration and the PDE activity in platelets as surrogate biomarkers of adipocytes.
Results of in vitro experiments had suggested that the concentrations of caffeine and coffee extracts required to obtain a half maximal inhibition were in the upper range of physiological conditions. Yet, it was demonstrated for the first time in vivo that moderate consumption of coffee can modulate the activity of platelet phosphodiesterases in humans in long and short term. In both studies, the first exposure to coffee showed a strong inhibition (p<0.001) of the PDE activity in the platelet lysates of the participants while the second coffee phase showed no or a slight effect when compared with the first coffee intervention.
In both studies a significant increase (p<0.001) in intraplatelet cAMP concentrations during the wash-out phase (after the first coffee phase) was observed. This response could be due to inhibition of the PDE activity in the previously phase extending in to the wash out phase. However, the behavior of cAMP in the following study phases cannot be easily explained. It may be hypothesized that this effect is attributable to adaptive effects to allow PDE inhibition. One possibility is the modulation of the expression of membrane-bound adenosine receptors in platelet precursors, which still have a nucleus. This may potentially influence adenylate cyclase activity in mature platelets. For the observed effects, in addition to caffeine other ingredients of coffee appear to play a role. The findings suggest that monitoring of cAMP homeostasis in platelets is not a useful surrogate biomarker for effects in other tissues.
Neither the activity of adenosine deaminase nor the adenosine concentrations in plasma were markedly modulated by the coffee consumption in both trials. This may reflect the fact that adenosine is subject to quick and effective enzymatic turnover by phosphorylation (adenosine kinase) or deamination (adenosine deaminase) allowing keep its concentration within a well balanced homeostasis. However, it is also well known, that considerable variability exists in the responses to coffee drinking. In part, such variability is due to caffeine tolerance, but there is also evidence for a genetic background.
Altogether the data reported here provide further evidence for the perception that coffee consumption is associated with beneficial health effects demonstrated for the cAMP enhancement in platelets, known to counteract platelet aggregation. The effects observed for the influence of cellular heme oxygenase (HO) are in line with the well documented antioxidative activity of coffee and its constituents.

Recently convex optimization models were successfully applied for solving various problems in image analysis and restoration. In this paper, we are interested in relations between convex constrained optimization problems of the form \(min\{\Phi(x)\) subject to \(\Psi(x)\le\tau\}\) and their non-constrained, penalized counterparts \(min\{\Phi(x)+\lambda\Psi(x)\}\). We start with general considerations of the topic and provide a novel proof which ensures that a solution of the constrained problem with given \(\tau\) is also a solution of the on-constrained problem for a certain \(\lambda\). Then we deal with the special setting that \(\Psi\) is a semi-norm and \(\Phi=\phi(Hx)\), where \(H\) is a linear, not necessarily invertible operator and \(\phi\) is essentially smooth and strictly convex. In this case we can prove via the dual problems that there exists a bijective function which maps \(\tau\) from a certain interval to \(\lambda\) such that the solutions of the constrained problem coincide with those of the non-constrained problem if and only if \(\tau\) and \(\lambda\) are in the graph of this function. We illustrate the relation between \(\tau\) and \(\lambda\) by various problems arising in image processing. In particular, we demonstrate the performance of the constrained model in restoration tasks of images corrupted by Poisson noise and in inpainting models with constrained nuclear norm. Such models can be useful if we have a priori knowledge on the image rather than on the noise level.

Recently convex optimization models were successfully applied
for solving various problems in image analysis and restoration.
In this paper, we are interested in relations between
convex constrained optimization problems
of the form
\({\rm argmin} \{ \Phi(x)\) subject to \(\Psi(x) \le \tau \}\)
and their penalized counterparts
\({\rm argmin} \{\Phi(x) + \lambda \Psi(x)\}\).
We recall general results on the topic by the help of an epigraphical projection.
Then we deal with the special setting \(\Psi := \| L \cdot\|\) with \(L \in \mathbb{R}^{m,n}\)
and \(\Phi := \varphi(H \cdot)\),
where \(H \in \mathbb{R}^{n,n}\) and \(\varphi: \mathbb R^n \rightarrow \mathbb{R} \cup \{+\infty\} \)
meet certain requirements which are often fulfilled in image processing models.
In this case we prove by incorporating the dual problems
that there exists a bijective function
such that
the solutions of the constrained problem coincide with those of the
penalized problem if and only if \(\tau\) and \(\lambda\) are in the graph
of this function.
We illustrate the relation between \(\tau\) and \(\lambda\) for various problems
arising in image processing.
In particular, we point out the relation to the Pareto frontier for joint sparsity problems.
We demonstrate the performance of the
constrained model in restoration tasks of images corrupted by Poisson noise
with the \(I\)-divergence as data fitting term \(\varphi\)
and in inpainting models with the constrained nuclear norm.
Such models can be useful if we have a priori knowledge on the image rather than on the noise level.

We present the derivation of a simple viscous damping model of Kelvin–Voigt type for geometrically exact
Cosserat rods from three–dimensional continuum theory. Assuming a homogeneous and isotropic material,
we obtain explicit formulas for the damping parameters of the model in terms of the well known stiffness
parameters of the rod and the retardation time constants defined as the ratios of bulk and shear viscosities to
the respective elastic moduli. We briefly discuss the range of validity of our damping model and illustrate
its behaviour with a numerical example.

Generic layout analysis--process of decomposing document image into homogeneous regions for a collection of diverse document images--has many important applications in document image analysis and understanding such as preprocessing of degraded warped, camera-captured document images, high performance layout analysis of document images containing complex cursive scripts, and word spotting in historical document images at page level. Many areas in this field like generic text line extraction method are considered as elusive goals so far, still beyond the reach of the state-of-the-art methods [NJ07, LSZT07, KB06]. This thesis addresses this problem in such a way that it presents generic, domain-independent, text line extraction and text and non-text segmentation methods, and then describes some important applications, that were developed based on these methods. An overview of the key contributions of this thesis is as follows.
The first part of this thesis presents a generic text line extraction method using a combination of matched filtering and ridge detection techniques, which are commonly used in computer vision. Unlike the state-of-the-art text line extraction methods in the literature, the generic text line extraction method can be equally and robustly applied to a large variety of document image classes including scanned and camera-captured documents, binary and grayscale documents, typed-text and handwritten documents, historical and contemporary documents, and documents containing different scripts. Different standard datasets are selected for performance evaluation that belong to different categories of document images such as the UW-III [GHHP97] dataset of scanned documents, the ICDAR 2007 [GAS07] and the UMD [LZDJ08] datasets of handwritten documents, the DFKI-I [SB07] dataset of camera-captured documents, Arabic/Urdu script documents dataset, and German calligraphic (Fraktur) script historical documents dataset. The generic text line extraction method achieves 86% (n = 23,763 text lines in 650 documents) text line detection accuracy which is better than the aggregate accuracy of 73% of the best performing domain-specific state-of-the-art methods. To the best of the author's knowledge, it is the first general-purpose text line extraction method that can be equally used for a diverse collection of documents.
This thesis also presents an active contour (snake) based curled text line extraction method for warped, camera-captured document images. The presented approach is applied to DFKI-I [SB07] dataset of camera-captured, Latin script document images for curled text line extraction. It achieves above 95% (n = 3,091 text lines in 102 documents) text line detection accuracy, which is significantly better than the competing state-of-the-art curled text line extraction methods. The presented text line extraction method can also be applied to document images containing different scripts like Chinese, Devanagari, and Arabic after small modifications.
The second part of this thesis presents an improved version of the state-of-the-art multiresolution morphology (Leptonica) based text and non-text segmentation method [Blo91], which is a domain-independent page segmentation approach and can be equally applied to a diverse collection of binarized document images. It is demonstrated that the presented improvements result in an increase in segmentation accuracy from 93% to 99% (n = 113 documents).
This thesis also introduces a discriminative learning based approach for page segmentation, where a self-tunable multi-layer perceptron (MLP) classifier [BS10] is trained for distinguishing between text and non-text connected components. Unlike other classification based page segmentation approaches in the literature, the connected components based discriminative learning based approach is faster than pixel based classification methods and does not require a block segmentation method beforehand. A segmentation accuracy of $96\%$ ($n = 113$ documents) is achieved in comparison to the state-of-the-art multiresolution morphology (Leptonica) based page segmentation method [Blo91] that achieves a segmentation accuracy of 93%. In addition to text and non-text segmentation of Latin script documents, the presented approach can also be adapted for document images containing other scripts as well as for other specialized layout analysis tasks such as digit and non-digit segmentation [HBSB12], orientation detection [RBSB09], and body-text and side-note segmentation [BAESB12].
Finally, this thesis presents important applications of the two generic layout analysis techniques, ridge-based text line extraction method and the multi-resolution morphology based text and non-text segmentation method, discussed above. First, a complete preprocessing pipeline is described for removing different types of degradations from grayscale warped, camera-captured document images that includes removal of grayscale degradations such as non-uniform shadows and blurring through binarization, noise cleanup applying page frame detection, and document rectification using monocular dewarping. Each of these preprocessing steps shows significant improvement in comparison to the analyzed state-of-the-art methods in the literature. Second, a high performance layout analysis method is described for complex Arabic script document images written in different languages such as Arabic, Urdu, and Persian and different styles for example Naskh and Nastaliq. The presented layout analysis system is robust against different types of document image degradations and shows better performance for text and non-text segmentation, text line extraction, and reading order determination on a variety of Arabic and Urdu document images as compared to the state-of-the-art methods. It can be used for large scale Arabic and Urdu documents' digitization processes. These applications demonstrate that the layout analysis methods, ridge-based text line extraction and the multi-resolution morphology based text and non-text segmentation, are generic and can be applied easily to a large collection of diverse document images.

Filtering, Approximation and Portfolio Optimization for Shot-Noise Models and the Heston Model
(2012)

We consider a continuous time market model in which stock returns satisfy a stochastic differential equation with stochastic drift, e.g. following an Ornstein-Uhlenbeck process. The driving noise of the stock returns consists not only of Brownian motion but also of a jump part (shot noise or compound Poisson process). The investor's objective is to maximize expected utility of terminal wealth under partial information which means that the investor only observes stock prices but does not observe the drift process. Since the drift of the stock prices is unobservable, it has to be estimated using filtering techniques. E.g., if the drift follows an Ornstein-Uhlenbeck process and without
jump part, Kalman filtering can be applied and optimal strategies can be computed explicitly. Also in other cases, like for an underlying
Markov chain, finite-dimensional filters exist. But for certain jump processes (e.g. shot noise) or certain nonlinear drift dynamics explicit computations, based on discrete observations, are no longer possible or existence of finite dimensional filters is no longer valid. The same
computational difficulties apply to the optimal strategy since it depends on the filter. In this case the model may be approximated by
a model where the filter is known and can be computed. E.g., we use statistical linearization for non-linear drift processes, finite-state-Markov chain approximations for the drift process and/or diffusion approximations for small jumps in the noise term.
In the approximating models, filters and optimal strategies can often be computed explicitly. We analyze and compare different approximation methods, in particular in view of performance of the corresponding utility maximizing strategies.

Unidirectional (UD) composites are the most competitive materials for the production
of high-end structures. Their field of application spreads from the aerospace up to
automotive and general industry sector. Typical examples of components made of
unidirectional reinforced composite materials are rocket motor cases, drive shafts or
pressure vessels for hydrogen storage. The filament winding technology, the pultrusion
process and the tape placement are processes suitable for the manufacturing
using UD semi-finished products. The demand for parts made of UD composites is
constantly increasing over the last years. A key feature for the success of this technology
is the improvement of the manufacturing procedure.
Impregnation is one of the most important steps in the manufacturing process. During
this step the dry continuous fibers are combined with the liquid matrix in order to create
a fully impregnated semi-finished product. The properties of the impregnated roving
have a major effect on the laminate quality, and the efficient processing of the
liquid matrix has a big influence on the manufacturing costs.
The present work is related to the development of a new method for the impregnation
of carbon fiber rovings with thermoset resin. The developed impregnation unit consists
of a sinusoidal cavity without any moving parts. The unit in combination with an
automated resin mixing-dosing system allows complete wet-out of the fibers, precise
calibration of the resin fraction, and stable processing conditions.
The thesis focuses on the modeling of the impregnation process. Mathematical expressions
for the fiber compaction, the gradual increase of the roving tension, the
static pressure, the capillarity inside the filaments of the roving, and the fiber permeation
are presented, discussed, and experimentally verified. These expressions were
implemented in a modeling algorithm. The model takes into account all the relevant
material and process parameters. An experimental set-up based on the filament
winding process was used for the validation of the model. Trials under different conditions
have been performed. The results proved that the model can accurately simulate
the impregnation process. The good impregnation degree of the wound samples
confirmed the efficiency of the developed impregnation unit. A techno economical
analysis has proved that the developed system will result to the reduction of the
manufacturing costs and to the increase of the productivity.

Induction welding is a technique for joining of thermoplastic composites. An alternating
electromagnetic field is used for contact-free and fast heating of the parts to be
welded. In case of a suitable reinforcement structure heat generation occurs directly
in the laminate with complete heating in thickness direction in the vicinity of the coil.
The resulting temperature field is influenced by the distance to the induction coil with
decreasing temperature for increasing distance. Consequently, the surface facing the
inductor yields the highest, the opposite surface the lowest temperature.
The temperature field described significantly complicates the welding process. Due to
complete heating the laminate has to be loaded with pressure in order to prevent delamination,
which requires the usage of complex and expensive welding tools. Additionally,
the temperature difference between the inductor and the opposite side may
be greater than the processing window, which is determined by the properties of the
matrix polymer.
The induction welding process is influenced by numerous parameters. Due to complexity
process development is mainly based on experimental studies. The investigation
of parameter influences and interactions is cumbersome and the measurement
of quality relevant parameters, especially in the bondline, is difficult. Process simulation
can reduce the effort of parameter studies and contribute to further analysis of
the induction welding process.
The objective of this work is the development of a process variant of induction welding
preventing complete heating of the laminate in thickness direction. For optimal
welding the bondline has to reach the welding temperature whereas the other domains
should remain below the melting temperature of the matrix polymer.
For control of the temperature distribution localized cooling by an impinging jet of
compressed air was implemented. The effect was assessed by static heating experiments
with carbon fiber reinforced polyetheretherketone (CF/PEEK) and polyphenylenesulfide
(CF/PPS).
The application of localized cooling could influence the temperature distribution in
thickness direction of the laminate, according to the specifications of the welding
process. The temperature maximum was shifted from the inductor to the opposite side. This enables heating of the laminate to welding temperature in the bondline and
concurrently preventing melting and effects connected to this on the outer surface.
Inductive heating and the process variant with localized cooling were implemented in
three-dimensional finite-element process models. For that purpose, the finiteelement-
software Comsol Multiphysics 4.1 was used for the development of fully
coupled electromagnetic-thermal models which have been validated experimentally.
A sensitivity analysis for determination of different processing parameters of inductive
heating was conducted. The coil current, field frequency, and heat capacity were
identified as significant parameters. The cooling effect of the impinging jets was estimated
by appropriate convection coefficients.
For transfer of the developed process variant to the continuous induction welding
process, a process model was created. It represents a single overlap joint with continuous
feed. With the help of process modeling a parameter set for welding of
CF/PEEK was determined and used for joining of specimens. In doing so, the desired
temperature field was achieved and melting of the outer layers could be prevented.

The goal of this work is to develop a simulation-based algorithm, allowing the prediction
of the effective mechanical properties of textiles on the basis of their microstructure
and corresponding properties of fibers. This method can be used for optimization of the
microstructure, in order to obtain a better stiffness or strength of the corresponding fiber
material later on. An additional aspect of the thesis is that we want to take into account the microcontacts
between fibers of the textile. One more aspect of the thesis is the accounting for the thickness of thin fibers in the
textile. An introduction of an additional asymptotics with respect to a small parameter,
the relation between the thickness and the representative length of the fibers, allows a
reduction of local contact problems between fibers to 1-dimensional problems, which
reduces numerical computations significantly.
A fiber composite material with periodic microstructure and multiple frictional microcontacts
between fibers is studied. The textile is modeled by introducing small geometrical
parameters: the periodicity of the microstructure and the characteristic
diameter of fibers. The contact linear elasticity problem is considered. A two-scale
approach is used for obtaining the effective mechanical properties.
The algorithm using asymptotic two-scale homogenization for computation of the
effective mechanical properties of textiles with periodic rod or fiber microstructure
is proposed. The algorithm is based on the consequent passing to the asymptotics
with respect to the in-plane period and the characteristic diameter of fibers. This
allows to come to the equivalent homogenized problem and to reduce the dimension
of the auxiliary problems. Further numerical simulations of the cell problems give
the effective material properties of the textile.
The homogenization of the boundary conditions on the vanishing out-of-plane interface
of a textile or fiber structured layer has been studied. Introducing additional
auxiliary functions into the formal asymptotic expansion for a heterogeneous
plate, the corresponding auxiliary and homogenized problems for a nonhomogeneous
Neumann boundary condition were deduced. It is incorporated into the right hand
side of the homogenized problem via effective out-of-plane moduli.
FiberFEM, a C++ finite element code for solving contact elasticity problems, is
developed. The code is based on the implementation of the algorithm for the contact
between fibers, proposed in the thesis.
Numerical examples of homogenization of geotexiles and wovens are obtained in the
work by implementation of the developed algorithm. The effective material moduli
are computed numerically using the finite element solutions of the auxiliary contact
problems obtained by FiberFEM.

The discrete nature of the dispersed phase (swarm of droplet) in stirred and pulsed liquid-liquid extraction columns makes its mathematical modelling of such complex system a tedious task. The dispersed phase is considered as a population of droplets distributed randomly with respect to their internal properties (such as: droplet size and solute concentration) at a specific location in space. Hence, the population balance equation has been emerged as a mathematical tool to model and describe such complex behaviour. However, the resulting model is too complicated. Accordingly, the analytical solution of such a mathematical model does not exist except for particular cases. Therefore, numerical solutions are resorted to in general. This is due to the inherent nonlinearities in the convective and diffusive terms as well as the appearance of many integrals in the source term. However, modelling and simulation of liquid extraction columns is not an easy task because of the discrete nature of the dispersed phase, which consist of population of droplets. The natural frame work for taking this into account is the population balance approach.
In part of this doctoral thesis work, a rigours mathematical model based on the bivariate population balance frame work (the base of LLECMOD ‘‘Liquid-Liquid Extraction Column Module’’) for the steady state and dynamic simulation of pulsed (sieve plate and packed) liquid-liquid extraction columns is developed. The model simulates the coupled hydrodynamic and mass transfer for pulsed (packed and sieve plate) extraction columns. The model is programmed using visual digital FORTRAN and then integrated into the LLECMOD program. Within LLECMOD the user can simulate different types of extraction columns including stirred and pulsed ones. The basis of LLECMOD depends on stable robust numerical algorithms based on an extended version of a fixed pivot technique after Attarakih et al., 2003 (to take into account interphase solute transfer) and advanced computational fluid dynamics numerical methods. Experimental validated correlations are used for the estimation of the droplet terminal velocity in extraction columns based on single and swarm droplet experiments in laboratory scale devices. Additionally, recent published correlations for turbulent energy dissipation, droplet breakage and coalescence frequencies are discussed as been used in this version of LLECMOD. Moreover, coalescence model from literature derived from a stochastical description have been modified to fit the deterministic population model. As a case study, LLECMOD is used here to simulate the steady state performance of pulsed extraction columns under different operating conditions, which include pulsation intensity and volumetric flow rates are simulated. The effect of pulsation intensity (on the holdup, mean droplet diameter and solute concentration) is found to have more profound effect on systems of high interfacial tension. On the hand, the variation of volumetric flow rates have substantial effect on the holdup, mean droplet diameter and solute concentration profiles for chemical systems with low interfacial tension. Two chemical test systems recommended by the European Federation of Chemical Engineering (water-acetone (solute)-n-butyl acetate and water-acetone (solute)-toluene) and an industrial test system are used in the simulation. Model predictions are successfully validated against steady state and transient experimental data, where good agreements are achieved. The simulated results (holdup, mean droplet diameter and mass transfer profiles) compared to the experimental data show that LLECMOD is a powerful simulation tool, which can efficiently predict the dynamic and steady state performance of pulsed extraction columns.
In other part of this doctoral thesis work, the steady state performance of extraction columns is studied taking into account the effect of dispersed phase inlet condition (light or heavy phase is dispersed) and the direction of mass transfer (from continuous to dispersed phase and vice versa) using the population balance framework. LLECMOD, a program that uses multivariate population balance models, is extended to take into account the direction of mass transfer and the dispersed phase inlet. As a case study, LLECMOD is used to simulate pilot plant RDC columns where the steady state mean flow properties (dispersed phase hold up and droplet mean diameter) and the solute concentration profiles are compared to the available experimental data. Three chemical systems were used: sulpholane–benzene–n-heptane, water–acetone–toluene and water–acetone–n-butyl acetate. The dispersed phase inlet and the direction of mass transfer as well as the chemical system physical properties are found to have profound effect on the steady state performance of the RDC column. For example, the mean droplet diameter is found to persist invariant when the heavy phase is dispersed and the extractor efficiency is higher when the direction of mass transfer is from the continuous to the dispersed phase. For the purpose of experimental validation, it is found that LLECMOD predictions are in good agreement with the available experimental data concerning the dispersed phase hold up, mean droplet diameter and solute concentration profiles in both phases.
In a further part of this doctoral thesis, a mathematical model is developed for liquid extraction columns based on the multivariate population balance equation (PBE) and the primary secondary particle method (PSPM) introduced by Attarakih, 2010 (US Patent Application: 0100106467). It is extended to include the momentum balance for the dispersed phase. The advantage of momentum balance is to eliminate the need for often conflicting correlations used in estimating the terminal velocity of single and swarm of droplets. The resulting mathematical model is complex due to the integral nature of the population balance equation. To reduce the complexity of this model, while maintaining most of the information drawn from the continuous population balance equation, the concept of the PSPM is used. Based on the multivariate population balance equation and the PSPM a mathematical model is developed for any liquid extraction column. The secondary particle could be envisaged as a fluid particle carrying information about the distribution as it is evolved in space and time, in the meanwhile the primary particles carry the mean properties of the population such as total droplet concentration; mean droplet diameter dispersed phase hold up and so on. This information reflects the particle-particle interactions (breakage and coalescence) and transport (convection and diffusion). The developed model is discretized in space using a first-order upwind method, while semi-implicit first-order scheme in time is used to simulate a pilot plant RDC extraction column. Here the effect of the number of primary particles (classes) on the final predicted solution is investigated. Numerical results show that the solution converge fast even as the number of primary particle is increased. The terminal droplet velocity of the individual primary particle is found the most sensitive to the number of primary particles. Other mean population properties like the droplet mean diameter, mean hold up and the concentration profiles are also found to converge along the column height by increasing the number of primary particles. The predicted steady state profiles (droplet diameter, holdup and the concentration profiles) along a pilot RDC extraction column are compared to the experimental data where good agreement is achieved.
In addition to this a robust rigorous mathematical model based on the bivariate population balance equation is developed to predict the steady state and dynamic behaviour of the interacting hydrodynamics and mass transfer in Kühni extraction columns. The developed model is extended to include the momentum balance for the calculation of the droplet velocity. The effects of step changes in the important input variables (such as volumetric flow rates, rotational speed, inlet solute concentrations etc.) on the output variables (dispersed phase holdup, mean droplet diameter and the concentration profiles) are investigated.
The last topic of this doctoral thesis is developed to transient problems. The unsteady state analysis reveals the fact that the largest time constant (slowest response) is due to the mass transfer. On the contrary, the hydrodynamic response of the dispersed phase holdup is very fast when compared to the mass transfer due to the relative fast motion of the dispersed droplets with respect to the continuous phase. The dynamic behaviour of the dispersed and continuous phases shows a lag time that increases away from the feed points of both phases. Moreover, the solute concentration response shows a highly nonlinear behaviour due to both positive and negative step changes in the input variables. The simulation results are in good agreement with the experimental ones and show the usefulness of the model.

Wechselnde Umweltbedingungen wie Temperaturveränderungen oder der Zugang zu Nährstoffen erfordern spezielle genetische Anpassungsprogramme, vor allem von sessilen Organismen wie Pflanzen. Ein solcher hochkonservierter Mechanismus, der unter anderem vor Temperaturspitzen schützt, ist die von Hitzeschockfaktoren (HSF) kontrollierte Hitzeschockantwort (HSR). Dabei werden vermehrt spezifische Hitzestressproteine (HSPs, Chaperone) gebildet, die Proteine vor Denaturierung schützen. In Pflanzen hat sich ein hochkomplexes regulatorisches Netzwerk gebildet, das aus über 20 HSFs besteht, das eine genaue Feinabstimmung der HSR auf die jeweiligen Stressbedingungen erlaubt.
Das hohe Maß an Komplexität der HSR in Pflanzen erschwert die wissenschaftliche Zugänglichkeit jedoch erheblich. Um die grundlegenden Prinzipien der HSR in Pflanzen zu verstehen griffen wir deshalb auf einen einfacheren Modellorganismus zurück, der Pflanzen sehr nahe steht aber nur einen einzigen HSF (HSF1) enthält, der einzelligen Grünalge Chlamydomonas reinhardtii. Im Rahmen dieser Arbeit wurden dazu drei Ansätze verfolgt.
Als erstes wurden verschiedene chemische Substanzen eingesetzt die unterschiedliche Schritte während der Aktivierung und Abschaltung der HSR hemmen um darüber die Regulation der HSR aufzuklären. Dabei wurde festgestellt, dass die Phosphorylierung von HSF1 eine entscheidende Rolle in der Aktivierung der HSR spielt, das auslösende Momentum die Anhäufung von falsch gefalteten Proteinen ist und das HSP90A aus dem Cytosol eine wichtige modulierende Rolle bei der HSR spielt.
Als zweites wurde die Veränderung sämtlicher Transkripte mithilfe von Microarrays gemessen, um vor allem pflanzenspezifische Prozesse zu identifizieren, die auf erhöhte Temperaturen gezielt angepasst werden müssen. Dabei konnte die Chlorophyll Biosynthese und der Transport von Proteinen in den Chloroplasten als neue, pflanzenspezifische Ziele der Stressantwort identifiziert werden. Des Weiteren konnte direkt gezeigt werden, das HSF1 auch plastidäre Chaperone reguliert, im Gegensatz zu mitochondrialen Chaperonen die getrennt gesteuert werden.
Als letztes wurde gezielt die Expression wichtiger Gene für die Stressantwort (HSF1/HSP70B) unterdrückt, um den Einfluss dieser Gene auf die HSR genauer zu studieren. Dazu habe ich ein in der einzelligen Grünalge neuartiges System entwickelt, basierend auf dem RNAi Mechanismus, dass es erlaubt abhängig von der Stickstoffquelle im Nährmedium auch essentielle Gene gezielt auszuschalten. Dieses System erlaubte es zu zeigen, dass HSF1 selbst während des Stresses die Expression seiner RNA erhöht, und dies gezielt tut um die Stressantwort weiter zu verstärken. Es konnte weiter gezeigt werden, dass das Chloroplasten Chaperon HSP70B ein essentielles Protein für das Zellwachstum ist, welches mithilfe des induzierbaren RNAi Systems genauer untersucht werden kann. Dabei wurde festgestellt, dass die HSP70B vermittelte Assemblierung und Disassemblierung des VIPP1 Proteins entscheidend ist für dessen Funktion in der Zelle. Des Weiteren konnte gezeigt werde, dass HSP70B wahrscheinlich verantwortlich ist für die Faltung eines oder mehrerer noch unbekannter Enzyme der Arginin Biosynthese oder der Stickstofffixierung, und das diese Prozesse wahrscheinlich die essentielle Funktion von HSP70B darstellen.

Development of New Methods for the Synthesis of Aldehydes, Arenes and Trifluoromethylated Compounds
(2012)

In the 1st project, successful development of 2nd generation of a palladium catalyst for the selective hydrogenation of carboxylic acids to aldehydes was accomplished. This project was done in cooperation with Dipl. Chem. Thomas Fett from Boeringer Ingelheim, Austria. The new catalyst is highly effective for the conversion of diversely functionalized aromatic, heteroaromatic and aliphatic carboxylic acids to the corresponding aldehydes in the presence of pivalic anhydride at 5 bar hydrogen pressure, which was otherwise achieved either at 30 bar of hydrogen pressure or by using waste intensive hypophosphite bases as reducing agent. Our method has increased the synthetic importance of this valuable transformation. Selective hydrogenation of carboxylic acids to the corresponding aldehydes is now possible with industrial hydrogenation equipment as well as laboratory scale glass autoclaves. It might also convince the synthetic organic chemists to use this transformation for routine aldehyde synthesis in the laboratories.
In the 2nd project, a microwave assisted Cu-catalyzed protodecarboxylation of arenecarboxylic acids to arenes is achieved. This work was done in collaboration with Dipl. Chem. Filipe Manjolinho under the supervision of Dr. Nuria Rodríguez. In the presence of 1-5 mol% of inexpensive CuI/1,10-phenanthroline catalyst generated in situ under microwave radiations, diversely functionalized arenes and heteroarene carboxylic acids have been decarboxylated to the corresponding arenes in good yields at 190 °C in 5-15 min. The loss of volatile arenes with the release of CO2 is controled by the use of sealed high pressure resistant microwave vessels. These reactions are highly beneficial for parallel synthesis in drug discovery due to their short reaction time. Microwave technology will also help in the future to develop more effective catalysts for protodecarboxylation rections.
Based on the microwave assisted protodecarboxylation strategy, decarboxylative coupling of arenecarboxylic acids with aryl triflates and tosylates was also conducted under microwave radiation which provided higher yields of the corresponding biphenyls from deactivated substrates in short reaction time compared to the conventional heating.
In the 3rd project, crystalline, potassium (trifluoromethyl)trimethoxyborate was successfully applied for the synthesis of benzotrifluorides under the oxidative conditions. This project was done in cooperation with Dipl. Chem. Annette Buba. In the presence of Cu(OAc)2 and molecular oxygen, arylboronates were coupled with K+[CF3B(OMe)3] in DMSO at 60 °C. A variety of benzotriflurides was synthesized in good yields under the optimized reaction conditions. This protocol for the oxidative trifluoromethylation of arylboronates is the base for the development of decarboxylative trifluoromethylation reaction of arenecarboxylic acids.
The 4th project discloses the simple and straightforward synthesis of trifluoromethylated alcohols by nucleophilic addition of potassium (trifluoromethyl)trimethoxyborate to carbonyl compounds. This project was done in cooperation with Dr. Thomas Knauber and Dipl. Chem. Annette Buba. In the presence of K+[CF3B(OMe)3] in THF at 60 °C, diversely functionalized aldehydes and ketones were successfully converted into the corresponding trifluoromethylated alcohols.
The 3rd and 4th projects demonstrate the successful establishment of crystalline and shelf stable potassium (trifluoromethyl)trimethoxyborate as highly versatile CF3-source in nucleophilic trifluoromethylation reactions. These new protocols are characterized by their user-friendliness and broad applicability under mild reaction conditions, thus they are beneficial for late stage introduction of CF3-group into organic molecules.

Capital budgeting or investment decisions have an essential influence on companies’ performance. Instead of a rational choice, capital budgeting might be regarded as a process of reality construction. Research suggests that decision makers have only limited control over their own cognitive biases in this construction process. It is in this perspective that this paper intends to answer the following research question: What are behavioral determinants for a successful capital-budgeting decision process? The authors identify and discuss three behavioral success factors (reflective prudence, critical communication and outcome independence) for five stages of the capital budgeting process against the backdrop of the findings of the managerial and organizational cognition theory and cognitive psychology.

In the presented work, we make use of the strong reciprocity between kinematics and geometry to build a geometrically nonlinear, shearable low order discrete shell model of Cosserat type defined on triangular meshes, from which we deduce a rotation–free Kirchhoff type model with the triangle vertex positions as degrees of freedom. Both models behave physically plausible already on very coarse meshes, and show good
convergence properties on regular meshes. Moreover, from the theoretical side, this deduction provides a
common geometric framework for several existing models.

Granular systems in solid-like state exhibit properties like stiffness
dependence on stress, dilatancy, yield or incremental non-linearity
that can be described within the continuum mechanical framework.
Different constitutive models have been proposed in the literature either based on relations between some components of the stress tensor or on a quasi-elastic description. After a brief description of these
models, the hyperelastic law recently proposed by Jiang and Liu [1]
will be investigated. In this framework, the stress-strain relation is
derived from an elastic strain energy density where the stable proper-
ties are linked to a Drucker-Prager yield criteria. Further, a numerical method based on the finite element discretization and Newton-
Raphson iterations is presented to solve the force balance equation.
The 2D numerical examples presented in this work show that the stress
distributions can be computed not only for triangular domains, as previoulsy done in the literature, but also for more complex geometries.
If the slope of the heap is greater than a critical value, numerical instabilities appear and no elastic solution can be found, as predicted by
the theory. As main result, the dependence of the material parameter
Xi on the maximum angle of repose is established.

We consider the maximum flow problem with minimum quantities (MFPMQ), which is a variant of the maximum flow problem where
the flow on each arc in the network is restricted to be either zero or above a given lower bound (a minimum quantity), which
may depend on the arc. This problem has recently been shown to be weakly NP-complete even on series-parallel graphs.
In this paper, we provide further complexity and approximability results for MFPMQ and several special cases.
We first show that it is strongly NP-hard to approximate MFPMQ on general graphs (and even bipartite graphs) within any positive factor.
On series-parallel graphs, however, we present a pseudo-polynomial time dynamic programming algorithm for the problem.
We then study the case that the minimum quantity is the same for each arc in the network and show that, under this restriction, the problem is still
weakly NP-complete on general graphs, but can be solved in strongly polynomial time on series-parallel graphs.
On general graphs, we present a \((2 - 1/\lambda) \)-approximation algorithm for this case, where \(\lambda\) denotes the common minimum quantity of all arcs.

I report on two experiments, which were designed to test theoretical predictions about individual behavior in a duopolistic setting. With quantity being the choice variable a simultaneous Cournot game and a sequential Stackelberg game were tested over two periods. The key feature of both models was that players were able to lower marginal cost for period two if they successfully outperformed their competition in period one in terms of profit. Experimental results suggest that in the Cournot game players are very competitive in period one but become Cournot players in period two. In the Stackelberg game Cournot play is modal, suggesting that players have preferences for equality in payoffs, which maybe brought about by punishment of Stackelberg followers and fear of punishment of Stackelberg leaders . Overall, players earned more money in the Stackelberg game than in the Cournot game.

In this paper we investigate the asymptotic behaviour of the parallel volume
of fixed non-convex bodies in Minkowski spaces as the distance \(r\) tends to infinity.
We will show that the difference of the parallel volume of the convex hull of a
body and the parallel volume of the body itself can at most have order \(r^{d-2}\) in a \(d\)-dimensional space. Then we will show that in Euclidean spaces this difference can at most have order \(r^{d-3}\). These results have several applications, e.g. we will use
them to compute the derivative of \(f_\mu(rK)\) in \(r = 0\), where \(f_\mu\) is the Wills functional
or a similar functional, \(K\) is a body and \(rK\) is the Minkowski-product of \(r\) and \(K\). Finally we present applications concerning Brownian paths and Boolean models and derive new proofs for formulae for intrinsic volumes.

Thermoplastic polymer-polymer composites consist of a polymeric matrix and a
polymeric reinforcement. The combination of these materials offers outstanding
mechanical properties at lower weight than standard fiber reinforced materials.
Furthermore, when both polymeric components originate from the same family or,
ideally, from the same polymer, their sustainability degree is higher than standard
fiber reinforced composites.
A challenge of polymer-polymer composites is the subsequent processing of their
semi-finished materials by heating techniques. Since the fibers are made of meltable
thermoplastic, the reinforcing fiber structure might be lost during the heating process.
Hence, the mechanical properties of an overheated polymer-polymer composite
would decline, and finally, they would be even lower than the neat matrix. A decrease
of process temperature to manage the heating challenge is not reasonable since the
cycle time would be increased at the same time. Therefore, this work pursues the
adaption of a fast and selective heating method on the use with polymer-polymer
composites. Inductively activatable particles, so-called susceptors, were distributed in
the matrix to evoke a local heating in the matrix when being exposed to an
alternating magnetic field. In this way, the energy input to the fibers is limited.
The experimental series revealed the induction particle heating effect to be mainly
related to susceptor material, susceptor fraction, susceptor distribution as well as
magnetic field strength, coupling distance, and heating time. A proper heating was
achieved with ferromagnetic particles at a filler content of only 5 wt-% in HDPE as
well as with its respective polymer fiber reinforced composites. The study included
the analysis of susceptor impact on mechanical and thermal matrix properties as well
as a degradation evaluation. The susceptors were identified to have only a marginal
impact on matrix properties. Furthermore, a semi-empiric simulation of the particle
induction heating was applied, which served for the investigation of intrinsic melting
processes.
The achieved results, the experimental as well as the analytic study, were
successfully adapted to a thermoforming process with a polymer-polymer material,
which had been preheated by means of particle induction.

Annual Report 2011
(2012)

Annual Report, Jahrbuch AG Magnetismus

Image restoration and enhancement methods that respect important features such as edges play a fundamental role in digital image processing. In the last decades a large
variety of methods have been proposed. Nevertheless, the correct restoration and
preservation of, e.g., sharp corners, crossings or texture in images is still a challenge, in particular in the presence of severe distortions. Moreover, in the context of image denoising many methods are designed for the removal of additive Gaussian noise and their adaptation for other types of noise occurring in practice requires usually additional efforts.
The aim of this thesis is to contribute to these topics and to develop and analyze new
methods for restoring images corrupted by different types of noise:
First, we present variational models and diffusion methods which are particularly well
suited for the restoration of sharp corners and X junctions in images corrupted by
strong additive Gaussian noise. For their deduction we present and analyze different
tensor based methods for locally estimating orientations in images and show how to
successfully incorporate the obtained information in the denoising process. The advantageous
properties of the obtained methods are shown theoretically as well as by
numerical experiments. Moreover, the potential of the proposed methods is demonstrated
for applications beyond image denoising.
Afterwards, we focus on variational methods for the restoration of images corrupted
by Poisson and multiplicative Gamma noise. Here, different methods from the literature
are compared and the surprising equivalence between a standard model for
the removal of Poisson noise and a recently introduced approach for multiplicative
Gamma noise is proven. Since this Poisson model has not been considered for multiplicative
Gamma noise before, we investigate its properties further for more general
regularizers including also nonlocal ones. Moreover, an efficient algorithm for solving
the involved minimization problems is proposed, which can also handle an additional
linear transformation of the data. The good performance of this algorithm is demonstrated
experimentally and different examples with images corrupted by Poisson and
multiplicative Gamma noise are presented.
In the final part of this thesis new nonlocal filters for images corrupted by multiplicative
noise are presented. These filters are deduced in a weighted maximum likelihood
estimation framework and for the definition of the involved weights a new similarity measure for the comparison of data corrupted by multiplicative noise is applied. The
advantageous properties of the new measure are demonstrated theoretically and by
numerical examples. Besides, denoising results for images corrupted by multiplicative
Gamma and Rayleigh noise show the very good performance of the new filters.

In this work, some model reduction approaches for performing simulations
with a pseudo-2D model of Li-ion battery are presented. A full pseudo-2D model of processes in Li-ion batteries is presented following
[3], and three methods to reduce the order of the full model are considered. These are: i) directly reduce the model order using proper
orthogonal decomposition, ii) using fractional time step discretization in order to solve the equations in decoupled way, and iii) reformulation
approaches for the diffusion in the solid phase. Combinations of above
methods are also considered. Results from numerical simulations are presented, and the efficiency and the accuracy of the model reduction approaches are discussed.

The paper production is a problem with significant importance for the society
and it is a challenging topic for scientific investigations. This study is concerned
with the simulations of the pressing section of a paper machine. A two-dimensional
model is developed to account for the water flow within the pressing zone. Richards’
type equation is used to describe the flow in the unsaturated zone. The dynamic capillary
pressure–saturation relation proposed by Hassanizadeh and co-workers (Hassanizadeh
et al., 2002; Hassanizadeh, Gray, 1990, 1993a) is adopted for the paper
production process.
The mathematical model accounts for the co-existence of saturated and unsaturated
zones in a multilayer computational domain. The discretization is performed
by the MPFA-O method. The numerical experiments are carried out for parameters
which are typical for the production process. The static and dynamic capillary
pressure–saturation relations are tested to evaluate the influence of the dynamic
capillary effect.

It is often helpful to compute the intrinsic volumes of a set of which only a pixel image is observed. A computational efficient approach, which is suggested by several authors and used in practice, is to approximate the intrinsic volumes by a linear functional of the pixel configuration histogram. Here we want to examine, whether there is an optimal way of choosing this linear functional, where we will use a quite natural optimality criterion that has already been applied successfully for the estimation of the surface area. We will see that for intrinsic volumes other than volume or surface area this optimality criterion cannot be used, since estimators which ignore the data and return constant values are optimal w.r.t. this criterion. This shows that one has to be very careful, when intrinsic volumes are approximated by a linear functional of the pixel configuration histogram.

This research for this thesis was conducted to develop a framework which supports the automatic configuration of project-specific software development processes by selecting and combining different technologies: the Process Configuration Framework. The research draws attention to the problem that while the research community develops new technologies, the industrial companies continue only using their well-known ones. Because of this, technology transfer takes decades. In addition, there is the fact that there is no solution which solves all problems in a software development project. This leads to a number of technologies which need to be combined for one project.
The framework developed and explained in this research mainly addresses those problems by building a bridge between research and industry as well as by supporting software companies during the selection of the most appropriate technologies combined in a software process. The technology transformation gap is filled by a repository of (new) technologies which are used as a foundation of the Process Configuration Framework. The process is configured by providing SPEM process pattern for each technology, so that the companies can build their process by plugging into each other.
The technologies of the repository were specified in a schema including a technology model, context model, and an impact model. With context and impact it is possible to provide information about a technology, for example, its benefits to quality, cost or schedule. The offering of the process pattern as output of the Process Configuration Framework is performed in several stages:
I Technology Ranking:
1 Ranking based on Application Domain, Project & Impact
2 Ranking based on Environment
3 Ranking based on Static Context
II Technology Combination:
4 Creation of all possible Technology Chains
5 Restriction of the Technology Chains
6 Ranking based on Static and Dynamic Context
7 Extension of the Chains by Quality Assurance
III Process Configuration:
8 Process Component Diagram
9 Extension of the Process Component Diagram
10 Instantiation of the Components by Technologies of the Technology Chain
11 Providing process patterns
12 Creation of the process based on Patterns
The effectiveness and quality of the Process Configuration Framework have additionally been evaluated in a case study. Here, the Technology Chains manually created by experts were compared to the chains automatically created by the framework after it was configured by those experts. This comparison depicted that the framework results are similar and therefore can be used as a recommendation.
We conclude from our research that support during the configuration of a process for software projects is important especially for non-experts. This support is provided by the Process Configuration Framework developed in this research. In addition our research has shown that this framework offers a possibility to speed up the technology transformation gap between the research community and industrial companies.

The direction splitting approach proposed earlier in [6], aiming at the efficient solution of Navier-Stokes equations, is extended and adopted here to solve the Navier-Stokes-Brinkman equations describing incompressible flows in plain and in porous media. The resulting pressure equation is a perturbation of the
incompressibility constrained using a direction-wise factorized operator as proposed in [6]. We prove that this approach is unconditionally stable for the unsteady Navier-Stokes-Brinkman problem. We also provide numerical illustrations of the method's accuracy and efficiency.

Modern society relies on convenience services and mobile communication. Cloud computing is the current trend to make data and applications available at any time on every device. Data centers concentrate computation and storage at central locations, while they claim themselves green due to their optimized maintenance and increased energy efﬁciency. The key enabler for this evolution is the microelectronics industry. The trend to power efﬁcient mobile devices has forced this industry to change its design dogma to: ”keep data locally and reduce data communication whenever possible”. Therefore we ask: is cloud computing repeating the aberrations of its enabling industry?