### Refine

#### Year of publication

- 2017 (51) (remove)

#### Document Type

- Doctoral Thesis (42)
- Preprint (3)
- Periodical Part (2)
- Diploma Thesis (1)
- Lecture (1)
- Other (1)
- Working Paper (1)

#### Language

- English (51) (remove)

#### Keywords

- A/D conversion (1)
- ADAS (1)
- AFDX (1)
- Ableitungsfreie Optimierung (1)
- Accountability (1)
- Achslage (1)
- Anion recognition (1)
- Automation (1)
- Backlog (1)
- Beschränkte Krümmung (1)

#### Faculty / Organisational entity

The detection and characterisation of undesired lead structures on shaft surfaces is a concern in production and quality control of rotary shaft lip-type sealing systems. The potential lead structures are generally divided into macro and micro lead based on their characteristics and formation. Macro lead measurement methods exist and are widely applied. This work describes a method to characterise micro lead on ground shaft surfaces. Micro lead is known as the deviation of main orientation of the ground micro texture from circumferential direction. Assessing the orientation of microscopic structures with arc minute accuracy with regard to circumferential direction requires exact knowledge of both the shaft’s orientation and the direction of surface texture. The shaft’s circumferential direction is found by calibration. Measuring systems and calibration procedures capable of calibrating shaft axis orientation with high accuracy and low uncertainty are described. The measuring systems employ areal-topographic measuring instruments suited for evaluating texture orientation. A dedicated evaluation scheme for texture orientation is based on the Radon transform of these topographies and parametrised for the application. Combining the calibration of circumferential direction with the evaluation of texture orientation the method enables the measurement of micro lead on ground shaft surfaces.

Divide-and-Conquer is a common strategy to manage the complexity of system design and verification. In the context of System-on-Chip (SoC) design verification, an SoC system is decomposed into several modules and every module is separately verified. Usually an SoC module is reactive: it interacts with its environmental modules. This interaction is normally modeled by environment constraints, which are applied to verify the SoC module. Environment constraints are assumed to be always true when verifying the individual modules of a system. Therefore the correctness of environment constraints is very important for module verification.
Environment constraints are also very important for coverage analysis. Coverage analysis in formal verification measures whether or not the property set fully describes the functional behavior of the design under verification (DuV). if a set of properties describes every functional behavior of a DuV, the set of properties is called complete. To verify the correctness of environment constraints, Assume-Guarantee Reasoning rules can be employed.
However, the state of the art assume-guarantee reasoning rules cannot be applied to the environment constraints specified by using an industrial standard property language such as SystemVerilog Assertions (SVA).
This thesis proposes a new assume-guarantee reasoning rule that can be applied to environment constraints specified by using a property language such as SVA. In addition, this thesis proposes two efficient plausibility checks for constraints that can be conducted without a concrete implementation of the considered environment.
Furthermore, this thesis provides a compositional reasoning framework determining that a system is completely verified if all modules are verified with Complete Interval Property Checking (C-IPC) under environment constraints.
At present, there is a trend that more of the functionality in SoCs is shifted from the hardware to the hardware-dependent software (HWDS), which is a crucial component in an SoC, since other software layers, such as the operating systems are built on it. Therefore there is an increasing need to apply formal verification to HWDS, especially for safety-critical systems.
The interactions between HW and HWDS are often reactive, and happen in a temporal order. This requires new property languages to specify the reactive behavior at the HW and SW interfaces.
This thesis introduces a new property language, called Reactive Software Property Language (RSPL), to specify the reactive interactions between the HW and the HWDS.
Furthermore, a method for checking the completeness of software properties, which are specified by using RSPL, is presented in this thesis. This method is motivated by the approach of checking the completeness of hardware properties.

In the present work the concept of decarboxylative couplings and the strategy to use carboxylates as directing groups for C-H functionalizations have been decisively improved in three ways. These concepts emphasize the multifaceted nature of aromatic carboxylic acids as expedient starting materials in homogeneous catalysis to construct highly desirable molecular scaffolds in a straightforward fashion.
In the first project, the restriction of decarboxylative biaryl synthesis to exclusively couple aryl halides with ortho-substituted benzoic acids has been overcome by a holistic optimization of a Cu/Pd bimetallic catalyst system. Long ago postulated, this is now the proof that decarboxylative cross-couplings are not intrinsically limited to different decarboxylation propensities of benzoic acids or hampered by excess halides, accessing for the first time the entire spectrum of aromatic carboxylic acids as starting materials for the decarboxylative biaryl synthesis. The second project uses the carboxyl moiety as directing group for the ortho-arylation with aryl bromides and -chlorides catalyzed by comparatively inexpensive ruthenium. The carboxylic acid group remains untouched after the ortho-functionalization giving the possibility to a wealth of further diversifications via decarboxylative ipso-substitutions. Within the same project, a Cu/Ru bimetallic catalyst system was found to be able to switch the decarboxylative biaryl coupling from the ipso- to the ortho-position, complementing the Cu/Pd system developed in the first project. In a third project, a redox neutral C-C bond formation revealed the full synthetic potential of the carboxyl group. The COOH moiety acts as a classical directing group for the C-H hydroarylation of internal alkynes to form highly desirable 2-vinyl benzoic acids. With propargylic alcohols the hydroarylation is followed by an in situ esterification, showing that after easing the C-H cleavage, the directing group can be transformed into another functional group, thus, acting as a transformable directing group. Most importantly, a new fascinating reaction mode is activated by embedding the decarboxylation within the C-H functionalization event. This mode of action is capable to solve regioselectivity issues that inherently occur when dealing with carboxylates as directing groups. A so-called deciduous directing group is cast off simultaneously within the C-H functionalization event, resulting in an inherently monoselective pathway.
These methods were developed with the permanent goal of ensuring high sustainability. They do require neither pre-functionalized starting materials nor additional oxidants and provide access to a number of chemically relevant molecules from abundant, inexpensive and toxicologically innocuous educts.

We continue in this paper the study of k-adaptable robust solutions for combinatorial optimization problems with bounded uncertainty sets. In this concept not a single solution needs to be chosen to hedge against the uncertainty. Instead one is allowed to choose a set of k different solutions from which one can be chosen after the uncertain scenario has been revealed. We first show how the problem can be decomposed into polynomially many subproblems if k is fixed. In the remaining part of the paper we consider the special case where k=2, i.e., one is allowed to choose two different solutions to hedge against the uncertainty. We decompose this problem into so called coordination problems. The study of these coordination problems turns out to be interesting on its own. We prove positive results for the unconstrained combinatorial optimization problem, the matroid maximization problem, the selection problem, and the shortest path problem on series parallel graphs. The shortest path problem on general graphs turns out to be NP-complete. Further, we present for minimization problems how to transform approximation algorithms for the coordination problem to approximation algorithms for the original problem. We study the knapsack problem to show that this relation does not hold for maximization problems in general. We present a PTAS for the corresponding coordination problem and prove that the 2-adaptable knapsack problem is not at all approximable.

We extend the standard concept of robust optimization by the introduction of an alternative solution. In contrast to the classic concept, one is allowed to chose two solutions from which the best can be picked after the uncertain scenario has been revealed. We focus in this paper on the resulting robust problem for combinatorial problems with bounded uncertainty sets. We present a reformulation of the robust problem which decomposes it into polynomially many subproblems. In each subproblem one needs to find two solutions which are connected by a cost function which penalizes if the same element is part of both solutions. Using this reformulation, we show how the robust problem can be solved efficiently for the unconstrained combinatorial problem, the selection problem, and the minimum spanning tree problem. The robust problem corresponding to the shortest path problem turns out to be NP-complete on general graphs. However, for series-parallel graphs, the robust shortest path problem can be solved efficiently. Further, we show how approximation algorithms for the subproblem can be used to compute approximate solutions for the original problem.

In this thesis viscoelastic material models are established to investigate the nature of continuous calving processes at Antarctic ice shelves. Physics-based descriptions of calving require appropriate fracture criteria to separate icebergs from the remaining ice shelf. Hence, criteria of the stress, the strain, and the self-similarity criterion are considered within finite-element computations. Crucial parameters in the models to determine the position of calving are the accurate knowledge of the geometry, especially the freeboard height, while the material parameters mainly influence the time span between two successive calving events. The extension to nonlinear material models is necessary to properly analyze the internal forces also for large deformations that occur for longer times of the viscous ice flow.

This thesis presents research studies on the fundamental interplay of diatomic molecules with transition metal compounds under cryogenic conditions. The utilized setup offers a multitude of opportunities to study isolated ions: The ions can either be generated by an ElectroSpray Ionization (ESI) source or a Laser VAPorization (LVAP) cluster ion source. The setup facilitates kinetic investigations of the ions with different reaction gases under well-defined isothermal conditions. Moreover it enables cryo InfraRed (Multiple) Photon Dissociation (IR-(M)PD) spectroscopy in combination with tunable OPO/OPA laser systems. In conjunction with density functional theory (DFT) modelling, the IR(M)-PD spectra allow for an assignment of geometric minimum structures. Furthermore DFT modelling helps to identify possible reaction pathways. Altogether the presented methods allow to gain fundamental insights into molecular structures and reactivity of the investigated systems.
The first part of this thesis focuses on the interplay of N2 with different transition metal clusters (Con+, Nin+, and Fen+) by cryo IR spectroscopy and cryo kinetics. In conjunction with DFT modelling the N2 coordination was elucidated (Con+), structures were assigned (Nin+), the concept of structure related surface adsorption behavior was introduced (Nin+), and the a first explanation for the inertness if Fe17+ was given (Fen+). Furthermore this thesis provides for a case study on the coadsorption of H2 and N2 on Ru8+ that elucidates the H migration on the Ru cluster. The last part of the thesis addresses the IR spectra of in vacuo generated [Hemin]+ complexes with N2, O2, and CO. Structures and spin states were assigned with the help of DFT modelling.

Nonwoven materials are used as filter media which are the key component of automotive filters such as air filters, oil filters, and fuel filters. Today, the advanced engine technologies require innovative filter media with higher performances. A virtual microstructure of the nonwoven filter medium, which has similar filter properties as the existing material, can be used to design new filter media from existing media. Nonwoven materials considered in this thesis prominently feature non-overlapping fibers, curved fibers, fibers with circular cross section, fibers of apparently infinite length, and fiber bundles. To this end, as part of this thesis, we extend the Altendorf-Jeulin individual fiber model to incorporate all the above mentioned features. The resulting novel stochastic 3D fiber model can generate geometries with good visual resemblance of real filter media. Furthermore, pressure drop, which is one of the important physical properties of the filter, simulated numerically on the computed tomography (CT) data of the real nonwoven material agrees well (with a relative error of 8%) with the pressure drop simulated in the generated microstructure realizations from our model.
Generally, filter properties for the CT data and generated microstructure realizations are computed using numerical simulations. Since numerical simulations require extensive system memory and computation time, it is important to find the representative domain size of the generated microstructure for a required filter property. As part of this thesis, simulation and a statistical approach are used to estimate the representative domain size of our microstructure model. Precisely, the representative domain size with respect to the packing density, the pore size distribution, and the pressure drop are considered. It turns out that the statistical approach can be used to estimate the representative domain size for the given property more precisely and using less generated microstructures than the purely simulation based approach.
Among the various properties of fibrous filter media, fiber thickness and orientation are important characteristics which should be considered in design and quality assurance of filter media. Automatic analysis of images from scanning electron microscopy (SEM) is a suitable tool in that context. Yet, the accuracy of such image analysis tools cannot be judged based on images of real filter media since their true fiber thickness and orientation can never be known accurately. A solution is to employ synthetically generated models for evaluation. By combining our 3D fiber system model with simulation of the SEM imaging process, quantitative evaluation of the fiber thickness and orientation measurements becomes feasible. We evaluate the state-of-the-art automatic thickness and orientation estimation method that way.

For many years, most distributed real-time systems employed data communication systems specially tailored to address the specific requirements of individual domains: for instance, Controlled Area Network (CAN) and Flexray in the automotive domain, ARINC 429 [FW10] and TTP [Kop95] in the aerospace domain. Some of these solutions were expensive, and eventually not well understood.
Mostly driven by the ever decreasing costs, the application of such distributed real-time system have drastically increased in the last years in different domains. Consequently, cross-domain communication systems are advantageous. Not only the number of distributed real-time systems have been increasing but also the number of nodes per system, have drastically increased, which in turn increases their network bandwidth requirements. Further, the system architectures have been changing, allowing for applications to spread computations among different computer nodes. For example, modern avionics systems moved from federated to integrated modular architecture, also increasing the network bandwidth requirements.
Ethernet (IEEE 802.3) [iee12] is a well established network standard. Further, it is fast, easy to install, and the interface ICs are cheap [Dec05]. However, Ethernet does not offer any temporal guarantee. Research groups from academia and industry have presented a number of protocols merging the benefits of Ethernet and the temporal guarantees required by distributed real-time systems. Two of these protocols are: Avionics Full-Duplex Switched Ethernet (AFDX) [AFD09] and Time-Triggered Ethernet (TTEthernet) [tim16]. In this dissertation, we propose solutions for two problems faced during the design of AFDX and TTEthernet networks: avoiding data loss due to buffer overflow in AFDX networks with multiple priority traffic, and scheduling of TTEthernet networks.
AFDX guarantees bandwidth separation and bounded transmission latency for each communication channel. Communication channels in AFDX networks are not synchronized, and therefore frames might compete for the same output port, requiring buffering to avoid data loss. To avoid buffer overflow and the resulting data loss, the network designer must reserve a safe, but not too pessimistic amount of memory of each buffer. The current AFDX standard allows for the classification of the network traffic with two priorities. Nevertheless, some commercial solutions provide multiple priorities, increasing the complexity of the buffer backlog analysis. The state-of-the-art AFDX buffer backlog analysis does not provide a method to compute deterministic upper bounds
iiifor buffer backlog of AFDX networks with multiple priority traffic. Therefore, in this dissertation we propose a method to address this open problem. Our method is based on the analysis of the largest busy period encountered by frames stored in a buffer. We identify the ingress (and respective egress) order of frames in the largest busy period that leads to the largest buffer backlog, and then compute the respective buffer backlog upper bound. We present experiments to measure the computational costs of our method.
In TTEthernet, nodes are synchronized, allowing for message transmission at well defined points in time, computed off-line and stored in a conflict-free scheduling table. The computation of such scheduling tables is a NP-complete problem [Kor92], which should be solved in reasonable time for industrial size networks. We propose an approach to efficiently compute a schedule for the TT communication channels in TTEthernet networks, in which we model the scheduling problem as a search tree. As the scheduler traverses the search tree, it schedules the communication channels on a physical link. We presented two approaches to traverse the search tree while progressively creating the vertices of the search tree. A valid schedule is found once the scheduler reaches a valid leaf. If on the contrary, it reaches an invalid leaf, the scheduler backtracks searching for a path to a valid leaf. We present a set of experiments to demonstrate the impact of the input parameters on the time taken to compute a feasible schedule or to deem the set of virtual links infeasible.

The proliferation of sensors in everyday devices – especially in smartphones – has led to crowd sensing becoming an important technique in many urban applications ranging from noise pollution mapping or road condition monitoring to tracking the spreading of diseases. However, in order to establish integrated crowd sensing environments on a large scale, some open issues need to be tackled first. On a high level, this thesis concentrates on dealing with two of those key issues: (1) efficiently collecting and processing large amounts of sensor data from smartphones in a scalable manner and (2) extracting abstract data models from those collected data sets thereby enabling the development of complex smart city services based on the extracted knowledge.
Going more into detail, the first main contribution of this thesis is the development of methods and architectures to facilitate simple and efficient deployments, scalability and adaptability of crowd sensing applications in a broad range of scenarios while at the same time enabling the integration of incentivation mechanisms for the participating general public. During an evaluation within a complex, large-scale environment it is shown that real-world deployments of the proposed data recording architecture are in fact feasible. The second major contribution of this thesis is the development of a novel methodology for using the recorded data to extract abstract data models which are representing the inherent core characteristics of the source data correctly. Finally – and in order to bring together the results of the thesis – it is demonstrated how the proposed architecture and the modeling method can be used to implement a complex smart city service by employing a data driven development approach.