Refine
Year of publication
Document Type
- Preprint (1037)
- Doctoral Thesis (928)
- Article (496)
- Report (399)
- Master's Thesis (30)
- Conference Proceeding (28)
- Diploma Thesis (24)
- Periodical Part (21)
- Working Paper (15)
- Lecture (11)
Language
- English (3025) (remove)
Keywords
- AG-RESY (47)
- PARO (25)
- Visualisierung (16)
- SKALP (15)
- Wavelet (13)
- finite element method (12)
- Case-Based Reasoning (11)
- Inverses Problem (11)
- Optimization (11)
- RODEO (11)
Faculty / Organisational entity
- Kaiserslautern - Fachbereich Mathematik (1045)
- Kaiserslautern - Fachbereich Informatik (739)
- Kaiserslautern - Fachbereich Physik (288)
- Kaiserslautern - Fachbereich Maschinenbau und Verfahrenstechnik (251)
- Fraunhofer (ITWM) (205)
- Kaiserslautern - Fachbereich Elektrotechnik und Informationstechnik (113)
- Kaiserslautern - Fachbereich Chemie (97)
- Kaiserslautern - Fachbereich Biologie (90)
- Kaiserslautern - Fachbereich Sozialwissenschaften (71)
- Kaiserslautern - Fachbereich Wirtschaftswissenschaften (31)
III/V semiconductor quantum dots (QD) are in the focus of optoelectronics research for about 25 years now. Most of the work
has been done on InAs QD on GaAs substrate. But, e.g., Ga(As)Sb (antimonide) QD on GaAs substrate/buffer have also gained
attention for the last 12 years.There is a scientific dispute on whether there is a wetting layer before antimonide QD formation, as
commonly expected for Stransky-Krastanov growth, or not. Usually ex situ photoluminescence (PL) and atomic force microscope
(AFM) measurements are performed to resolve similar issues. In this contribution, we show that reflectance anisotropy/difference
spectroscopy (RAS/RDS) can be used for the same purpose as an in situ, real-time monitoring technique. It can be employed not
only to identify QD growth via a distinct RAS spectrum, but also to get information on the existence of a wetting layer and its
thickness. The data suggest that for antimonide QD growth the wetting layer has a thickness of 1 ML (one monolayer) only.
Modern society relies on convenience services and mobile communication. Cloud computing is the current trend to make data and applications available at any time on every device. Data centers concentrate computation and storage at central locations, while they claim themselves green due to their optimized maintenance and increased energy efficiency. The key enabler for this evolution is the microelectronics industry. The trend to power efficient mobile devices has forced this industry to change its design dogma to: ”keep data locally and reduce data communication whenever possible”. Therefore we ask: is cloud computing repeating the aberrations of its enabling industry?
The plasma membrane transporter SOS1 (SALT-OVERLY SENSITIVE1) is vital for plant survival under salt stress. SOS1 activity is tightly regulated, but little is known about the underlying mechanism. SOS1 contains a cytosolic, autoinhibitory C-terminal tail (abbreviated as SOS1 C-term), which is targeted by the protein kinase SOS2 to trigger its transport activity. Here, to identify additional binding proteins that regulate SOS1 activity, we synthesized the SOS1 C-term domain and used it as bait to probe Arabidopsis thaliana cell extracts. Several 14-3-3 proteins, which function in plant salt tolerance, specifically bound to and interacted with the SOS1 C-term. Compared to wild-type plants, when exposed to salt stress, Arabidopsis plants overexpressing SOS1 C-term showed improved salt tolerance, significantly reduced Na+ accumulation in leaves, reduced induction of the salt-responsive gene WRKY25, decreased soluble sugar, starch, and proline levels, less impaired inflorescence formation and increased biomass. It appears that overexpressing SOS1 C-term leads to the sequestration of inhibitory 14-3-3 proteins, allowing SOS1 to be more readily activated and leading to increased salt tolerance. We propose that the SOS1 C-term binds to previously unknown proteins such as 14-3-3 isoforms, thereby regulating salt tolerance. This finding uncovers another regulatory layer of the plant salt tolerance program
Previously in this journal we have reported on fundamental transversemode selection (TMS#0) of broad area semiconductor lasers
(BALs) with integrated twice-retracted 4f set-up and film-waveguide lens as the Fourier-transform element. Now we choose and
report on a simpler approach for BAL-TMS#0, i.e., the use of a stable confocal longitudinal BAL resonator of length L with a
transverse constriction.The absolute value of the radius R of curvature of both mirror-facets convex in one dimension (1D) is R = L
= 2f with focal length f.The round trip length 2L = 4f againmakes up for a Fourier-optical 4f set-up and the constriction resulting
in a resonator-internal beam waist stands for a Fourier-optical low-pass spatial frequency filter. Good TMS#0 is achieved, as long
as the constriction is tight enough, but filamentation is not completely suppressed.
1. Introduction
Broad area (semiconductor diode) lasers (BALs) are intended
to emit high optical output powers (where “high” is relative
and depending on the material system). As compared to
conventional narrow stripe lasers, the higher power is distributed
over a larger transverse cross-section, thus avoiding
catastrophic optical mirror damage (COMD). Typical BALs
have emitter widths of around 100 ????m.
Thedrawback is the distribution of the high output power
over a large number of transverse modes (in cases without
countermeasures) limiting the portion of the light power in
the fundamental transverse mode (mode #0), which ought to
be maximized for the sake of good light focusability.
Thus techniques have to be used to support, prefer, or
select the fundamental transverse mode (transverse mode
selection TMS#0) by suppression of higher order modes
already upon build-up of the laser oscillation.
In many cases reported in the literature, either a BAL
facet, the
2D quantum dilaton gravitational Hamiltonian, boundary terms and new definition for total energy
(1995)
The ADM and Bondi mass for the RST model have been first discussed from Hawking and Horowitz's argument. Since there is a nonlocal term in the RST model, the RST lagrangian has to be localized so that Hawking and Horowitz's proposal can be carried out. Expressing the localized RST action in terms of the ADM formulation, the RST Hamiltonian can be derived, meanwhile keeping track of all boundary terms. Then the total boundary terms can be taken as the total energy for the RST model. Our result shows that the previous expression for the ADM and Bondi mass actually needs to be modified at quantum level, but at classical level, our mass formula can be reduced to that given by Bilal and Kogan [5] and de Alwis [6]. It has been found that there is a new contribution to the ADM and Bondi mass from the RST boundary due to the existence of the hidden dynamical field. The ADM and Bondi mass with and without the RST boundary for the static and dynamical solutions have been discussed respectively in detail, and some new properties have been found. The thunderpop of the RST model has also been encountered in our new Bondi mass formula.
This paper considers the numerical solution of a transmission boundary-value problem for the time-harmonic Maxwell equations with the help of a special finite volume discretization. Applying this technique to several three-dimensional test problems, we obtain large, sparse, complex linear systems, which are solved by using BiCG, CGS, BiCGSTAB resp., GMRES. We combine these methods with suitably chosen preconditioning matrices and compare the speed of convergence.
Destructive diseases of the lung like lung cancer or fibrosis are still often lethal. Also in case of fibrosis in the liver, the only possible cure is transplantation.
In this thesis, we investigate 3D micro computed synchrotron radiation (SR\( \mu \)CT) images of capillary blood vessels in mouse lungs and livers. The specimen show so-called compensatory lung growth as well as different states of pulmonary and hepatic fibrosis.
During compensatory lung growth, after resecting part of the lung, the remaining part compensates for this loss by extending into the empty space. This process is accompanied by an active vessel growing.
In general, the human lung can not compensate for such a loss. Thus, understanding this process in mice is important to improve treatment options in case of diseases like lung cancer.
In case of fibrosis, the formation of scars within the organ's tissue forces the capillary vessels to grow to ensure blood supply.
Thus, the process of fibrosis as well as compensatory lung growth can be accessed by considering the capillary architecture.
As preparation of 2D microscopic images is faster, easier, and cheaper compared to SR\( \mu \)CT images, they currently form the basis of medical investigation. Yet, characteristics like direction and shape of objects can only properly be analyzed using 3D imaging techniques. Hence, analyzing SR\( \mu \)CT data provides valuable additional information.
For the fibrotic specimen, we apply image analysis methods well-known from material science. We measure the vessel diameter using the granulometry distribution function and describe the inter-vessel distance by the spherical contact distribution. Moreover, we estimate the directional distribution of the capillary structure. All features turn out to be useful to characterize fibrosis based on the deformation of capillary vessels.
It is already known that the most efficient mechanism of vessel growing forms small torus-shaped holes within the capillary structure, so-called intussusceptive pillars. Analyzing their location and number strongly contributes to the characterization of vessel growing. Hence, for all three applications, this is of great interest. This thesis provides the first algorithm to detect intussusceptive pillars in SR\( \mu \)CT images. After segmentation of raw image data, our algorithm works automatically and allows for a quantitative evaluation of a large amount of data.
The analysis of SR\( \mu \)CT data using our pillar algorithm as well as the granulometry, spherical contact distribution, and directional analysis extends the current state-of-the-art in medical studies. Although it is not possible to replace certain 3D features by 2D features without losing information, our results could be used to examine 2D features approximating the 3D findings reasonably well.
The various uses of fiber-reinforced composites, for example in the enclosures of planes, boats and cars, generates the demand for a detailed analysis of these materials. The final goal is to optimize fibrous materials by the means of “virtual material design”. New fibrous materials are virtually created as realizations of a stochastic model and evaluated with physical simulations. In that way, materials can be optimized for specific use cases, without constructing expensive prototypes or performing mechanical experiments. In order to design a practically fabricable material, the stochastic model is first adapted to an existing material and then slightly modified. The virtual reconstruction of the existing material requires a precise knowledge of the geometry of its microstructure. The first part of this thesis describes a fiber quantification method by the means of local measurements of the fiber radius and orientation. The combination of a sparse chord length transform and inertia moments leads to an efficient and precise new algorithm. It outperforms existing approaches with the possibility to treat different fiber radii within one sample, with high precision in continuous space and comparably fast computing time. This local quantification method can be directly applied on gray value images by adapting the directional distance transforms on gray values. In this work, several approaches of this kind are developed and evaluated. Further characterization of the fiber system requires a segmentation of each single fiber. Using basic morphological operators with specific structuring elements, it is possible to derive a probability for each pixel describing if the pixel belongs to a fiber core in a region without overlapping fibers. Tracking high probabilities leads to a partly reconstruction of the fiber cores in non crossing regions. These core parts are then reconnected over critical regions, if they fulfill certain conditions ensuring the affiliation to the same fiber. In the second part of this work, we develop a new stochastic model for dense systems of non overlapping fibers with a controllable level of bending. Existing approaches in the literature have at least one weakness in either achieving high volume fractions, producing non overlapping fibers, or controlling the bending or the orientation distribution. This gap can be bridged by our stochastic model, which operates in two steps. Firstly, a random walk with the multivariate von Mises-Fisher orientation distribution defines bent fibers. Secondly, a force-biased packing approach arranges them in a non overlapping configuration. Furthermore, we provide the estimation of all parameters needed for the fitting of this model to a real microstructure. Finally, we simulate the macroscopic behavior of different microstructures to derive their mechanical and thermal properties. This part is mostly supported by existing software and serves as a summary of physical simulation applied to random fiber systems. The application on a glass fiber reinforced polymer proves the quality of the reconstruction by our stochastic model, as the effective properties match for both the real microstructure and the realizations of the fitted model. This thesis includes all steps to successfully perform virtual material design on various data sets. With novel and efficient algorithms it contributes to the science of analysis and modeling of fiber reinforced materials.
The fifth-generation (5G) of wireless networks promises to bring new advances, such as a huge increase in mobile data rates, a plunge in communications latency, and an increase in the quality of experience perceived by users that can cope with the ever-increasing demand in Internet traffic. However, the high cost of capital and operational expenditure (CAPEX/OPEX) of the new 5G network and the lack of a killer application hinder its rapid adoption. In this context, Mobile Network Operators (MNOs) have turned their attention to the following idea: opening up their infrastructure so that vertical businesses can leverage the new 5G network to improve their primary businesses and develop new ones. However, deploying multiple isolated vertical applications on top of the same infrastructure poses unique challenges that must be addressed. In this thesis, we provide critical contributions to developing 5G networks to accommodate different vertical applications in an isolated, flexible, and automated manner. This thesis contributions spawn on three main areas: (i) the development of an integrated fronthaul and backhaul network, (ii) the development of a network slicing overbooking algorithm, and (iii) the development of a method to mitigate the noisy neighbors' problem in a vRAN deployment.
Sensing location information in indoor scenes requires a high accuracy and is a challenging task, mainly because of multipath and NLoS (non-line-of-sight) propagation. GNSS signals cannot penetrate well in indoor environment. Satellite-based navigation and positioning systems cannot therefore be used for indoor positioning.. Other technologies have been suggested for indoor usage, among them, Wi-Fi (802.11) and 5G NR (New Radio). The primary aim of this study is to discuss the advantages and drawbacks of 5G and Wi-Fi positioning techniques for indoor localization.
This paper presents a new approach to parallel path planning for industrial robot arms with six degrees of freedom in an on-line given 3D environment. The method is based a best-first search algorithm and needs no essential off-line computations. The algorithm works in an implicitly discrete configuration space. Collisions are detected in the Cartesian workspace by hierarchical distance computation based on polyhedral models of the robot and the obstacles. By decomposing the 6D configuration space into hypercubes and cyclically mapping them onto multiple processing units, a good load distribution can be achieved. We have implemented the parallel path planner on a workstation cluster with 9 PCs and tested the planner for several benchmark environments. With optimal discretisation, the new approach usually shows very good speedups. In on-line provided environments with static obstacles, the parallel planning times are only a few seconds.
We have presented here a two-dimensional kinetical scheme for equations governing the motion of a compressible flow of an ideal gas (air) based on the Kaniel method. The basic flux functions are computed analytically and have been used in the organization of the flux computation. The algorithm is implemented and tested for the 1D shock and 2D shock-obstacle interaction problems.
In this paper a three dimensional stochastic model for the lay-down of fibers on a moving conveyor belt in the production process of nonwoven materials is derived. The model is based on stochastic diferential equations describing the resulting position of the fiber on the belt under the influence of turbulent air ows. The model presented here is an extension of an existing surrogate model, see [6, 3].
The World Wide Web is a medium through which a manufacturer may allow Internet visitors to customize or compose his products. Due to missing or rapidly changing standards these applications are often restricted to relatively simple CGI or JAVA based scripts. Usually, results like images or movies are stored in a database and are transferred on demand to the web-user. Viper (Visualisierung parametrisch editierbarer Raumkomponenten) is a Toolkit [VIP96] written in C++ and JAVA which provides 3D-modeling and visualization methodsfor developing complex web-based applications. The Toolkit has been designed to built a prototype, which can be used to construct and visualize prefabricated homes on the Internet. Alternative applications are outlined in this paper. Within Viper, all objects are stored in a scene graph (VSSG ), which is the basic data structure of the Toolkit. To show the concept and structure of the Toolkit, functionality, and implementation of the prototype are described.
The classic approach in robust optimization is to optimize the solution with respect to the worst case scenario. This pessimistic approach yields solutions that perform best if the worst scenario happens, but also usually perform bad on average. A solution that optimizes the average performance on the other hand lacks in worst-case performance guarantee.
In practice it is important to find a good compromise between these two solutions. We propose to deal with this problem by considering it from a bicriteria perspective. The Pareto curve of the bicriteria problem visualizes exactly how costly it is to ensure robustness and helps to choose the solution with the best balance between expected and guaranteed performance.
Building upon a theoretical observation on the structure of Pareto solutions for problems with polyhedral feasible sets, we present a column generation approach that requires no direct solution of the computationally expensive worst-case problem. In computational experiments we demonstrate the effectivity of both the proposed algorithm, and the bicriteria perspective in general.
We consider the problem of evacuating a region with the help of buses. For a given set of possible collection points where evacuees gather, and possible shelter locations where evacuees are brought to, we need to determine both collection points and shelters we would like to use, and bus routes that evacuate the region in minimum time.
We model this integrated problem using an integer linear program, and present a branch-cut-and-price algorithm that generates bus tours in its pricing step. In computational experiments we show that our approach is able to solve instances of realistic size in sufficient time for practical application, and considerably outperforms the usage of a generic ILP solver.
A building-block model reveals new insights into the biogenesis of yeast mitochondrial ribosomes
(2020)
Most of the mitochondrial proteins in yeast are encoded in the nuclear genome, get synthesized by cytosolic ribosomes and are imported via TOM and TIM23 into the matrix or other subcompartments of mitochondria. The mitochondrial DNA in yeast however also encodes a small set of 8 proteins from which most are hydrophobic membrane proteins and build core components of the OXPHOS complexes. They get synthesized by mitochondrial ribosomes which are descendants of bacterial ribosomes and still have some similarities to them. On the other hand, mitochondrial ribosomes experienced various structural and functional changes during evolution that specialized them for the synthesis of the mitochondrial encoded membrane proteins. The mitoribosome contains mitochondria-specific ribosomal proteins and replaced the bacterial 5S rRNA by mitochondria-specific proteins and rRNA extensions. Furthermore, the mitoribosome is tethered to the inner mitochondrial membrane to facilitate a co-translational insertion of newly synthesized proteins. Thus, also the assembly process of mitoribosomes differs from that of bacteria and is to date not well understood.
Therefore, the biogenesis of mitochondrial ribosomes in yeast should be investigated. To this end, a strain was generated in which the gene of the mitochondrial RNA-polymerase RPO41 is under control of an inducible GAL10-promoter. Since the scaffold of ribosomes is built by ribosomal RNAs, the depletion of the RNA-polymerase subsequently leads to a loss of mitochondrial ribosomes. Reinduction of Rpo41 initiates the assembly of new mitoribosomes, which makes this strain an attractive model to study mitoribosome biogenesis.
Initially, the effects of Rpo41 depletion on cellular and mitochondrial physiology was investigated. Upon Rpo41 depletion, growth on respiratory glycerol medium was inhibited. Furthermore, mitochondrial ribosomal 21S and 15S rRNA was diminished and mitochondrial translation was almost completely absent. Also, mitochondrial DNA was strongly reduced due to the fact that mtDNA replication requires RNA primers that get synthesized by Rpo41.
Next, the effect of reinduction of Rpo41 on mitochondria was tested. Time course experiments showed that mitochondrial translation can partially recover from 48h Rpo41 depletion within a timeframe of 4.5h. Sucrose gradient sedimentation experiments further showed that the mitoribosomal constitution was comparable to wildtype control samples during the time course of 4.5h of reinduction, suggesting that the ribosome assembly is not fundamentally altered in Gal-Rpo41 mitochondria. In addition, the depletion time was found to be critical for recovery of mitochondrial translation and mitochondrial RNA levels. It was observed that after 36h of Rpo41 depletion, the rRNA levels and mitochondrial translation recovered to almost 100%, but only within a time course of 10h.
Finally, mitochondria from Gal-Rpo41 cells isolated after different timepoints of reinduction were used to perform complexome profiling and the assembly of mitochondrial protein complexes was investigated. First, the steady state conditions and the assembly process of mitochondrial respiratory chain complexes were monitored. The individual respiratory chain complexes and the super-complexes of complex III, complex IV and complex V were observed. Furthermore, it was seen that they recovered from Rpo41 depletion within 4.5h of reinduction. Complexome profiles of the mitoribosomal small and large subunit discovered subcomplexes of mitoribosomal proteins that were assumed to form prior to their incorporation into assembly intermediates. The complexome profiles after reinduction indeed showed the formation of these subcomplexes before formation of the fully assembled subunit. In the mitochondrial LSU one subcomplex builds the membrane facing protuberance and a second subcomplex forms the central protuberance. In contrast to the preassembled subcomplexes, proteins that were involved in early assembly steps were exclusively found in the fully assembled subunit. Proteins that assemble at the periphery of the mitoribosome during intermediate and late assembly steps where found in soluble form suggesting a pool of unassembled proteins that supply assembly intermediates with proteins.
Taken together, the findings of this thesis suggest a so far unknow building-block model for mitoribosome assembly in which characteristic structures of the yeast mitochondrial ribosome form preassembled subcomplexes prior to their incorporation into the mitoribosome.
3D integration of solid-state memories and logic, as demonstrated by the Hybrid Memory Cube (HMC), offers major opportunities for revisiting near-memory computation and gives new hope to mitigate the power and performance losses caused by the “memory wall”. In this paper we present the first exploration steps towards design of the Smart Memory Cube (SMC), a new Processor-in-Memory (PIM) architecture that enhances the capabilities of the logic-base (LoB) in HMC. An accurate simulation environment has been developed, along with a full featured software stack. All offloading and dynamic overheads caused by the operating system, cache coherence, and memory management are considered, as well. Benchmarking results demonstrate up to 2X performance improvement in comparison with the host SoC, and around 1.5X against a similar host-side accelerator. Moreover, by scaling down the voltage and frequency of PIM’s processor it is possible to reduce energy by around 70% and 55% in comparison with the host and the accelerator, respectively.
Contrary to symbolic learning approaches, which represent a learned concept explicitly, case-based approaches describe concepts implicitly by a pair (CB; sim), i.e. by a measure of similarity sim and a set CB of cases. This poses the question if there are any differences concerning the learning power of the two approaches. In this article we will study the relationship between the case base, the measure of similarity, and the target concept of the learning process. To do so, we transform a simple symbolic learning algorithm (the version space algorithm) into an equivalent case- based variant. The achieved results strengthen the hypothesis of the equivalence of the learning power of symbolic and case-based methods and show the interdependency between the measure used by a case-based algorithm and the target concept.
Retrieving multiple cases is supposed to be an adequate retrieval strategy for guiding partial-order planners because of the recognized flexibility of these planners to interleave steps in the plans. Cases are combined by merging them. In this paper, we will examine two different kinds of merging cases in the context of partial-order planning. We will see that merging cases can be very difficult if the cases are merged eagerly. On the other hand, if cases are merged by avoiding redundant steps, the guidance of the additional cases tends to decrease with the number of covered goals and retrieved cases in domains having a certain kind of interactions. Thus, to retrieve a single case covering many of the goals of the problem or to retrieve fewer cases covering many of the goals is at least equally effective as to retrieve several cases covering all goals in these domains.