Refine
Year of publication
- 2000 (96) (remove)
Document Type
- Preprint (74)
- Article (12)
- Report (7)
- Diploma Thesis (1)
- Doctoral Thesis (1)
- Periodical Part (1)
Language
- English (96) (remove)
Keywords
- resonances (6)
- Quantum mechanics (5)
- lifetimes (5)
- Wannier-Stark systems (4)
- AG-RESY (3)
- HANDFLEX (3)
- branch and cut (3)
- facets (3)
- facility location (3)
- hub location (3)
Faculty / Organisational entity
- Kaiserslautern - Fachbereich Physik (46)
- Kaiserslautern - Fachbereich Mathematik (33)
- Kaiserslautern - Fachbereich Informatik (10)
- Fraunhofer (ITWM) (4)
- Kaiserslautern - Fachbereich Elektrotechnik und Informationstechnik (1)
- Kaiserslautern - Fachbereich Maschinenbau und Verfahrenstechnik (1)
- Kaiserslautern - Fachbereich Wirtschaftswissenschaften (1)
In this paper we show how Metropolis Light Transport can be extended both in the underlying theoretical framework and the algorithmic implementation to incorporate volumetric scattering.
We present a generalization of the path integral formulation thathandles anisotropic scattering in non-homogeneous media. Based on this framework we introduce a new mutation strategy that is
specifically designed for participating media. It exploits the locality of light propagation by perturbing certain interaction points within the medium. To efficiently sample inhomogeneous media a new ray marching method has been developed that avoids aliasing artefacts and is significantly faster than stratified sampling. The resulting global illumination algorithm provides a physically correct simulation of light transport in the presence of participating media that includes effects such as volume caustics and multiple volume scattering. It is not restricted to certain classes of geometry and scattering models and has minimal memory requirements. Furthermore, it is unbiased and robust, in the sense that it produces satisfactory results for a wide range of input scenes and lighting situations within acceptable time bounds. In particular, we found that it is weil suited for complex scenes with many light sources.
For most applications the used transport service providers are predetermined during the development of the application. This makes it difficult to consider the application communication requirements and to exploit specific features of the network technology. Specialized protocols that are more efficient and offer a qualitative improved service are typically not supported by most applications because they are not commonly available. In this paper we propose a concept for the realization of protocol independent transport services. Only a transport service is predetermined during the development of the application and an appropriate transport service provider is dynamically selected at run time. This enables to exploit specialized protocols if possible, but standard protocols could still be used if necessary. The main focus of this paper is how a transport service could provide a new transport service provider transparently to existing applications. A prototype is presented that maps TCP/IP based applications to an ATM specific transport service provider which offers a reliable and unreliable transport service like TCP/IP.
The photon map provides a powerful tool for approximating the irradiance in global illumination computations independent from geometry. By presenting new importance sampling techniques, we dramatically improve the memory footprint of the photon map, simplify the caustic generation, and allow for a much faster sampling of direct illumination in complicated models as they arise in a production environment.
The quality of freeform surfaces is one of the major topics of CAD/CAM. Aesthetic and technical demands require the construction of high quality surfaces with strong shape conditions. Quality diminishing properties like dents or flat points have to be eliminated while approximation conditions must hold at the same time. Our approach combines quality and approximation criteria to a nonlinear multicriteria optimization problem and achieves an automatic approximation and fitting process.
Given a public transportation system represented by its stops and direct connections between stops, we consider two problems dealing with the prices for the customers: The fare problem in which subsets of stops are already aggregated to zones and "good" tariffs have to be found in the existing zone system. Closed form solutions for the fare problem are presented for three objective functions. In the zone problem the design of the zones is part of the problem. This problem is NP hard and we therefore propose three heuristics which prove to be very successful in the redesign of one of Germany's transportation systems
We examine the feasibility polyhedron of the uncapacitated hub location problem (UHL) with multiple allocation, which has applications in the fields of air passenger and cargo transportation, telecommunication and postal delivery services. In particular we determine the dimension and derive some classes of facets of this polyhedron. We develop some general rules about lifting facets from the uncapacitated facility location (UFL) for UHL and projecting facets from UHL to UFL. By applying these rules we get a new class of facets for UHL which dominates the inequalities in the original formulation. Thus we get a new formulation of UHL whose constraints are all facet–defining. We show its superior computational performance by benchmarking it on a well known data set.
Within this paper we review image distortion measures. A distortion measure is a criterion that assigns a "quality number" to an image. We distinguish between mathematical distortion measures and those distortion measures in-cooperating a priori knowledge about the imaging devices ( e.g. satellite images), image processing algorithms or the human physiology. We will consider representative examples of different kinds of distortion measures and are going to discuss them.
In this short note we prove some general results on semi-stable sheaves on P_2 and P_3 with arbitrary linear Hilbert polynomial. Using Beilinson's spectral sequence, we compute free resolutions for this class of semi-stable sheaves and deduce that the smooth moduli spaces M_{r m + s}(P_2) and M_{r m + r - s}(P_2) are birationally equivalent if r and s are coprime.
Power-ordered sets are not always lattices. In the case of distributive lattices we give a description by disjoint of chains. Finite power-ordered sets have a polarity. We introduct the leveled lattices and show examples with trivial tolerance. Finally we give a list of Hasse diagrams of power-ordered sets.
Abstract: This paper presents a solution to a problem from superanalysis about the existence of Hilbert-Banach superalgebras. Two main results are derived: 1) There exist Hilbert norms on some graded algebras (infinite-dimensional superalgebras included) with respect to which the multiplication is continuous. 2) Such norms cannot be chosen to be submultiplicative and equal to one on the unit of the algebra.
Abstract: Local field effects on the rate of spontaneous emission and Lamb shift in a dense gas of atoms are discussed taking into account correlations of atomic center-of-mass coordinates. For this the exact retarded propagator in the medium is calculated in independent scattering approximation and employing a virtual-cavity model. The resulting changes of the atomic polarizability lead to modi cations of the medium response which can be of the same order of magnitude but of opposite sign than those due to local field corrections of the dielectric function derived by Morice, Castin, and Dalibard [Phys.Rev.A 51, 3896 (1995)].
Abstract: We identify form-stable coupled excitations of light and matter ("dark-state polaritons") associated with the propagation of quantum fields in Electromagnetically Induced Transparency. The properties of the dark-state polaritons such as the group velocity are determined by the mixing angle between light and matter components and can be controlled by an external coherent field as the pulse propagates. In particular, light pulses can be decelerated and "trapped" in which case their shape and quantum state are mapped onto metastable collective states of matter. Possible applications of this reversible coherent-control technique are discussed.
Abstract: We analyze systematic (classical) and fundamental (quantum) limitations of the sensitivity of optical magnetometers resulting from ac-Stark shifts. We show that incontrast to absorption-based techniques, the signal reduction associated with classical broadening can be compensated in magnetometers based on phase measurements using electromagnetically induced transparency (EIT). However due to ac-Stark associated quantum noise the signal-to-noise ratio of EIT-based magnetometers attains a maximum value at a certain laser intensity. This value is independent on the quantum statistics of the light and defines a standard quantum limit of sensitivity. We demonstrate that an EIT-based optical magnetometer in Faraday configuration is the best candidate to achieve the highest sensitivity of magnetic field detection and give a detailed analysis of such a device.
Abstract: We analyze the above-threshold behavior of a mirrorless parametric oscillator based on resonantly enhanced four wave mixing in a coherently driven dense atomic vapor. It is shown that, in the ideal limit, an arbitrary small flux of pump photons is sufficient to reach the oscillator threshold. We demonstrate that due to the large group velocity delays associated with coherent media, an extremely narrow oscillator linewidth is possible, making a narrow-band source of non-classical radiation feasible.
Abstract: We describe a technique for manipulating quantum information stored in collective states of mesoscopic ensembles. Quantum processing is accomplished by optical excitation into states with strong dipole-dipole interactions. The resulting "dipole blockade" can be used to inhibit transitions into all but singly excited collective states. This can be employed for a controlled generation of collective atomic spin states as well as non-classical photonic states and for scalable quantum logic gates. An example involving a cold Rydberg gas is analyzed.
Abstract: The recently proposed idea to generate entanglement between photon states via exchange interactions in an ensemble of atoms (J. D. Franson and T. B. Pitman, Phys. Rev. A 60 , 917 (1999) and J. D. Franson et al., (quant- ph/9912121)) is discussed using an S -matix approach. It is shown that if the nonlinear response of the atoms is negligible and no additional atom-atom interactions are present, exchange interactions cannot produce entanglement between photons states in a process that returns the atoms to their initial state. Entanglement generation requires the presence of a nonlinear atomic response or atom-atom interactions.
Introduction: Recent developments in quantum communication and computing [1-3] stimulated an intensive search for physical systems that can be used for coherent processing of quantum information. It is generally believed that quantum entanglement of distinguishable quantum bits (qubits) is at the heart of quantum information processing. Significant efforts have been directed towards the design of elementary logic gates, which perform certain unitary processes on pairs of qubits. These gates must be capable of generating specific, in general entangled, superpositions of the two qubits and thus require a strong qubit-qubit interaction. Using a sequence of single and two-bit operations, an arbitrary quantum computation can be performed [2]. Over the past few years many systems have been identified for potential implementations of logic gates and several interesting experiments have been performed. Proposals for strong qubit-qubit interaction involve e.g. the vibrational coupling of cooled trapped ions [4], near dipole-dipole or spin-spin interactions such as in nuclear magnetic resonance [5], collisional interactions of confined cooled atoms [6] or radiative interactions between atoms in cavity QED [7]. The possibility of simple preparation and measurement of qubit states as well as their relative insensitivity to a thermal environment makes the latter schemes particularly interesting for quantum information processing. Most theoretical proposals on cavity-QED systems focus on fundamental systems involving a small number of atoms and few photons. These systems are sufficiently simple to allow for a first-principle description. Their experimental implementation is however quite challenging. For example, extremely high-Q micro-cavities are needed to preserve coherence during all atom-photon interactions. Furthermore, single atoms have to be confined inside the cavities for a sufficiently long time. This requires developments of novel cooling and trapping techniques, which is in itself a fascinating direction of current research. Despite these technical obstacles, a remarkable progress has been made in this area: quantum processors consisting of several coupled qubits now appear to be feasible.
Dynamics of Excited Electrons in Copper and Ferromagnetic Transition Metals: Theory and Experiment
(2000)
Both theoretical and experimental results for the dynamics of photoexcited electrons at surfaces of Cu and the ferromagnetic transition metals Fe, Co, and Ni are presented. A model for the dynamics of excited electrons is developed, which is based on the Boltzmann equation and includes effects of photoexcitation, electron-electron scattering, secondary electrons (cascade and Auger electrons), and transport of excited carriers out of the detection region. From this we determine the time-resolved two-photon photoemission (TR-2PPE). Thus a direct comparison of calculated relaxation times with experimental results by means of TR-2PPE becomes possible. The comparison indicates that the magnitudes of the spin-averaged relaxation time t and of the ratio t_up/t_down of majority and minority relaxation times for the different ferromagnetic transition metals result not only from density-of-states effects, but also from different Coulomb matrix elements M. Taking M_Fe > M_Cu > M_Ni = M_Co we get reasonable agreement with experiments.
Abstract: We analyse 4-dimensional massive "phi" ^ 4 theory at finite temperature T in the imaginary-time formalism. We present a rigorous proof that this quantum field theory is renormalizable, to all orders of the loop expansion. Our main point is to show that the counterterms can be chosen temperature independent, so that the temperature flow of the relevant parameters as a function of T can be followed. Our result confirms the experience from explicit calculations to the leading orders. The proof is based on flow equations, i.e. on the (perturbative) Wilson renormalization group. In fact we will show that the difference between the theories at T > 0 and at T = 0 contains no relevant terms. Contrary to BPHZ type formalisms our approach permits to lay hand on renormalization conditions and counterterms at the same time, since both appear as boundary terms of the renormalization group flow. This is crucial for the proof.
Abstract: Let H_1 , H_2 be complex Hilbert spaces, H be their Hilbert tensor product and let tr_2 be the operator of taking the partial trace of trace class operators in H with respect to the space H_2 . The operation tr_2 maps states in H (i.e. positive trace class operators in H with trace equal to one) into states in H_1 . In this paper we give the full description of mappings that are linear right inverse to tr_2 . More precisely, we prove that any affine mapping F(W) of the convex set of states in H_1 into the states in H that is right inverse to tr_2 is given by W -> W x D for some state D in H_2 . In addition we investigate a representation of the quantum mechanical state space by probability measures on the set of pure states and a representation - used in the theory of stochastic Schrödinger equations - by probability measures on the Hilbert space. We prove that there are no affine mappings from the state space of quantum mechanics into these spaces of probability measures.
Abstract: We develop a method of singularity analysis for conformal graphs which, in particular, is applicable to the holographic image of AdS supergravity theory. It can be used to determine the critical exponents for any such graph in a given channel. These exponents determine the towers of conformal blocks that are exchanged in this channel. We analyze the scalar AdS box graph and show that it has the same critical exponents as the corresponding CFT box graph. Thus pairs of external fields couple to the same exchanged conformal blocks in both theories. This is looked upon as a general structural argument supporting the Maldacena hypothesis.
Abstract: The calculation of absorption cross sections for minimal scalars in supergravity backgrounds is an important aspect of the investigation of AdS/CFT correspondence and requires a matching of appropriate wave functions. The low energy case has attracted particular attention. In the following the dependence of the cross section on the matching point is investigated. It is shown that the low energy limit is independent of the matching point and hence exhibits universality. In the high energy limit the independence is not maintained, but the result is believed to possess the correct energy dependence.
Abstract: The transition from the instanton-dominated quantum regime to the sphaleron-dominated classical regime is studied in the d = 2 abelian-Higgs model when the spatial coordinate is compactified to S1. Contrary to the noncompactified case, this model allows both sharp first-order and smooth second-order transitions depending on the size of the circle. This finding may make the model a useful toy model for the analysis of baryon number violating processes.
Abstract: Standard methods of nonlinear dynamics are used to investigate the stability of particles, branes and D-branes of abelian Born-Infeld theory. In particular the equation of small fluctuations about the D-brane is derived and converted into a modified Mathieu equation and - complementing earlier low-energy investigations in the case of the dilaton-axion system - studied in the high-energy domain. Explicit expressions are derived for the S-matrix and absorption and reflection amplitudes of the scalar fluctuation in the presence of the D-brane. The results confirm physical expectations and numerical studies of others. With the derivation and use of the (hitherto practically unknown) high energy expansion of the Floquet exponent our considerations also close a gap in earlier treatments of the Mathieu equation.
Abstact. The tunnel splitting in biaxial antiferromagnetic particles is studied with a magnetic field applied along the hard anisotropy axis. We observe the oscillation of tunnel splitting as a function of the magnetic field due to the quantum phase interference of two tunneling paths of opposite windings. The oscillation is similar to the recent experimental result with Fe8 molecular clusters.
Abstract: The point-particle-like Hamiltonian of a biaxial spin particle with external magnetic field along the hard axis is obtained in terms of the potential field description of spin systems with exact spin-coordinate correspondence. The Zeeman energy term turns out to be an effective gauge potential which leads to a nonintegrable phase of the Euclidean Feynman propagator. The phase interference between clockwise and anticlockwise under barrier propagations is recognized explicitly as the Aharonov-Bohm effect. An additional phase which is significant for quantum phase interference is discovered with the quantum theory of spin systems besides the known phase obtained with the semiclassical treatment of spin. We also show the energy dependence of the effect and obtain the tunneling splitting at excited states with the help of periodic instantons.
Abstract: Following our earlier investigations we examine the quantum-classical winding number transition in the Abelian-Higgs system. It is demonstrated that the winding number transition in this system is of the smooth second order type in the full range of parameter space. Comparison of the action of classical vortices with that of the sphaleron supports our finding.
Abstract: A Born-Infeld theory describing a D2-brane coupled to a 4-form RR field strength is considered, and the general solutions of the static and Euclidean time equations are derived and discussed. The period of the bounce solutions is shown to allow a consideration of tunneling and quantum-classical transitions in the sphaleron region. The order of such transitions, depending on the strength of the RR field strength, is determined. A criterion is then derived to confirm these findings.
Abstract: The functional relation between interquark potential and interquark distance is explicitly derived by considering the Nambu-Goto action in the AdS5 X S 5 background. It is also shown that a similar relation holds in a general background. The implications of this relation for confinement are briefly discussed.
Abstract: Winding number transitions from quantum to classical behavior are studied in the case of the 1+1 dimensional Mottola-Wipf model with the space coordinate on a circle for exploring the possibility of obtaining transitions of second order. The model is also studied as a prototype theory which demonstrates the procedure of such investigations. In the model at hand we find that even on a circle the transitions remain those of first order.
A new method is used to investigate the tunneling between two weakly-linked Bose-Einstein con- densates confined in double-well potential traps. The nonlinear interaction between the atoms in each well contributes to a finite chemical potential, which, with consideration of periodic instantons, leads to a remarkably high tunneling frequency. This result can be used to interpret the newly found Macroscopic Quantum Self Trapping (MQST) effect. Also a new kind of first-order crossover between different regions is predicted.
The increasing parallelisation of development processes as well as the ongoing trends towards virtual product development and outsourcing of development activities strengthen the need for 3D co-operative design via communication networks. Regarding the field of CAx, none of the existing systems meets all the requirements of very complex process chain. This leads to a tremendous need for the integration of heterogeneous CAx systems. Therefore, MACAO, a platform-independent client for a distributed CAx component system, the so-called ANICA CAx object bus, is presented. The MACAO client is able to access objects and functions provided by different CAx servers distributed over a communication network. Thus, MACAO is a new solution for engineering design and visualisation in shared distributed virtual environments. This paper describes the underlying concepts, the actual prototype implementation, as well as possible application scenarios in the area of co-operative design and visualisation.
Abstract: The self-duality of chiral p-forms was originally investigated by Pasti, Sorokin and Tonin in a manifestly Lorentz covariant action with non-polynomial auxiliary fields. The investigation was then extended to other chiral p-form actions. In this paper we point out that the self-duality appears in a wider context of theoretical models that relate to chiral p-forms. We demonstrate this by considering the interacting model of Floreanini- Jackiw chiral bosons and gauge fields, the generalized chiral Schwinger model (GCSM) and the latter's gauge invariant formulation, and discover that the self-duality of the GCSM corresponds to the vector and axial vector current duality.
Abstract: It has recently been shown that the equation of motion of a massless scalar field in the background of some specific p branes can be reduced to a modified Mathieu equation. In the following the absorption rate of the scalar by a D3 brane in ten dimensions is calculated in terms of modified Mathieu functions of the first kind, using standard Mathieu coefficients. The relation of the latter to Dougall coefficients (used by others) is investigated. The S-matrix obtained in terms of modified Mathieu functions of the first kind is easily evaluated if known rapidly convergent low energy expansions of these in terms of products of Bessel functions are used. Leading order terms, including the interesting logarithmic contributions, can be obtained analytically.
Abstract: The duality symmetries of various chiral boson actions are investigated using D = 2 and D = 6 space-time dimensions as examples. These actions involve the Siegel, Floreanini-Jackiw, Srivastava and Pasti-Sorokin-Tonin formulations. We discover that the Siegel, Floreanini-Jackiw and Pasti-Sorokin-Tonin actions have self-duality with respect to a common anti-dualization of chiral boson fields in D = 2 and D = 6 dimensions, respectively, while the Srivastava action is self-dual with respect to a generalized dualization of chiral boson fields. Moreover, the action of the Floreanini-Jackiw chiral bosons interacting with gauge fields in D = 2 dimensions also has self-duality but with respect to a generalized anti-dualization of chiral boson fields.
A pure Yang-Mills theory extended by addition of a quartic term is considered in order to study the transition from the quantum tunneling regime to that of classical, i.e. thermal, behaviour. The periodic field confiurations are found, which interpolate between the vacuum and sphaleron field configurations. It is shown by explicit calculation that only smooth second order transitions occur for all permissible values of the parameter A introduced with the quartic term. The theory is one of the rare cases which canbe handled analytically.
Abstract: The transition from the quantum to the classical regime of the nucleation of the closed Robertson-Walker Universe with spacially homogeneous matter fields is investigated with a perturbation expansion around the sphaleron configuration. A criterion is derived for the occurrence of a first-order type transition, and the related phase diagram for scalar and vector fields is obtained. For scalar fields both the first and second order transitions can occur depending on the shape of the potential barrier. For a vector field, here that of an O (3) nonlinear o-model, the transition is seen to be only of the first order. PACS numbers: 11.15.Kc, 03.65Sq, 05.70.Fh, 98.80.Cq
Corporate environmental reporting makes good business and environmental sense. A big challenge for companies is to utilize the technical benefit of state of the art IT, especially of Internet-technologies and Internet-services. In this paper an approach of internet-based environmental reports by companies is presented. Three different levels are discussed: The first level deals with the basics of corporate environmental reports (CER) by companies. Illustrating the order within the emerging field of CERs a morphological box is suggested (section 1). Building on this, general requirements for corporate environmental reports are outlined (section 2). On the second level, the general reporting requirements are specified by IT-relevant challenges, seen as starting points for internet-based environmental reports (section 3). The immense technical benefit of using the Internet towards efficient, integrated, interactive, hypermedia-featured, dialog-oriented, and customised environmental reporting is analysed (section 4). On the basis of the technical benefit analysis, the state of the art of internet-based CERs is presented (section 5). The third level refers to the IT-application turning from the basics, IT-challenges and technical benefit to consequences for environmental reporting companies in practice. Thereby a fundamental framework for internet-based CERs is sketched (section 6). Grounded on this framework a basic architecture of an IT-implementation is explained (section 7).
Software development organizations are recognizing the increasing importance of investing in the build-up of core competencies for their competitiveness in software system development. This is supported by reuse and experience repository systems that assist in capturing and reusing all kinds of software artifacts (e. g., code, patterns, frameworks) and processes as well as experiences related to these artifacts and processes. To justify such an investment and guide its improvement, it must be evaluated according to the business case, that is, a measurement program has to be developed that is oriented towards the business goals of such a reuse and experience repository system. In this paper, we suggest an approach to iteratively build up measurement programs for gaining feedback and, thereby, controlling and improving such a reuse and experience repository system. The focus is placed on guidelines for the evolution of such measurement programs over time, rather than providing directly applicable metrics or questionnaires. In order to illustrate the feasibility of the approach, examples of running measurement programs at different stages of evolutions are given.
Comprehensive reuse and systematic evolution of reuse artifacts as proposed by the Quality Improvement Paradigm (QIP) do not only require tool support for mere storage and retrieval. Rather, an integrated management of (potentially reusable) experience data as well as project-related data is needed. This paper presents an approach exploiting object-relational database technology to implement QIP-driven reuse repositories. Requirements, concepts, and implementational aspects are discussed and illustrated through a running example, namely the reuse and continuous improvement of SDL patterns for developing distributed systems. Our system is designed to support all phases of a reuse process and the accompanying improvement cycle by providing adequate functionality. Its implementation is based on object-relational database technology along with an infrastructure well suited for these purposes.
Chaotic Billiards
(2000)
The frictionless motion of a particle on a plane billiard table The frictionless motion of a particle on a plane billiard table bounded by a closed curve provides a very simple example of a conservative classical system with non-trivial, chaotic dynamics. The limiting cases of strictly regular ("integrable") and strictly irregular ("ergodic") systems can be illustrated, as well as the typical case which shows an intricate mixture of regular and irregular behavior. Irregular orbits are characterized by an extremely sensitivity with respect to the initial conditions. Such billiard systems are exemplarily suited for educational purposes as models for simple systems with complicated dynamics as well as for far-reaching fundamental investigations.
A simple method of calculating the Wannier-Stark resonances in 2D lattices is suggested. Using this method we calculate the complex Wannier-Stark spectrum for a non-separable 2D potential realized in optical lattices and analyze its general structure. The dependence of the lifetime of Wannier-Stark states on the direction of the static field (relative to the crystallographic axis of the lattice) is briefly discussed.
An extremely simple and convenient method is presented for computing eigenvalues in quantum mechanics by representing position and momentum operators in a simple matrix form. The simplicity and success of the method is illustrated by numerical results concerning eigenvalues of bound systems and resonances for hermitian and non-hermitian Hamiltonians as well as driven quantum systems.
The analyticity property of the one-dimensional complex Hamiltonian system H(x,p)=H_1(x_1,x_2,p_1,p_2)+iH_2(x_1,x_2,p_1,p_2) with p=p_1+ix_2, x=x_1+ip_2 is exploited to obtain a new class of the corresponding two-dimensional integrable Hamiltonian systems where H_1 acts as a new Hamiltonian and H_2 is a second integral of motion. Also a possible connection between H_1 and H_2 is sought in terms of an auto-B"acklund transformation.
The paper studies the dynamics of transitions between the levels of a Wannier-Stark ladder induced by a resonant periodic driving. The analysis of the problem is done in terms of resonance quasienergy states, which take into account the metastable character of the Wannier-Stark states. It is shown that the periodic driving creates from a localized Wannier-Stark state an extended Bloch-like state with a spatial length varying in time as ~ t^1/2. Such a state can find applications in the field of atomic optics because it generates a coherent pulsed atomic beam.
We consider investment problems where an investor can invest in a savings account, stocks and bonds and tries to maximize her utility from terminal wealth. In contrast to the classical Merton problem we assume a stochastic interest rate. To solve the corresponding control problems it is necessary to prove averi cation theorem without the usual Lipschitz assumptions.
In the Black-Scholes type financial market, the risky asset S 1 ( ) is supposed to satisfy dS 1 ( t ) = S 1 ( t )( b ( t ) dt + Sigma ( t ) dW ( t ) where W ( ) is a Brownian motion. The processes b ( ), Sigma ( ) are progressively measurable with respect to the filtration generated by W ( ). They are known as the mean rate of return and the volatility respectively. A portfolio is described by a progressively measurable processes Pi1 ( ), where Pi1 ( t ) gives the amount invested in the risky asset at the time t. Typically, the optimal portfolio Pi1 ( ) (that, which maximizes the expected utility), depends at the time t, among other quantities, on b ( t ) meaning that the mean rate of return shall be known in order to follow the optimal trading strategy. However, in a real-world market, no direct observation of this quantity is possible since the available information comes from the behavior of the stock prices which gives a noisy observation of b ( ). In the present work, we consider the optimal portfolio selection which uses only the observation of stock prices.
In multicriteria optimization problems the connectedness of the set of efficient solutions (pareto set) is of special interest since it would allow the determination of the efficient solutions without considering non-efficient solutions in the process. In the case of the multicriteria problem to minimize matchings the set of efficient solutions is not connected. The set of minimal solutions E pot with respect to the power ordered set contains the pareto set. In this work theorems about connectedness of E pot are given. These lead to an automated process to detect all efficient solutions.
Many polynomially solvable combinatorial optimization problems (COP) become NP when we require solutions to satisfy an additional cardinality constraint. This family of problems has been considered only recently. We study a newproblem of this family: the k-cardinality minimum cut problem. Given an undirected edge-weighted graph the k-cardinality minimum cut problem is to find a partition of the vertex set V in two sets V 1 , V 2 such that the number of the edges between V 1 and V 2 is exactly k and the sum of the weights of these edges is minimal. A variant of this problem is the k-cardinality minimum s-t cut problem where s and t are fixed vertices and we have the additional request that s belongs to V 1 and t belongs to V 2 . We also consider other variants where the number of edges of the cut is constrained to be either less or greater than k. For all these problems we show complexity results in the most significant graph classes.
Performance of some preconditioners for the p - and hp -version of the finite element method in 3D
(2000)
We consider the problem of locating a line or a line segment in three- dimensional space, such that the sum of distances from the linear facility to a given set of points is minimized. An example is planning the drilling of a mine shaft, with access to ore deposits through horizontal tunnels connecting the deposits and the shaft. Various models of the problem are developed and analyzed, and effcient solution methods are given.
We examine the feasibility polyhedron of the uncapacitated hub location problem (UHL) with multiple allocation, which has applications in the fields of air passenger and cargo transportation, telecommunication and postal delivery services. In particular we determine the dimension and derive some classes of facets of this polyhedron. We develop some general rules about lifting facets from the uncapacitated facility location (UFL) for UHL and projecting facets from UHL to UFL. By applying these rules we get a new class of facets for UHL which dominates the inequalities in the original formulation. Thus we get a new formulation of UHL whose constraints are all facet defining. We show its superior computational performance by benchmarking it on a well known data set.
In this paper we deal with single facility location problems in a general normed space where the existing facilities are represented by sets. The criterion to be satis ed by the service facility is the minimization of an increasing function of the distances from the service to the closest point ofeach demand set. We obtain a geometrical characterization of the set of optimal solutions for this problem. Two remarkable cases - the classical Weber problem and the minmax problem with demand sets - are studied as particular instances of our problem. Finally, for the planar polyhedral case we give an algorithmic description of the solution set of the considered problems.
The balance space approach (introduced by Galperin in 1990) provides a new view on multicriteria optimization. Looking at deviations from global optimality of the different objectives, balance points and balance numbers are defined when either different or equal deviations for each objective are allowed. Apportioned balance numbers allow the specification of proportions among the deviations. Through this concept the decision maker can be involved in the decision process. In this paper we prove that the apportioned balance number can be formulated by a min-max operator. Furthermore we prove some relations between apportioned balance numbers and the balance set, and see the representation of balance numbers in the balance set. The main results are necessary and sufficient conditions for the balance set to be exhaustive, which means that by multiplying a vector of weights (proportions of deviation) with its corresponding apportioned balance number a balance point is attained. The results are used to formulate an interactive procedure for multicriteria optimization. All results are illustrated by examples.
In this paper we investigate the problem offending the Nadir point for multicriteria optimization problems (MOP). The Nadir point is characterized by the component wise maximal values of efficient points for (MOP). It can be easily computed in the bicriteria case. However, in general this problem is very difficult. We review some existing methods and heuristics and propose some new ones. We propose a general method to compute Nadir values for the case of three objectives, based on theoretical results valid for any number of criteria. We also investigate the use of the Nadir point for compromise programming, when the goal is to be as far away as possible from the worst outcomes. We prove some results about (weak) Pareto optimality of the resulting solutions. The results are illustrated by examples.
In this paper we address the question of how many objective functions are needed to decide whether a given point is a Pareto optimal solution for a multicriteria optimization problem. We extend earlier results showing that the set of weakly Pareto optimal points is the union of Pareto optimal sets of subproblems and show their limitations. We prove that for strictly quasi-convex problems in two variables Pareto optimality can be decided by consideration of at most three objectives at a time. Our results are based on a geometric characterization of Pareto, strict Pareto and weak Pareto solutions and Helly's Theorem. We also show that a generalization to quasi-convex objectives is not possible, and state a weaker result for this case. Furthermore, we show that a generalization to strictly Pareto optimal solutions is impossible, even in the convex case.
This paper provides an annotated bibliography of multiple objective combinatorial optimization, MOCO. We present a general formulation of MOCO problems, describe the main characteristics of MOCO problems, and review the main properties and theoretical results for these problems. One section is devoted to a brief description of the available solution methodology, both exact and heuristic. The main part of the paper is devoted to an annotation of the existing literature in the field organized problem by problem. We conclude the paper by stating open questions and areas of future research. The list of references comprises more than 350 entries.
At present the standardization of third generation (3G) mobile radio systems is the subject of worldwide research activities. These systems will cope with the market demand for high data rate services and the system requirement for exibility concerning the offered services and the transmission qualities. However, there will be de ciencies with respect to high capacity, if 3G mobile radio systems exclusively use single antennas. Very promising technique developed for increasing the capacity of 3G mobile radio systems the application is adaptive antennas. In this thesis, the benefits of using adaptive antennas are investigated for 3G mobile radio systems based on Time Division CDMA (TD-CDMA), which forms part of the European 3G mobile radio air interface standard adopted by the ETSI, and is intensively studied within the standardization activities towards a worldwide 3G air interface standard directed by the 3GPP (3rd Generation Partnership Project). One of the most important issues related to adaptive antennas is the analysis of the benefits of using adaptive antennas compared to single antennas. In this thesis, these bene ts are explained theoretically and illustrated by computer simulation results for both data detection, which is performed according to the joint detection principle, and channel estimation, which is applied according to the Steiner estimator, in the TD-CDMA uplink. The theoretical explanations are based on well-known solved mathematical problems. The simulation results illustrating the benefits of adaptive antennas are produced by employing a novel simulation concept, which offers a considerable reduction of the simulation time and complexity, as well as increased exibility concerning the use of different system parameters, compared to the existing simulation concepts for TD-CDMA. Furthermore, three novel techniques are presented which can be used in systems with adaptive antennas for additionally improving the system performance compared to single antennas. These techniques concern the problems of code-channel mismatch, of user separation in the spatial domain, and of intercell interference, which, as it is shown in the thesis, play a critical role on the performance of TD-CDMA with adaptive antennas. Finally, a novel approach for illustrating the performance differences between the uplink and downlink of TD-CDMA based mobile radio systems in a straightforward manner is presented. Since a cellular mobile radio system with adaptive antennas is considered, the ultimate goal is the investigation of the overall system efficiency rather than the efficiency of a single link. In this thesis, the efficiency of TD-CDMA is evaluated through its spectrum efficiency and capacity, which are two closely related performance measures for cellular mobile radio systems. Compared to the use of single antennas, the use of adaptive antennas allows impressive improvements of both spectrum efficiency and capacity. Depending on the mobile radio channel model and the user velocity, improvement factors range from six to 10.7 for the spectrum efficiency, and from 6.7 to 12.6 for the spectrum capacity of TD-CDMA. Thus, adaptive antennas constitute a promising technique for capacity increase of future mobile communications systems.
We consider the determination of optimal portfolios under the threat of a crash. Our main assumption is that upper bounds for both the crash size and the number of crashes occurring before the time horizon are given. We make no probabilistic assumption on the crash size or the crash time distribution. The optimal strategies in the presence of a crash possibility are characterized by a balance problem between insurance against the crash and good performance in the crash-free situation. Explicit solutions for the log-utility case are given. Our main finding is that constant portfolios are no longer optimal ones.
We consider some continuous-time Markowitz type portfolio problems that consist of maximizing expected terminal wealth under the constraint of an upper bound for the Capital-at-Risk. In a Black-Scholes setting we obtain closed form explicit solutions and compare their form and implications to those of the classical continuous-time mean-variance problem. We also consider more general price processes which allow for larger uctuations in the returns.
Linearized flows past slender bodies can be asymptotically described by a linear Fredholm integral equation. A collocation method to solve this equation is presented. In cases where the spectral representation of the integral operator is explicitly known, the collocation method recovers the spectrum of the continuous operator. The approximation error is estimated for two discretizations of the integral operator and the convergence is proved. The collocation scheme is validated in several test cases and extended to situations where the spectrum is not explicit.
The satellite-to-satellite tracking (SST) problems are characterized from mathematical point of view. Uniqueness results are formulated. Moreover, the basic relations are developed between (scalar) approximation of the earth's gravitational potential by "scalar basis systems" and (vectorial) approximation of the gravitational eld by "vectorial basis systems". Finally, the mathematical justication is given for approximating the external geopotential field by finite linear combinations of certain gradient fields (for example, gradient fields of multi-poles) consistent to a given set of SST data.
Being interested in (rotation-)invariant pseudodi erential equations of satellite problems corresponding to spherical orbits, we are reasonably led to generating kernels that depend only on the spherical distance, i. e. in the language of modern constructive approximation form spherical radial basis functions. In this paper approximate identities generated by such (rotation-invariant) kernels which are additionally locally supported are investigated in detail from theoretical as well as numerical point of view. So-called spherical di erence wavelets are introduced. The wavelet transforms are evaluated by the use of a numerical integration rule, that is based on Weyl's law of equidistribution. This approximate formula is constructed such that it can cope with millions of (satellite) data. The approximation error is estimated on the orbital sphere. Finally, we apply the developed theory to the problems of satellite-to-satellite tracking (SST) and satellite gravity gradiometry (SGG).
The aim of this article is to show that moment approximations of kinetic equations based on a Maximum Entropy approach can suffer from severe drawbacks if the kinetic velocity space is unbounded. As example, we study the Fokker Planck equation where explicit expressions for the moments of solutions to Riemann problems can be derived. The quality of the closure relation obtained from the Maximum Entropy approach as well as the Hermite/Grad approach is studied in the case of five moments. It turns out that the Maximum Entropy closure is even singular in equilibrium states while the Hermite/Grad closure behaves reasonably. In particular, the admissible moments may lead to arbitrary large speeds of propagation, even for initial data arbitrary close to global eqilibrium.
The paper concerns the equilibrium state of ultra small semiconductor devices. Due to the quantum drift diffusion model, electrons and holes behave as a mixture of charged quantum fluids. Typically the involved scaled Plancks constants of holes, \(\xi\), is significantly smaller than the scaled Plancks constant of electrons. By setting formally \(\xi=0\) a well-posed differential-algebraic system arises. Existence and uniqueness of an equilibrium solution is proved. A rigorous asymptotic analysis shows that this equilibrium solution is the limit (in a rather strong sense) of quantum systems as \(\xi \to 0\). In particular the ground state energies of the quantum systems converge to the ground state energy of the differential-algebraic system as \(\xi \to 0\).
An asymptotic preserving numerical scheme (with respect to diffusion scalings) for a linear transport equation is investigated. The scheme is adopted from a class of recently developped schemes. Stability is proven uniformly in the mean free path under a CFL type condition turning into a parabolic CFL condition in the diffusion limit.
Starting with general hyperbolic systems of conservation laws, a special sub - class is extracted in which classical solutions can be expressed in terms of a linear transport equation. A characterizing property of this sub - class which contains, for example, all linear systems and non - linear scalar equations, is the existence of so called exponentially exact entropies.
Based on general partitions of unity and standard numerical flux functions, a class of mesh-free methods for conservation laws is derived. A Lax-Wendroff type consistency analysis is carried out for the general case of moving partition functions. The analysis leads to a set of conditions which are checked for the finite volume particle method FVPM. As a by-product, classical finite volume schemes are recovered in the approach for special choices of the partition of unity.
In this chapter, the quantitative numerical simulation of the behavior of deformable linear objects, such as hoses, wires and leaf springs is studied. We first give a short review of the physical approach and the basic solution principle. Then, we give a more detailed description of some key aspects: We introduce a novel approach concerning dynamics based on an algorithm very similar to the one used for (quasi-) static computation. Then, we look at the plastic workpiece deformation, involving a modified computation algorithm and a special representation of the workpiece shape. Then, we give alternative solutions for two key aspects of the algorithm, and investigate the problem of performing the workpiece simulation efficiently, i.e., with desired precision in a short time. In the end, we introduce the inverse modeling problem which must be solved when the gripper trajectory for a given task shall be generated.
A new and systematic basic approach to force- and vision-based robot manipulation of deformable (non-rigid) linear objects is introduced. This approach reduces the computational needs by using a simple state-oriented model of the objects. These states describe the relation between the deformable and rigid obstacles, and are derived from the object image and its features. We give an enumeration of possible contact states and discuss the main characteristics of each state. We investigate the performance of robust transitions between the contact states and derive criteria and conditions for each of the states and for two sensor systems, i.e. a vision sensor and a force/torque sensor. This results in a new and task-independent approach in regarding the handling of deformable objects and in a sensor-based implementation of manipulation primitives for industrial robots. Thus, the usage of sensor processing is an appropriate solution for our problem. Finally, we apply the concept of contact states and state transitions to the description of a typical assembly task. Experimental results show the feasibility of our approach: A robot performs several contact state transitions which can be combined for solving a more complex task.
This essay discusses the multileaf collimator leaf sequencing problem, which occurs in every treatment planning in radiation therapy. The problem is to find a good realization in terms of a leaf sequence in the multileaf collimator such that the time needed to deliver the given dose profile is minimized. A mathematical model using an integer programming formulation has been developed. Additionally, a heuristic, based on existing algorithms and an integer programming formulation, has been developed to enhance the quality of the solutions. Comparing the results to those provided by other algorithms, a significant improvement can be observed.
Mean field equations arise as steady state versions of convection-diffusion systems where the convective field is determined as solution of a Poisson equation whose right hand side is affine in the solutions of the convection-diffusion equations. In this paper we consider the repulsive coupling case for a system of 2 convection-diffusion equations. For general diffusivities we prove the existence of a unique solution of the mean field equation by a variational technique. Also we analyse the small-Debye-length limit and prove convergence to either the so-called charge-neutral case or to a double obstacle problem for the limiting potential depending on the data.
Besides the work in the field of manipulating rigid objects, currently, there are several research and development activities going on in the field of manipulating non-rigid or deformable objects. Several papers have been published on international conferences in this field from various projects and countries. But there has been no comprehensive work which provides both a representative overview of the state of the art and identifies the important aspects in this field. Thus, we collected these activities and invited the corresponding working groups to present an overview of their research. Altogether, nineteen authors coming from Japan, Germany, Italy, Greece, United Kingdom, and Australia contributed to this book. Their research work covers all the different aspects that occur when manipulating deformable objects. The contributions can be characterized and grouped by the following four aspects: * object modeling and simulation, * planning and control strategies, * collaborative systems, and * applications and industrial experiences. In the following, we give a short motivation and overview of the single chapters of the book. The simulation of deformable objects is one way to approach the problem of manipulating these objects by robots. Based on a physical model of the object and the occurring constraints, the resulting object shape is calculated. In Chapter 2, Hirai presents an energy-based approach, where the internal energy under the geometric constraints is minimized. Frugoli et al. introduce a force-based approach, where the forces between discrete particles are minimized meeting given constraints. Finally, Remde and Henrich extend the energy-based approach to plastic deformation and give a solution of the inverse simulation problem. Even if the object behavior is predicted by simulation, there is still the question of how to control the robot during a single manipulation operation. An additional question is how to retrieve an overall plan for the concatenated manipulation operations. In Chapter 3, Wada investigates the control problems when positioning multiple points of a planar deformable object. McCarrager proposes a control scheme exploiting the flexibility, rather than minimizing it. Abegg et al. use a simple contact state model to describe typical assembly tasks and to derive robust manipulation primitives. Finally, Ono presents an automatic sewing system and suggests a strategy for unfolding fabric. In several manipulation tasks, it is reasonable to apply more than one robot. Especially in cases, where the deformable object has to take a specific shape. Since the robots working at the same object are influencing each other, different control algorithms have to be introduced. In Chapter 4, Yoshida and Kosuge investigates this problem for the task of bending a sheet of metal and exploits the relation ship between the static object deformation and the bending moments. Tanner and Kyriakopoulos regard the deformable object as underactuated mechanical system and make use of the existence of non-holonomic constraints. Both approaches model the deformable object as finite elements. All of the above aspects have their counterpart in different applications and industrial experiences. In Chapter 5, Rizzi et al. present test cases and applications of their approach to simulate the manipulation of fabric, wires, cables, and soft bags. Buckingham and Graham give an overview of two European projects processing white fish including locating, gripping, and deheading the fish. Maruyama outlines the three development phases of a robot system for performing outage-free maintenance of live-line power supply in Japan. Finally, Kämper presents the development of a flexible automatic cabling unit for the wiring of long-tube lighting with plug components.
We present a detailed analysis of a scalar conformal four-point function obtained from AdS/CFT correspondence. We study the scalar exchange graphs in AdS and discuss their analytic properties. Using methods of conformal partial wave analysis, we present a general procedure to study conformal four-point functions in terms of exchanges of scalar and tensor fields. The logarithmic terms in the four-point functions are connected to the anomalous dimensions of the exchanged fields. Comparison of the results from AdS graphs with the conformal partial wave analysis, suggests a possible general form for the operator product expansion of scalar fields in the boundary CFT.
We discuss the analytic properties of AdS scalar exchange graphs in the crossed channel. We show that the possible non-analytic terms drop out by virtue of non-trivial properties of generalized hypergeometric functions. The absence of non-analytic terms is a necessary condition for the existence of an operator product expansion for CFT amplitudes obtained from AdS/CFT correspondence.
The basic idea behind selective multiscale reconstruction of functions from error-affected data is outlined on the sphere. The selective reconstruction mechanism is based on the premise that multiscale approximation can be well-represented in terms of only a relatively small number of expansion coefficients at various resolution levels. An attempt is made within a tree algorithm (pyramid scheme) to remove the noise component from each scale coefficient using a priori statistical information (provided by an error covariance kernel of a Gaussian, stationary stochastic model).
For transferring existing knowledge into new projects, reuse has become an important factor in today's software industry. However, to set reuse into practice, reusable artifacts have to be stored somewhere, and must be offered to (re-)users on demand. For this purpose, advanced reuse repository systems like, for instance, instantiations of the Experience Base concept, are quite frequently used. Many people, from different projects, have to access such a repository at various phases of software development processes to retrieve or store reusable data. In order to fulfill the given tasks, each of these user has specific needs. Taking this into account, a reuse repository has to offer tailored user interfaces and functions for different user groups. Furthermore, since the contents of such a repository usually represent the state of the art of an organization's (core) competencies, not everyone should be allowed to freely access each and every repository entry. This isespecially true for persons that are not part of the organization. This report discusses role concepts that can be applied to reuse repository systems to overcome some of the stated access problems. Commonly used roles for software development and reuse repository management are listed. Based on these roles, a basic set of roles, as implemented in the SFB 501 Experience Base, is introduced.
In this paper we construct a multiscale solution method for the gravimetry problem, which is concerned with the determination of the earth's density distribution from gravitational measurements. For this purpose isotropic scale continuous wavelets for harmonic functions on a ball and on a bounded outer space of a ball, respectively, are constructed. The scales are discretized and the results of numerical calculations based on regularization wavelets are presented. The obtained solutions yield topographical structures of the earth's surface at different levels of localization ranging from continental boundaries to local structures such as Ayer's Rock and the Amazonas area.
Spherical Tikhonov Regularization Wavelets in Satellite Gravity Gradiometry with Random Noise
(2000)
This paper considers a special class of regularization methods for satellite gravity gradiometry based on Tikhonov spherical regularization wavelets with particular emphasis on the case of data blurred by random noise. A convergence rate is proved for the regularized solution, and a method is discussed for choosing the regularization level a posteriori from the gradiometer data.
Linear viscoelastic properties for a dilute polymer solution are predicted by modeling the solution as a suspension of non-interacting bead-spring chains. The present model, unline the Rouse model, can describe the solution's rheological behavior even when the solvent quality is good, since excluded volume effects are explicitly taken into account through a narrow Gaussian repulsive potential between pairs of beads in a bead-spring chain. The use of the narrow Gaussian potential, which tends to the more commonly used delta-function repulsive potential in the limit of a width parameter d going to zero, enables the performance of Brownian dynamics simulations. The simulations results, which describe the exact behavior of the model, indicate that for chains of arbitrary but finite length, a delta-function potential leads to equilibrium and zero shear rate properties which are identical to the predictions of the Rouse model. On the other hand, a non-zero value of d gives rise to a predictionof swelling at equilibrium, and an increase in zero shear rate properties relative to their Rouse model values. The use of a delta-function potential appears to be justified in the limit of infinite chain length. The exact simulation results are compared with those obtained with an approximate solution, which is based on the assumption that the non-equilibrium configurational distribution function is Gaussian. The Gaussian approximation is shown to be exact to first order in the strength of excluded volume interaction, and is used to explore long chain rheological properties by extrapolating results obtained numerically for finite chains, to the limit of infinite chain length.
The paper studies the effect of a weak periodic driving on metastable Wannier-Stark states. The decay rate of the ground Wannier-Stark states as a continuous function of the driving frequency is calculated numerically. The theoretical results are compared with experimental data of Wilkinson et at. [Phys.Rev.Lett.76, 4512 (1996)] obtained for cold sodium atoms in an accelerated optical lattice.
The statistics of the resonance widths and the behavior of the survival probability is studied in a particular model of quantum chaotic scattering (a particle in a periodic potential subject to static and time-periodic forces) introduced earlier in Ref. [5,6]. The coarse-grained distribution of the resonance widths is shown to be in good agreement with the prediction of Random Matrix Theory (RMT). The behavior of the survival probability shows, however, some deviation from RMT.
We study the transitions between the ground and excited Wannier states induced by a weak ac field. Because the upper Wannier states are several order of magnitude less stable than the ground states, these transitions decrease the global stability of the system characterized by the rate of probability leakage or decay rate. Using nonhermitian resonant perturbation theory we obtain an analytical expression for this induced decay rate. The analytical results are compared with exact numerical calculations of the system decay rate.
The quasienergy spectrum of a Bloch electron affected by dc-ac fields is known to have a fractal structure as function of the so-called electric matching ratio, which is the ratio of the ac field frequency and the Bloch frequency. This paper studies a manifestation of the fractal nature of the spectrum in the system "atom in a standing laser wave", which is a quantum optical realization of a Bloch electron. It is shown that for an appropriate choice of the system parameters the atomic survival probability (a quantity measured in laboratory experiments) also develops a fractal structure as a function of the electric matching ratio. Numerical simulations under classically chaotic scattering conditions show good agreement with theoretical predictions based on random matrix theory.
FeNi/FeMn exchange bias samples with a large exchange bias field at room temperature have been prepared on a Cu buffer layer. Upon irradiation with He ions, both the exchange bias field and the coercive field are modified. For low ion doses the exchange bias field is enhanced by nearly a factor of 2. Above a threshold dose of 0.3olsi 10 15 ions/cm 2 , the exchange bias field decreases continuously as the ion dose increases. The ob-served modifications are explained in terms of defect creation acting as pinning sites for domain walls and atomic intermixing.
For the next generation of high data rate magnetic recording above 1 Gbit/s, a better understanding of the switching processes for both recording heads and media will be required. In order to maximize the switch-ing speed for such devices, the magnetization precession after the magnetic field pulse termination needs to be suppressed to a maximum degree. It is demonstrated experimentally for ferrite films that the appropriate adjustment of the field pulse parameters and/or the static applied field may lead to a full suppression of the magnetization precession immediately upon termination of the field pulse. The suppression is explained by taking into account the actual direction of the magnetization with respect to the static field direction at the pulse termination.
It is well-known that some of the classical location problems with polyhedral gauges can be solved in polynomial time by finding a finite dominating set, i.e. a finite set of candidates guaranteed to contain at least one optimal location. In this paper it is first established that this result holds for a much larger class of problems than currently considered in the literature. The model for which this result can be proven includes, for instance, location problems with attraction and repulsion, and location-allocation problems. Next, it is shown that the approximation of general gauges by polyhedral ones in the objective function of our general model can be analyzed with regard to the subsequent error in the optimal objective value. For the approximation problem two different approaches are described, the sandwich procedure and the greedy algorithm. Both of these approaches lead - for fixed epsilon - to polynomial approximation algorithms with accuracy epsilon for solving the general model considered in this paper.
Phase velocities of surface acoustic waves in several boron nitride films were investigated by Brillouin light scattering. In the case of films with predominantly hexagonal crystal structure, grown under conditions close to the nucleation threshold of cubic BN, four independent elastic constants have been determined from the dispersion of the Rayleigh and the first Sezawa mode. The large elastic anisotropy of up to c11/c33 = 0.1 is attributed to a pronounced texture with the c-axes of the crystallites parallel to the film plane. In the case of cubic BN films the dispersion of the Rayleigh wave provides evidence for the existence of a more compliant layer at the substrate-film interface. The observed broadening of the Rayleigh mode is identified to be caused by the film morphology.
Annual Report 1999
(2000)