Refine
Year of publication
- 1997 (97) (remove)
Document Type
- Preprint (66)
- Article (17)
- Report (9)
- Doctoral Thesis (2)
- Diploma Thesis (1)
- Master's Thesis (1)
- Periodical (1)
Keywords
- AG-RESY (7)
- PARO (7)
- SKALP (2)
- Anisotropic smoothness classes (1)
- Bayesrisiko (1)
- Bewegungsplanung (1)
- Brownian motion (1)
- C (1)
- CAx-Anwendungen (1)
- CODET (1)
- CoMo-Kit (1)
- Dense gas (1)
- Diffusionsprozess (1)
- Elliptic-parabolic equation (1)
- Enskog equation (1)
- Function of bounded variation (1)
- Integral transform (1)
- Intelligent Object Fusion (1)
- Internet knowledge base (1)
- Internet knowledge reuse (1)
- Jacobian (1)
- Java (1)
- Kohonen's SOM (1)
- Laplace transform (1)
- Locally stationary processes (1)
- Moment sequence (1)
- Netz-Architekturen (1)
- Netzwerkmanagement (1)
- Neural networks (1)
- PVM (1)
- Panel clustering (1)
- Parallel Virtual Machines (1)
- Robotik (1)
- Scalar-type operator (1)
- Software Agents (1)
- Stieltjes transform (1)
- Suchve (1)
- Tcl (1)
- Workstation-Cluster (1)
- adaptive estimation (1)
- asymptotic analysis (1)
- authentication (1)
- automated theorem proving (1)
- autonomous systems (1)
- average density (1)
- business process reengineering (1)
- byte code (1)
- compact operator equation (1)
- density distribution (1)
- drift-diffusion limit (1)
- dynamical systems (1)
- entropy (1)
- finite pointset method (1)
- finite-difference methods (1)
- higher-order calculi (1)
- interpreter (1)
- kinetic semiconductor equations (1)
- kinetic theory (1)
- lacunarity distribution (1)
- local stationarity (1)
- localization (1)
- logarithmic averages (1)
- migration (1)
- minimax estimation (1)
- motion planning (1)
- multi-language (1)
- mutiresolution (1)
- non-linear wavelet thresholding (1)
- non-stationary time series (1)
- numerical integration (1)
- numerical methods for stiff equations (1)
- object-oriented software modeling (1)
- occupation measure (1)
- one-dimensional self-organization (1)
- optimal rate of convergence (1)
- order-three density (1)
- parallel algorithms (1)
- parallel numerical algorithms (1)
- parallel processing (1)
- parallelism and concurrency (1)
- particle method (1)
- persistence (1)
- phase-space (1)
- porous media (1)
- quantum chaos (1)
- quantum mechanics (1)
- quantum tunneling (1)
- regularization wavelets (1)
- review (1)
- robot control (1)
- robot kinematics (1)
- robotics (1)
- security domain (1)
- semiclassical quantisation (1)
- shock wave (1)
- software reuse (1)
- spline and wavelet based determination of the geoid and the gravitational potential (1)
- stationarity (1)
- tensor product basis (1)
- test (1)
- threshold choice (1)
- time series (1)
- time-frequency plan (1)
- time-varying covariance (1)
- wavelet thresholding (1)
- wavelets (1)
- winner definition (1)
Faculty / Organisational entity
- Kaiserslautern - Fachbereich Mathematik (36)
- Kaiserslautern - Fachbereich Informatik (35)
- Kaiserslautern - Fachbereich Physik (18)
- Kaiserslautern - Fachbereich Maschinenbau und Verfahrenstechnik (4)
- Kaiserslautern - Fachbereich Wirtschaftswissenschaften (3)
- Kaiserslautern - Fachbereich Elektrotechnik und Informationstechnik (1)
Diese Diplomarbeit gibt eine kurze Einführung in das Gebiet der Diffusionsprozesse (beschrieben als Lösungen stochastischer Differentialgleichungen) und der großen Abweichungen. Mit Methoden aus dem Gebiet der großen Abweichungen wird dann das asymptotische Verhalten des Bayesrisikos für die unterscheidung zweier Diffusionsprozesse untersucht.
The Internet has fallen prey to its most successful service, the World-Wide Web. The networksdo not keep up with the demands incurred by the huge amount of Web surfers. Thus, it takeslonger and longer to obtain the information one wants to access via the World-Wide Web.Many solutions to the problem of network congestion have been developed in distributed sys-tems research in general and distributed file and database systems in particular. The introduc-tion of caching and replication strategies has proven to help in many situations and thereforethese techniques are also applied to the WWW. Although most problems and associated solu-tions are known, some circumstances are different with the Web, forcing the adaptation ofknown strategies. This paper gives an overview about these differences and about currentlydeployed, developed, and evaluated solutions.
We derive minimax rates for estimation in anisotropic smoothness classes. This rate is attained by a coordinatewise thresholded wavelet estimator based on a tensor product basis with separate scale parameter for every dimension. It is shown that this basis is superior to its one-scale multiresolution analog, if different degrees of smoothness in different directions are present.; As an important application we introduce a new adaptive wavelet estimator of the time-dependent spectrum of a locally stationary time series. Using this model which was resently developed by Dahlhaus, we show that the resulting estimator attains nearly the rate, which is optimal in Gaussian white noise, simultaneously over a wide range of smoothness classes. Moreover, by our new approach we overcome the difficulty of how to choose the right amount of smoothing, i.e. how to adapt to the appropriate resolution, for reconstructing the local structure of the evolutionary spectrum in the time-frequency plane.
We present a distributed system, Dott, for approximately solving the Trav-eling Salesman Problem (TSP) based on the Teamwork method. So-calledexperts and specialists work independently and in parallel for given time pe-riods. For TSP, specialists are tour construction algorithms and experts usemodified genetic algorithms in which after each application of a genetic operatorthe resulting tour is locally optimized before it is added to the population. Aftera given time period the work of each expert and specialist is judged by a referee.A new start population, including selected individuals from each expert and spe-cialist, is generated by the supervisor, based on the judgments of the referees.Our system is able to find better tours than each of the experts or specialistsworking alone. Also results comparable to those of single runs can be found muchfaster by a team.
The intuitionistic calculus mj for sequents, in which no other logical symbols than those for implication and universal quantification occur, is introduced and analysed. It allows a simple backward application, called mj-reduction here, for searching for derivation trees. Terms needed in mj-reduction can be found with the unification algorithm. mj-Reduction with unification can be seen as a natural extension of SLD-resolution. mj-Derivability of the sequents considered here coincides with derivability in Johansson's minimal intuitionistic calculus LHM in [6]. Intuitionistic derivability of formulae with negation and classical derivability of formulae with all usual logical symbols can be expressed with mj-derivability and hence be verified by mj-reduction. mj-Derivations can be easily translated into LJ-derivations without
"Schnitt", or into NJ-derivations in a slightly sharpened form of Prawitz' normal form. In the first three sections, the systematic use of mj-reduction for proving in predicate logic is emphasized. Although the fourth section, the last and largest, is exclusively devoted to the mathematical analysis of the calculus mj, the first three sections may be of interest to a wider readership, including readers looking for applications of symbolic logic. Unfortunately, the mathematical analysis of the calculus mj, as the study of Gentzen's calculi, demands a large amount of technical work that obscures the natural unfolding of the argumentation. To alleviate this, definitions and theorems are completely embedded in the text to provide a fluent and balanced mathematical discourse: new concepts are indicated with bold-face, proofs of assertions are outlined, or omitted when it is assumed that the reader can provide them.
Primary decomposition of an ideal in a polynomial ring over a field belongs to the indispensable theoretical tools in commutative algebra and algebraic geometry. Geometrically it corresponds to the decomposition of an affine variety into irreducible components and is, therefore, also an important geometric concept.The decomposition of a variety into irreducible components is, however, slightly weaker than the full primary decomposition, since the irreducible components correspond only to the minimal primes of the ideal of the variety, which is a radical ideal. The embedded components, although invisible in the decomposition of the variety itself, are, however, responsible for many geometric properties, in particular, if we deform the variety slightly. Therefore, they cannot be neglected and the knowledge of the full primary decomposition is important also in a geometric context.In contrast to the theoretical importance, one can find in mathematical papers only very few concrete examples of non-trivial primary decompositions because carrying out such a decomposition by hand is almost impossible. This experience corresponds to the fact that providing efficient algorithms for primary decomposition of an ideal I ae K[x1; : : : ; xn], K a field, is also a difficult task and still one of the big challenges for computational algebra and computational algebraic geometry.All known algorithms require Gr"obner bases respectively characteristic sets and multivariate polynomial factorization over some (algebraic or transcendental) extension of the given field K. The first practical algorithm for computing the minimal associated primes is based on characteristic sets and the Ritt-Wu process ([R1], [R2], [Wu], [W]), the first practical and general primary decomposition algorithm was given by Gianni, Trager and Zacharias [GTZ]. New ideas from homological algebra were introduced by Eisenbud, Huneke and Vasconcelos in [EHV]. Recently, Shimoyama and Yokoyama [SY] provided a new algorithm, using Gr"obner bases, to obtain the primary decompositon from the given minimal associated primes.In the present paper we present all four approaches together with some improvements and with detailed comparisons, based upon an analysis of 34 examples using the computer algebra system SINGULAR [GPS]. Since primary decomposition is a fairly complicated task, it is, therefore, best explained by dividing it into several subtasks, in particular, while sometimes only one of these subtasks is needed in practice. The paper is organized in such a way that we consider the subtasks separately and present the different approaches of the above-mentioned authors, with several tricks and improvements incorporated. Some of these improvements and the combination of certain steps from the different algorithms are essential for improving the practical performance.
In this report we treat an optimization task, which should make the choice of nonwoven for making diapers faster. A mathematical model for the liquid transport in nonwoven is developed. The main attention is focussed on the handling of fully and partially saturated zones, which leads to a parabolic-elliptic problem. Finite-difference schemes are proposed for numerical solving of the differential problem. Paralle algorithms are considered and results of numerical experiments are given.
We show that the occupation measure on the path of a planar Brownian motion run for an arbitrary finite time intervalhas an average density of order three with respect to thegauge function t^2 log(1/t). This is a surprising resultas it seems to be the first instance where gauge functions other than t^s and average densities of order higher than two appear naturally. We also show that the average densityof order two fails to exist and prove that the density distributions, or lacunarity distributions, of order threeof the occupation measure of a planar Brownian motion are gamma distributions with parameter 2.
We describe a platform for the portable and secure execution of mobile agents writtenin various interpreted languages on top of a common run-time core. Agents may migrate at anypoint in their execution, fully preserving their state, and may exchange messages with otheragents. One system may contain many virtual places, each establishing a domain of logicallyrelated services under a common security policy governing all agents at this place. Agents areequipped with allowances limiting their resource accesses, both globally per agent lifetime andlocally per place. We discuss aspects of this architecture and report about ongoing work.
Sudakov's typical marginals, random linear functionals and a conditional central limit theorem
(1997)
V.N. Sudakov [Sud78] proved that the one-dimensional marginals of a highdimensional second order measure are close to each other in most directions. Extending this and a related result in the context of projection pursuit of P. Diaconis and D. Freedman [Dia84], we give for a probability measure P and a random (a.s.) linear functional F on a Hilbert space simple sufficient conditions under which most of the one-dimensional images of P under F are close to their canonical mixture which turns out to be almost a mixed normal distribution. Using the concept of approximate conditioning we deduce a conditional central limit theorem (theorem 3) for random averages of triangular arrays of random variables which satisfy only fairly weak asymptotic orthogonality conditions.
Um stationäre bzw. quasi-stationäre Ohmsche Ströme in leitenden Medien berechnen zu können, wird aus komplexifizierten Maxwellschen Gleichungen mittels des Clifford Produktes eine vereinheitlichte hyperkomplexe Feldgleichung hergeleitet. Für, längs einer Achse translationsinvariante, komplexe Leitfähigkeitsfelder wird eine Dimension absepariert und die verbleibenden 2 Raumdimensionen mit der komplexen Zahlenebene identifiziert. Diese Identifikation kann durch den Clifford Formalismus explizit und völlig kanonisch definiert werden, da sowohl die komplexen Zahlen als auch Ortsvektoren in der Clifford Algebra enthalten sind. Da direkt die Spinor Feldgleichung gelöst wird, treten Eichprobleme, wie sie bei entsprechenden Potentialgleichungen üblich sind, erst gar nicht auf. Durch die Liftung der Spinor Feldgleichung vom \(\mathbb{R}^2 \to \mathbb{C}^2\) wird sofort ersichtlich wie wichtig monogene (holomorphe) Funktionen für die Lösung dieser Gleichung sind.
Die zugehörige Randbedingung ist im allgemeinen weder rein vom Neumannschen noch vom Diricheltschen Typ. Ausgehend von elementaren Lösungen für \(\delta\)-Quellen in Gebieten konstanter Leitfähigkeit, werden durch Fortsetzung dieser Lösungen mittels der Randbedingung Feldlösungen für zusammengesetzte Gebiete konstruiert.
Im Gegensatz zu Gebieten mit nur einem Rand, ist es für mehrfach berandete Gebiete viel schwieriger, die lokalen Lösungen so anzupassen, daß alle Randbedingungen erfüllt sind. Deshalb wird eine neue Lösungsmethode vorgestellt, welche die lokalen Feldgleichungen und alle Randbedingungen durch sukzessive Konstruktion von Spiegelpolreihen löst. Dieses Verfahren wird anhand einiger Klassen von geometrischen Konfigurationen erläutert, deren topologische Unterschiede sich direkt auf die Struktur der Spiegelpolverteilungen auswirkt.
Bei der Diskussion wird besonders der Fall von N kreisförmigen Anomalien in einer Kreisscheibe hervorgehoben, da diese Klasse von Problemen auch von besonderem Interesse in der medizinischen Physik, im Bereich der Impedanz-Tomographie ist. Die Lösungen erlauben die Variation der Zusammenhangszahl über die relativen Leitfähigkeitsdifferenzen. Studien der Potentialverteilung auf dem Rand, wie sie für die elektrische Impedanz-Tomographie wesentlich sind, werden zum Teil durch numerische, als auch durch analytische Berechnungen durchgeführt. Komplexe Potentiale können aus den Feldlösungen leicht berechnet werden, indem die typischen Polterme \(\displaystyle{1 \over z-p}\) durch die komplexen Logarithmen \(- \log(z-p)\)
ersetzt werden.
Das elektrische Potential ergibt sich aus dem Komplexen als dessen Realteil. Der Imaginärteil hat eine große Bedeutung bei der Visualisierung der Vektorfelder. Es wird gezeigt, daß die Höhenlinien dieses Imaginärteils, der aus der Strömungsmechanik auch als Strömungsfunktion bekannt ist, gerade die Feldlinien des zugehörigen Feldes liefert.
Für die elektrische Impedanz-Tomographie wird am Beispiel einer kleinen, konzentrisch positionierten Anomalie das Auflösungsvermögen diskutiert, woraus unter anderem eine optimale Lage der Einprägepole resultiert. Aus den analytischen Ergebnissen ist eindeutig zu erkennen, daß sich maximale Potentialänderungen auf dem Rand bei diametral angeordneten Einprägepolen ergeben.
Die für die Visualisierung der Felder nötigen Studien von Strömungsfunktionen, lieferte unter anderem auch eine Berechnungsmöglichkeit von Strömungsfunktionen für Felder im \(\mathbb{R}^3\)! Des weitern wird eine mögliche Wahl der Schnitte dieser mehrblättrigen Funktion für den Fall der Kreisscheibe mit N Anomalien explizit gegeben und die Vorteile dieser speziellen Wahl anhand numerischer Studien aufgezeigt. Typische Darstellungen von Feld- und Potentiallinien, von Verteilungen von Spiegelpolen, sowie von Potential und Strömungsfunktionen selbst, verdeutlichen die Vorteile dieses Lösungsverfahrens. Für sehr viele, in der Praxis wichtige Konfigurationen ist vor allem die große Konvergenzgeschwindigkeit ein Vorteil, welcher es ermöglicht Feldlinienbilder dieser Lösungen in kurzer Zeit auf einem PC zu erstellen.
The concept of algebraic simplification is of great importance for the field of symbolic computation in computer algebra. In this paper we review somefundamental concepts concerning reduction rings in the spirit of Buchberger. The most important properties of reduction rings are presented. Thetechniques for presenting monoids or groups by string rewriting systems are used to define several types of reduction in monoid and group rings. Gröbnerbases in this setting arise naturally as generalizations of the corresponding known notions in the commutative and some non-commutative cases. Severalresults on the connection of the word problem and the congruence problem are proven. The concepts of saturation and completion are introduced formonoid rings having a finite convergent presentation by a semi-Thue system. For certain presentations, including free groups and context-free groups, theexistence of finite Gröbner bases for finitely generated right ideals is shown and a procedure to compute them is given.
Static magnetic and spin wave properties of square lattices of permalloy micron dots with thicknesses of 500 Å and 1000 Å and with varying dot separations have been investigated. The spin wave frequencies can be well described taking into account the demagnetization factor of each single dot. A magnetic four-fold anisotropy was found for the lattice with dot diameters of 1 micrometer and a dot separation of 0.1 micrometer. The anisotropy is attributed to an anisotropic dipole-dipole interaction between magnetically unsaturated parts of the dots. The anisotropy strength (order of 100000 erg/cm^3 ) decreases with increasing in-plane applied magnetic field.
In modern approximation methods linear combinations in terms of (space localizing) radial basis functions play an essential role. Areas of application are numerical integration formulas on the uni sphere omega corresponding to prescribed nodes, spherical spline interpolation, and spherical wavelet approximation. the evaluation of such a linear combination is a time consuming task, since a certain number of summations, multiplications and the calculation of scalar products are required. This paper presents a generalization of the panel clustering method in a spherical setup. The economy and efficiency of panel clustering is demonstrated for three fields of interest, namely upward continuation of the earth's gravitational potential, geoid computation by spherical splines and wavelet reconstruction of the gravitational potential.
Das Modell des Intelligenten ist eine Abstraktion von Telefonvermittlungs-systemen und beschreibt auch deren Erweiterungen. Zunächst wird ein einfachesBasissystem spezifiziert, das dann um weitere Leistungsmerkmale, sog. Features, erweitert wird. Im Rahmen dieser Arbeit haben wir ein bereits bestehendes, in Estellespezifiziertes Basissystem um sechs Features erweitert. Dabei konnten wir verschiedene Stile für die Featurespezifikation in Estelle überprüfen. Wir entwerfen Prinzipien füreine verhaltenerhaltende Transformation, die geeignete Ansatzpunkte für neueFeatures schaffen kann. Für das Ergänzen von neuen Rufnummern haben wir eine einfache Methode entwickelt. Wir zeigen zwei Schwächen von Estelle beim Erweitern vonSystemen auf. Schließlich berichten wir über unsere Erfahrungen mit dem im IN-Modellverwendeten Prinzip der Detection Points.
Viele Entwicklungsprozesse, wie sie z.B. beim Entwurf von grossen Softwaresystemen benötigt werden, basieren in erster Linie auf dem Wissen der mit der Entwicklung betrauten Mitarbeiter. Mit wachsender Komplexität der Entwurfsaufgaben und mit wachsender Anzahl der Mitarbeiter in einem Projekt wird die Koordination und Verteilung dieses Wissens immer problematischer. Aus diesem Grund versucht man zunehmend, das Wissen der Mitarbeiter in elektronischer Form, d.h. in Rechnern zu speichern und zu verwalten. Dadurch, dass der Entwurf eines komplexen Systems ebenfalls am Rechner modelliert wird, steht benötigtes Wissen sofort zur Verfügung und kann zur Entscheidungsunterstützung herangezogen werden. Gerade bei der Planung grosser Projekte stehen jedoch oft Entscheidungen aus, die erst später, während der Abwicklung getroffen werden können. Da gängige Workflow-Management-System zumeist eine komplette Modellierung verlangen, bevor die Abwicklung eines Projektmodells beginnen kann, habt sich dieser Ansatz gerade für umfangreiche Projekte als eher ungeeignet herausgestellt.
In this paper a group of participants of the 12th European Summer Institute which took place in Tenerifa, Spain in June 1995 present their views on the state of the art and the future trends in Locational Analysis. The issue discussed includes modelling aspects in discrete, network and continuous location, heuristic techniques, the state of technology and undesirable facility location. Some general questions are stated reagrding the applicability of location models, promising research directions and the way technology affects the development of solution techniques.
Sokrates und das Nichtwissen
(1997)
Software Products As Objects
(1997)
This paper describes our experiences in modeling entire software products (trees of software files) as objects. Container pnodes (product nodes) have user-defined Internetunique names, data types, and methods (operations). Pnodes can contain arbitrary collections of software files that represent programs, libraries, documents, or other software products. Pnodes can contain multiple software products, so that header files, libraries, and program products may all be stored within one pnode. Pnodes can contain views that list other pnodes in order to form large conceptual structures of pnodes. Typical pnode -object methods include: fetching and storing into version controlled repositories; dynamic analysis of pnode contents to generate makefiles of arbitrary complexity; local automated build operations; Internet-scalable distributed repository synchroni- zations; Internet-scalable, multi-platform, distributed build operations; extraction and generation of online API documen- tation, spell checking of document pnodes, and so on. Since methods are user-defined, they can be arbitrarily complex. Modelling software products as objects provides a large amount of effort leverage, since one person can define the methods and many people can use them in extensively automated ways.
Techniques for modular software design are presented applying software agents. The conceptual designs are domain independent and make use of specificdomain aspects applying Multiagent AI. The stages of conceptualization, design and implementation are defined by new techniques coordinated by objects. Software systemsare designed by knowledge acquisition, specification, and multiagent implementations.
Skyrme Sphalerons of an O(3)-oe Model and the Calculation of Transition Rates at Finite Temperature
(1997)
The reduced O(3)-oe model with an O(3) ! O(2) symmetry breaking potential is considered with an additional Skyrmionic term, i. e. a totally antisymmetric quartic term in the field derivatives. This Skyrme term does not affect the classical static equations of motion which, however, allow an unstable sphaleron solution. Quantum fluctuations around the static classical solution are considered for the determination of the rate of thermally induced transitions between topologically distinct vacua mediated by the sphaleron. The main technical effect of the Skyrme term is to produce an extra measure factor in one of the fluctuation path integrals which is therefore evaluated using a measure-modified Fourier-Matsubara decomposition (this being one of the few cases permitting this explicit calculation). The resulting transition rate is valid in a temperature region different from that of the original Skyrme-less model, and the crossover from transitions dominated by thermal fluctuations to those dominated by tunneling at the lower limit of this range depends on the strength of the Skyrme coupling.
For periodically driven systems, quantum tunneling between classical resonant stability islands in phase space separated by invariant KAM curves or chaotic regions manifests itself by oscillatory motion of wave packets centered on such an island, by multiplet splittings of the quasienergy spectrum, and by phase space localisation of the quasienergy states on symmetry related ,ux tubes. Qualitatively di,erent types of classical resonant island formation | due to discrete symmetries of the system | and their quantum implications are analysed by a (uniform) semiclassical theory. The results are illustrated by a numerical study of a driven non-harmonic oscillator.
Here the self-organization property of one-dimensional Kohonen's algorithm in its 2k-neighbour setting with a general type of stimuli distribution and non-increasing learning rate is considered. We prove that the probability of self-organization for all initial values of neurons is uniformly positive. For the special case of a constant learning rate, it implies that the algorithm self-organizes with probability one.
Metaharmonic wavelets are introduced for constructing the solution of theHelmholtz equation (reduced wave equation) corresponding to Dirichlet's orNeumann's boundary values on a closed surface approach leading to exactreconstruction formulas is considered in more detail. A scale discrete version ofmultiresolution is described for potential functions metaharmonic outside theclosed surface and satisfying the radiation condition at infinity. Moreover, wediscuss fully discrete wavelet representations of band-limited metaharmonicpotentials. Finally, a decomposition and reconstruction (pyramid) scheme foreconomical numerical implementation is presented for Runge-Walsh waveletapproximation.
Ist "Programmieren ganz ohne Code" auch im CAx-Bereich möglich? Die Vielzahl heterogener CAx-Anwendungen und die wachsende Komplexität der Entwicklungsprozesse bedarf neuer Lösun-gen in der CAx-Technik. Ziel dieses Beitrages ist es, die richtungsweisende Rolle der Komponenten-technologie im CAx-Bereich aufzuzeigen. Es werden die Grundlagen der Komponenten sowie die wichtigen Komponentenarchitekturen (ActiveX und Java Beans) vorgestellt. Die Erwartungen der Anwender und der Systemhersteller, die Potentiale und die Auswirkungen dieser Technologie auf die neuen Systeme werden analysiert. Die zur Zeit verfügbaren ersten Ansätze werden präsentiert. Die Rolle der internationalen Standards für die technische Umsetzung und für die Akzeptanz von CAx-Komponentensystemen wird aufgezeigt.
The Filter-Diagonalization Method is used to ,nd the broad and even overlapping resonances of a 1D Hamiltonian used before as a test model for new resonance theories and computational methods. It is found that the use of several complex-scaled cross-correlation probability amplitudes from short time propagation enables the calculation of broad overlapping resonances, which can not be resolved from the amplitude of a single complex-scaled autocorrelation calculation.
A first explicit connection between finitely presented commutative monoids and ideals in polynomial rings was used 1958 by Emelichev yielding a solution tothe word problem in commutative monoids by deciding the ideal membership problem. The aim of this paper is to show in a similar fashion how congruenceson monoids and groups can be characterized by ideals in respective monoid and group rings. These characterizations enable to transfer well known resultsfrom the theory of string rewriting systems for presenting monoids and groups to the algebraic setting of subalgebras and ideals in monoid respectively grouprings. Moreover, natural one-sided congruences defined by subgroups of a group are connected to one-sided ideals in the respective group ring and hencethe subgroup problem and the ideal membership problem are directly related. For several classes of finitely presented groups we show explicitly howGröbner basis methods are related to existing solutions of the subgroup problem by rewriting methods. For the case of general monoids and submonoidsweaker results are presented. In fact it becomes clear that string rewriting methods for monoids and groups can be lifted in a natural fashion to definereduction relations in monoid and group rings.
Many problems arising in (geo)physics and technology can be formulated as compact operator equations of the first kind \(A F = G\). Due to the ill-posedness of the equation a variety of regularization methods are in discussion for an approximate solution, where particular emphasize must be put on balancing the data and the approximation error. In doing so one is interested in optimal parameter choice strategies. In this paper our interest lies in an efficient algorithmic realization of a special class of regularization methods. More precisely, we implement regularization methods based on filtered singular value decomposition as a wavelet analysis. This enables us to perform, e.g., Tikhonov-Philips regularization as multiresolution. In other words, we are able to pass over from one regularized solution to another one by adding or subtracting so-called detail information in terms of wavelets. It is shown that regularization wavelets as proposed here are efficiently applicable to a future problem in satellite geodesy, viz. satellite gravity gradiometry.
Like other industries, the aircraft industry is under high pressure to meet drastically increased customer goals for market price and flexibility. This while at the same time share holders request for short term profit guarantees. Daimler-Benz Aerospace Airbus has met this challenge using business process reengineering methods which led to total company restructuring from functional orientation to customer and product orientation. This paper will show how business process modelling techniques have been applied. Especially concurrent engineering methods are used to integrate the various disciplines involved from market analysts over design, commercial to industrialization staff.
This paper discusses the benefits and drawbacks of caching and replication strategies in the WWW with respect to the Internet infrastructure. Bandwidth consumption, latency, and overall error rates are considered to be most important from a network point of view. The dependencies of these values with input parameters like degree of replication, document popularity, actual cache hit rates, and error rates are highlighted. In order to determine the influence of different caching and replication strategies on the behavior of a single proxy server with respect to these values, trace-based simulations are used. Since the overall effects of such strate- gies can hardly be decided with this approach alone, a mathematical model has been developed to deal with their influence on the network as a whole. Together, this two-tiered approach permits us to propose quantita- tive assessments on the influence different caching and replication proposals (are going to) have on the Inter- net infrastructure.
Process Chain in Automotive Industry - Present Day Demands versus Long Term Open CAD/CAM Strategies
(1997)
The automotive industry was a pioneer in using CAD/CAM technology. Now the car manufacturers development process is almost completely done with this technology. Substantial initiative for the standardisation of CAD/CAM technics comes from the automotive industry, as e.g. for neutral CAD data interfaces. The R&D departments of German car manufacturers have founded a working group ii with the aim to develop a common long term CAD/CAM strategy. One important result is the concept of a future CAx iii architecture based on the standard data structure STEP iv . The commitment of the car manufactures to STEP and open system architectures is in contradiction to their attitude towards suppliers and subcontractors: Recently, more and more contractors are contractually bound to use exactly the same CAD system as the orderer. The German car industry tries to find a way out of this contradiction and to improve the co-operation between the companies in short term. Therefore they proposed a "Dual CAD Strategy", i.e. to put improvements in CAD communication into practice which are possible today - even proprietary solutions - and in parallel to invest in strategic concepts to prepare tomorrow's open system landscape.
The problem of constructing a geometric model of an existing object from a set of boundary points arises in many areas of industry. In this paper we present a new solution to this problem which is an extension of Boissonnat's method [2]. Our approach uses the well known Delaunay triangulation of the data points as an intermediate step. Starting with this structure, we eliminate tetrahedra until we get an appropriate approximation of the desired shape. The method proposed in this paper is capable of reconstructing objects with arbitrary genus and can cope with different point densities in different regions of the object. The
problems which arise during the elimination process, i.e. which tetrahedra can be eliminated, which order has to be used to control the process and finally, how to stop the elimination procedure at the right time, are discussed in detail. Several examples are given to show the validity of the method.
This paper provides a description of PLATIN. With PLATIN we present an imple-mented system for planning inductive theorem proofs in equational theories that arebased on rewrite methods. We provide a survey of the underlying architecture ofPLATIN and then concentrate on details and experiences of the current implementa-tion.
This paper presents the different possibilities for parallel processing in robot control architectures. At the beginning, we shortly review the historic development of control architectures. Then, a list of requirements for control architectures is set up from a parallel processing point of view. As our main topic, we identify the levels of parallel processing in robot control architectures. With each level of parallelism, examples for a typical robot control architecture are presented. Finally, a list of keywords is provided for each previous work we refer to.
The observation of an ergodic Markov chain asymptotically allows perfect identification of the transition matrix. In this paper we determine the rate of the information contained in the first n observations, provided the unknown transition matrix belongs to a known finite set. As an essential tool we prove new refinements of the large deviation theory of the empirical pair measure of finite Markov chains. Keywords: Markov Chain, Entropy, Bayes risk, Large Deviations.
In the modeling of biological phenomena, in living organisms whether the measurements are of blood pressure, enzyme levels, biomechanical movements or heartbeats, etc., one of the important aspects is time variation in the data. Thus, the recovery of a "smooth" regression or trend function from noisy time-varying sampled data becomes a problem of particular interest. Here we use non-linear wavelet thresholding to estimate a regression or a trend function in the presence of additive noise which, in contrast to most existing models, does not need to be stationary. (Here, nonstationarity means that the spectral behaviour of the noise is allowed to change slowly over time.). We develop a procedure to adapt existing threshold rules to such situations, e.g., that of a time-varying variance in the errors. Moreover, in the model of curve estimation for functions belonging to a Besov class with locally stationary errors, we derive a near-optimal rate for the L2-risk between the unknown function and our soft or hard threshold estimator, which holds in the general case of an error distribution with bounded cumulants. In the case of Gaussian errors, a lower bound on the asymptotic minimax rate in the wavelet coefficient domain is also obtained. Also it is argued that a stronger adaptivity result is possible by the use of a particular location and level dependent threshold obtained by minimizing Stein's unbiased estimate of the risk. In this respect, our work generalizes previous results, which cover the situation of correlated, but stationary errors. A natural application of our approach is the estimation of the trend function of nonstationary time series under the model of local stationarity. The method is illustrated on both an interesting simulated example and a biostatistical data-set, measurements of sheep luteinizing hormone, which exhibits a clear nonstationarity in its variance.
MP Prototype Specification
(1997)
We study the problem of global solution of Fredholm integral equations. This means that we seek to approximate the full solution function (as opposed to the local problem, where only the value of the solution in a single point or a functional of the solution is sought). We analyze the Monte Carlo complexity, i.e. the complexity of stochastic solution of this problem. The framework for this analysis is provided by information based complexity theory. Our investigations complement previous ones on stochastic complexity of local solution and on deterministic complexity of
both local and global solution. The results show that even in the global case Monte Carlo algorithms can perform better than deterministic ones, although the difference is not as large as in the local case.
In this paper we provide a semantical meta-theory that will support the development of higher-order calculi for automated theorem proving like the corresponding methodology has in first-order logic. To reach this goal, we establish classes of models that adequately characterize the existing theorem-proving calculi, that is, so that they are sound and complete to these calculi, and a standard methodology of abstract consistency methods (by providing the necessary model existence theorems) needed to analyze completeness of machine-oriented calculi.
Liegruppen
(1997)
Formale Beschreibungstechniken (FDTs) erlauben durch ihre formale Syntax und Semantik eine präzise Systembeschreibung und sind Grundlage für die formale Verifikation. Bei der Implementierung von Systemen wird jedoch nach wie vor von Hand implementiert, selbst wenn ausgereifte Werkzeuge zur automatischen Generierung von Kode direkt aus der formalen Spezifikation existieren. Die Ursache dafür liegt in dem Ruf dieser Werkzeuge, Kode mit extrem geringer Leistungsfähigkeit zu erzeugen. Es gibt jedoch kaum quantitative Leistungsvergleiche zwischen manuell und automatisch generierten Implementierungen, die dieses Vorurteil stützen oder widerlegen könnten. In diesem Beitrag wird ein solcher Leistungsvergleich anhand des Hochleistungsprotokolls XTP und der FDT Estelle vorgestellt. Er liefert eine Bestandsaufnahme des momentanen Entwicklungsstandes bei der automatischen Generierung von Kode aus Estelle-Spezifikationen im direkten Vergleich zu gut optimierten Handimplementierungen. Es zeigt sich, daß in dem betrachteten Fall eines komplexen Protokolls die Handimplementierung zwar merklich leistungsstärker ist. Dieser Leistungsvorteil wird jedoch durch einen sehr hohen Implementierungsaufwand sowie die Schwierigkeit, die Korrektheit bzgl. der Spezifikation sicherzustellen, erkauft. Im einzelnen Anwendungsfall kann es daher trotz der Leistungseinbußen durchaus vorteilhaft sein, automatisch Kode zu erzeugen, zumal in der Bestandsaufnahme festgestellt wurde, daß automatisch generierte Implementierungen z.T. besser abschneiden als erwartet. Zudem besteht - anders als bei der bereits umfassend optimierten Handimplementierung - noch ein erhebliches ungenutztes Potential zur Leistungsverbesserung der automatisch generierten Implementierung.
We report on the observation of quantized surface spin waves in periodic arrays of magnetic Ni81Fe19 wires by means of Brillouin light scattering spectroscopy. At small wavevectors (q_1 = 0 - 0.9*100000 cm^-1 ) several discrete, dispersionless modes with a frequency splitting of up to 0.9 GHz were observed for the wavevector oriented perpendicular to the wires. From the frequencies of the modes and the wavevector interval, where each mode is observed, the modes are identified as dipole-exchange surface spin wave modes of the film with quantized wavevector values determined by the boundary conditions at the lateral edges of the wires. With increasing wavevector the separation of the modes becomes smaller, and the frequencies of the discrete modes converge to the dispersion of the dipole-exchange surface mode of a continuous film.
We investigate in how far interpolation mechanisms based on the nearest-neighbor rule (NNR) can support cancer research. The main objective is to usethe NNR to predict the likelihood of tumorigenesis based on given risk factors.By using a genetic algorithm to optimize the parameters of the nearest-neighbourprediction, the performance of this interpolation method can be improved sub-stantially. Furthermore, it is possible to detect risk factors which are hardly ornot relevant to tumorigenesis. Our preliminary studies demonstrate that NNR-based interpolation is a simple tool that nevertheless has enough potential to beseriously considered for cancer research or related research.
In dieser Arbeit wird eine Integration der temporallogischen Verarbeitungskonzepte
der Programmiersprache ExTeLL in die objektorientierte Wirtssprache \(C^{++}\) vorgestellt. Dabei war unser Ziel eine Schnittstelle zur komfortablen Kommunikation der Sprachkomponenten zu entwickeln, derart daß die Sprachsynthese eine homogene Gesamtsprache darstellt . Hierbei haben wir besonderen Wert auf die Nutzung der Möglichkeiten der jeweils hinzugefügten Sprachkomponente und einen syntaktisch einheitlichen Aufbau der Gesamtsprache gelegt. Dies erforderte insbesondere die Integration des Typkonzepts von \(C^{++}\) sowie der Mechanismen zur Überladung von Funktionen und Prozeduren in ExTeLL und in der zugrundeliegenden Temporallogik
EITeL.
Instant Radiosity
(1997)
We present a fundamental procedure for instant rendering from the radiance equation. Operating directly on the textured scene description, the very efficient and simple algorithm produces photorealistic images without any kernel or solution discretization of the underlying integral equation. Rendering rates of a few seconds are obtained by exploiting graphics hardware, the deterministic
technique of the quasi-random walk for the solution of the global illumination problem, and the new method of jittered low discrepancy sampling.
In dieser Arbeit wird die Problematik der sich rapide wandelnden industriellen CAx-Anwendungen betrachtet. Durch die Einfu"hrung der Feature-Technologie scheinen einige Probleme der Parallelisierung der Prozesse, des Simultaneous und des Concurrent Engineering sowie des Outsourcing überwindbar zu sein. Allerdings entwickelte sich die Feature-Technologie bisher ohne ausreichenden Bezug zur Konstruktionspraxis, was zu erheblichen Defiziten im industriellen Einsatz führte. Untersuchungen in der Automobilindustrie (AIFEMInitiative) zeigen, dass dies vielfach auf mangelnde Kommunikation zwischen Konstrukteuren und CAx-Experten zurückgeführt werden kann. Aufgrund des jetzigen Ansatzes der Feature-Technologie im Zusammenwirken mit dem extremen Zeitdruck in der Produktentwicklung besteht aber die Gefahr, die Produktdefinitionsprozesse nur nach den Kriterien Entwicklungszeit, Kosten und Produktqualität zu optimieren. Features dienen dabei nur als speziell angepasste Werkzeuge. Damit wird eine echte Innovation der Produkte behindert. Es wird aufgezeigt, wie die Feature-Technologie erweitert werden muss, um die Kreativität der Konstrukteure zu fördern und somit neuartige Produkte zu ermöglichen. Näher ausgeführt werden die Aspekte der benutzerdefinierten Features, der Datenstandardisierung, der Verarbeitung unvollsta"ndiger Information und der dynamischen Prozessunterstützung.
In this note, answering a question of N. Maslova, we give a two-dimensional elementary example of the phenomenon indicated in the title. Perhaps this simple example may serve as an object of comparison for more refined models like in the theory of kinetic differential equations where similar questions still seem to be unsettled.
Starting from the mollified version of the Enskog equation for a hard-sphere fluid, a grid-free algorithm to obtain the solution is proposed. The algorithm is based on the finite pointset method. For illustration, it is applied to a Riemann problem. The shock-wave solution is compared to the results of Frezzotti and Sgarra where a good agreement is found.
In the Banach space co there exists a continuous function of bounded semivariation which does not correspond to a countably additive vector measure. This result is in contrast to the scalar case, and it has consequences for the characterization of scalar-type operators. Besides this negative result we introduce the notion of functions of unconditionally bounded variation which are exactly the generators of countably additive vector measures.
The Fock space of bosons and fermions and its underlying superalgebra are represented by algebras of functions on a superspace. We define Gaussian integration on infinite dimensional superspaces, and construct superanalogs of the classical function spaces with a reproducing kernel - including the Bargmann-Fock representation - and of the Wiener-Segal representation. The latter representation requires the investigation of Wick ordering on Z 2 -graded algebras. As application we derive a Mehler formula for the Ornstein-Uhlenbeck semigroup on the Fock space.
\(C^0\)-scalar-type spectrality criterions for operators \(A\), whose resolvent set contains the negative reals, are provided. The criterions are given in terms of growth conditions on the resolvent of \(A\) and the semi-group generated by \(A\).These criterions characterize scalar-type operators on the Banach space \(X\), if and only if \(X\) has no subspace isomorphic to the space of complex null-sequences.
We present a method for making use of past proof experience called flexiblere-enactment (FR). FR is actually a search-guiding heuristic that uses past proofexperience to create a search bias. Given a proof P of a problem solved previouslythat is assumed to be similar to the current problem A, FR searches for P andin the "neighborhood" of P in order to find a proof of A.This heuristic use of past experience has certain advantages that make FRquite profitable and give it a wide range of applicability. Experimental studiessubstantiate and illustrate this claim.This work was supported by the Deutsche Forschungsgemeinschaft (DFG).
We present a general framework for developing search heuristics for au-tomated theorem provers. This framework allows for the construction ofheuristics that are on the one hand able to replay (parts of) a given prooffound in the past but are on the other hand flexible enough to deviate fromthe given proof path in order to solve similar proof problems. We substanti-ate the abstract framework by the presentation of three distinct techniquesfor learning appropriate search heuristics based on soADcalled features. Wedemonstrate the usefulness of these techniques in the area of equational de-duction. Comparisons with the renowned theorem prover Otter validatethe applicability and strength of our approach.
One of the many features needed to support the activities of autonomous systems is the ability of motion planning. It enables robots to move in their environment securely and to accomplish given tasks. Unfortunately, the control loop comprising sensing, planning, and acting has not yet been closed for robots in dynamic environments. One reason involves the long execution times of the motion planning component. A solution for this problem is offered by the use of highly computational parallelism. Thus, an important task is the parallelization of existing motion planning algorithms for robots so that they are suitable for highly computational parallelism. In several cases, completely new algorithms have to be designed, so that a parallelization is feasible. In this survey, we review recent approaches to motion planning using parallel computation. As a classification scheme, we use the structure given by the different approaches to the robot's motion planning. For each approach, the available parallel processing methods are discussed. Each approach is uniquely assigned a class. Finally, for each referenced research work, a list of keywords is given.
A formula suitable for a quantitative evaluation of the tunneling effect in a ferromagnetic particle is derived with the help of the instanton method. The tunneling between n-th degenerate states of neighboring wells is dominated by a periodic pseudoparticle configuration. The low-lying level-splitting previously obtained with the LSZ method in field theory in which the tunneling is viewed as the transition of n bosons induced by the usual(vacuum) instanton is recovered.The observation made with our new result is that the tunneling effect increases at excited states. The results should be useful in analyzing results of experimental tests of macroscopic quantum coherence in ferromagnetic particles.
Die Bewegungsplanung für Industrieroboter ist eine notwendige Voraussetzung, damit sich autonome Systeme kollisionsfrei durch die Umwelt bewegen können. Die Berücksichtigung von dynamischen Hindernissen zur Laufzeit erfordert allerdings leistungsfähige Algorithmen, zur Lösung dieser Aufgabenstellung in Echtzeit. Eine Möglichkeit zur Beschleunigung der Algorithmen ist der effiziente Einsatz von skalierbarer Parallelverarbeitung. Die softwaretechnische Umsetzung kann aber nur dann erfolgreich sein, wenn ein Parallelrechner zur Verfügung steht, der einen hohen Datendurchsatz bei geringer Latenzzeit bietet. Darüber hinaus muß dieser Parallelrechner unter vertretbarem Aufwand bedienbar sein und ein gutes Preisleistungsverhältnis aufweisen, damit die Parallelverarbeitung verstärkt in der Industrie zum Einsatz kommt. In diesem Artikel wird ein Workstation-Cluster auf der Basis von neun Standard- PCs vorgestellt, die über eine spezielle Kommunikationskarte miteinander vernetzt sind. In den einzelnen Abschnitten werden die gesammelten Erfahrungen bei der Inbetriebnahme, Systemadministration und Anwendung geschildert. Als Beispiel für eine Anwendung auf diesem Cluster wird ein paralleler Bewegungsplaner für Industrieroboter beschrieben.
The first observation of self-focusing of dipolar spin waves in garnet film media is reported. In particular, we show that the quasi-stationary diffraction of a finite-aperture spin wave beam in a focusing medium leads to the concentration of the wave power in one focal point rather than along a certain line (channel). The obtained results demonstrate the wide applicability of non-linear spin wave media to study non-linear wave phenomena using an advanced combined microwave-Brillouin light scattering technique for a two-dimensional mapping of the spin wave amplitudes.
We compare different notions of differentiability of a measure along a vector field on a locally convex space. We consider in the \(L^2\)-space of a differentiable measure the analoga of the classical concepts of gradient, divergence and Laplacian (which coincides with the Ornstein-Uhlenbeck
operator in the Gaussian case). We use these operators for the extension of the basic results of Malliavin and Stroock on the smoothness of finite dimensional image measures under certain nonsmooth mappings to the case of non-Gaussian measures. The proof of this extension is quite direct and does not use any Chaos-decomposition. Finally, the role of this Laplacian in the
procedure of quantization of anharmonic oscillators is discussed.
An analogue of the classical Riemann-Siegel integral formula for Dirichlet series associated to cusp forms is developed. As an application of the formula, we give a comparatively simple proof of the approximate functional equation for this type of Dirichlet series.
The tunneling splitting of the energy levels of a ferromagnetic particle in the presence of an applied magnetic field - previously derived only for the ground state with the path integral method - is obtained in a simple way from Schr"odinger theory. The origin of the factors entering the result is clearly understood, in particular the effect of the asymmetry of the barriers of the potential. The method should appeal particularly to experimentalists searching for evidence of macroscopic spin tunneling.
Brillouin light scattering investigations of exchange biased (110)-oriented NiFe/FeMn bilayers
(1997)
All contributing magnetic anisotropies in (110)-oriented exchange biased Ni 80 Fe 20 /Fe 50 Mn 50 double layers prepared by molecular beam epitaxy on Cu(110) single crystals have been determined by means of Brillouin light scattering. Upon covering the Ni 80 Fe 20 films by Fe 50 Mn 50 , a unidirectional anisotropy contribution appears, which is consistent with the measured exchange bias field. The uniaxial and fourfold in-plane anisotropy contributions are largely modified by an amount, which scales with the Ni 80 Fe 20 thickness, indicating an interface effect. The strong uniaxial anisotropy contribution shows an in-plane switching of the easy axis from [110] to [001] with increasing Ni 80 Fe 20 -layer thickness. The large mode width of the spin wave excitations, which exceeds the linewidth of uncovered Ni 80 Fe 20 films by a factor of more than six, indicates large spatial variations of the exchange coupling constant. (C) 1998 American Institute of Physics.
An unusual interlayer coupling, recently discovered in layered magnetic systems, is analysed from the experimental and theoretical points of view. This coupling favours the 90 orientation of the magnetization of the adjacent magnetic films. It can be phenomenologically described by a term in the energy expression, which is biquadratic with respect to the magnetizations of the two films. The main experimental findings, as well as the theoretical models, explaining the phenomenon are discussed.
Fabric reinforced thermoplastic composites, suitable for the production of thin-walled, high
strength structural parts, are available on the market today with various fibre/matrix combinations.
However, further market penetration and series production are inhibited as long as forming
technologies are not well understood. In this thesis, the potential for series production of different
forming technologies is evaluated. Stamp forming is an efficient way to produce parts in
very short cycle times. A limiting factor to part complexibilty is undesired wrinkle formation as
a consequence of insufficient fabric shear. Fabric shear and other important deformations of impregnated
fabrics were examined by means of new test devices. Evidence was found that membrane
tension is the crucial factor to avoid wrinkle formation. New tool concepts and processing
Windows were developed to produce fabric reinforced thermoplastic parts free of wrinkles and
distortions.
Annual Report
(1997)
Static magnetic and spin wave properties of square lattices of permalloy micron dots with thicknesses of 500 Å and 1000 Å and with varying dot separations have been investigated. A magnetic fourfold anisotropy was found for the lattice with dot diameters of 1 micrometer and a dot separation of 0.1 micrometer. The anisotropy is attributed to an anisotropic dipole-dipole interaction between magnetically unsaturated parts of the dots. The anisotropy strength (order of 100000 erg/cm^3 ) decreases with increasing in-plane applied magnetic field.
This paper describes an Internet-scalable knowledge base infrastructure for managing the knowledge used by an in-telligent software productivity infrastructure system. The infrastructure provides workable solutions for several significant issues: (1) Internetunique names for pieces of knowledge; (2) multi-platform, multi-language support; (3) distributed knowledge base synchronization mechanisms; (4) support for extensive customized variations in knowledge content, and (5) knowledge caching mechanisms for improved system performance. The infrastructure described here is a workable example of the kind of infrastructure that will be required to manage the evolution and reuse of millions of pieces of knowledge in the future.
An asymptotic-induced scheme for nonstationary transport equations with thediffusion scaling is developed. The scheme works uniformly for all ranges ofmean free paths. It is based on the asymptotic analysis of the diffusion limit ofthe transport equation. A theoretical investigation of the behaviour of thescheme in the diffusion limit is given and an approximation property is proven.Moreover, numerical results for different physical situations are shown and atheuniform convergence of the scheme is established numerically.
This paper is a continuation of a joint paper with B. Martin [MS] dealing with the problem of direct sum decompositions. The techniques of that paper areused to decide wether two modules are isomorphic or not. An positive answer to this question has many applications - for example for the classification ofmaximal Cohen-Macaulay module over local algebras as well as for the study of projective modules. Up to now computer algebra is normally dealing withequality of ideals or modules which depends on chosen embeddings. The present algorithm allows to switch to isomorphism classes which is more natural inthe sense of commutative algebra and algebraic geometry.
This paper shows an approach to profit from type information about planning objects in a partial-order planner. The approach turns out to combine representational and computational advantages. On the one hand, type hierarchies allow better structuring of domain specifications. On the other hand, operators contain type constraints which reduce the search space of the planner as they partially achieve the functionality of filter conditions.
We report on Brillouin light scattering investigations of the elastic properties in Co/Ni superlattices which exhibit localized electronic eigenstates near the Fermi level causing an oscillation of the resistivity as a function of the superlattice periodicity A. No oscillations of the Rayleigh and Sezawa mode as a function of A could be observed within an error margin of +- 2% indicating that the localized electronic states do not contribute to the elastic constants.
We develop a test for stationarity of a time series against the alternative of a time-changing covariance structure. Using localized versions of the periodogram, we obtain empirical versions of a reasonable notion of a time-varying spectral density. Coefficients w.r.t. a Haar wavelet series expansion of such a time-varying periodogram are a possible indicator whether there is some deviation from covariance stationarity. We propose a test based on the limit distribution of these empirical coefficients.
Due to continuously increasing demands in the area of advanced robot control, it became necessary to speed up the computation. One way to reduce the computation time is to distribute the computation onto several processing units. In this survey we present different approaches to parallel computation of robot kinematics and Jacobian. Thereby, we discuss both the forward and the reverse problem. We introduce a classification scheme and classify the references by this scheme.
An asymptotic-induced scheme for kinetic semiconductor equations with the diffusion scaling is developed. The scheme is based on the asymptotic analysis of the kinetic semiconductor equation. It works uniformly for all ranges of mean free paths. The velocity discretization is done using quadrature points equivalent to a moment expansion method. Numerical results for different physical situations are presented.
The Multiple Objective Median Problem involves locating a new facility so that a vector of performance criteria is optimized over a given set of existing facilities. A variation of this problem is obtained if the existing facilities are situated on two sides of a linear barrier. Such barriers like rivers, highways, borders, or mountain ranges are frequently encountered in practice. In this paper, theory of the Multiple Objective Median Problem with line barriers is developped. As this problem is nonconvex but specially-structured, a reduction to a series of convex optimization problems is proposed. The general results lead to a polynomial algorithm for finding the set of efficient solutions. The algorithm is proposed for bi-criteria problems with different measures of distance.
It is of basic interest to assess the quality of the decisions of a statistician, based on the outcoming data of a statistical experiment, in the context of a given model class P of probability distributions. The statistician picks a particular distribution P , suffering a loss by not picking the 'true' distribution P' . There are several relevant loss functions, one being based on the the relative entropy function or Kullback Leibler information distance. In this paper we prove a general 'minimax risk equals maximin (Bayes) risk' theorem for the Kullback Leibler loss under the hypothesis of a dominated and compact family of distributions over a Polish observation space with suitably integrable densities. We also find that there is always an optimal Bayes strategy (i.e. a suitable prior) achieving the minimax value. Further, we see that every such minimax optimal strategy leads to the same distribution P in the convex closure of the model class. Finally, we give some examples to illustrate the results and to indicate, how the minimax result reflects in the structure of least favorable priors. This paper is mainly based on parts of this author's doctorial thesis.
Estelle is an internationally standardized formal description technique (FDT) designed for the specification of distributed systems, in particular communication protocols. An Estelle specification describes a system of communicating components (module instances). The specified system is closed in a topological sense, i.e. it has no ability to interact with some environment. Because of this restriction, open systems can only be specified together with and incorporated with an environment. To overcome this restriction, we introduce a compatible extension of Estelle, called "Open Estelle". It allows the specification of (topologically) open systems, i.e. systems that have the ability to communicate with any environment through a well-defined external interface. We define aformal syntax and a formal semantics for Open Estelle, both based on and extending the syntax and semantics of Estelle. The extension is compatible syntactically and semantically, i.e. Estelle is a subset of Open Estelle. In particular, the formal semantics of Open Estelle reduces to the Estelle semantics in the special case of a closed system. Furthermore, we present a tool for the textual integration of open systems into environments specified in Open Estelle, and a compiler for the automatic generation of implementations directly from Open Estelle specifications.
Retrieving multiple cases is supposed to be an adequate retrieval strategy for guiding partial-order planners because of the recognized flexibility of these planners to interleave steps in the plans. Cases are combined by merging them. In this paper, we will examine two different kinds of merging cases in the context of partial-order planning. We will see that merging cases can be very difficult if the cases are merged eagerly. On the other hand, if cases are merged by avoiding redundant steps, the guidance of the additional cases tends to decrease with the number of covered goals and retrieved cases in domains having a certain kind of interactions. Thus, to retrieve a single case covering many of the goals of the problem or to retrieve fewer cases covering many of the goals is at least equally effective as to retrieve several cases covering all goals in these domains.