Kaiserslautern - Fachbereich Mathematik
Refine
Year of publication
Document Type
- Preprint (608)
- Doctoral Thesis (292)
- Report (121)
- Article (84)
- Diploma Thesis (26)
- Lecture (25)
- Master's Thesis (8)
- Course Material (5)
- Part of a Book (4)
- Study Thesis (4)
Keywords
- Mathematische Modellierung (19)
- MINT (16)
- Schule (16)
- Wavelet (14)
- Inverses Problem (12)
- Mehrskalenanalyse (12)
- Modellierung (12)
- Mathematikunterricht (9)
- praxisorientiert (9)
- Approximation (8)
- Boltzmann Equation (8)
- Regularisierung (8)
- Lineare Algebra (7)
- Location Theory (7)
- Numerical Simulation (7)
- Optimization (7)
- integer programming (7)
- Algebraische Geometrie (6)
- Finanzmathematik (6)
- Gravitationsfeld (6)
- Navier-Stokes-Gleichung (6)
- Portfolio Selection (6)
- modelling (6)
- wavelets (6)
- Elastizität (5)
- Elastoplastizität (5)
- Modellierungswoche (5)
- Numerische Mathematik (5)
- Stochastische dynamische Optimierung (5)
- haptotaxis (5)
- isogeometric analysis (5)
- nonparametric regression (5)
- portfolio optimization (5)
- time series (5)
- Combinatorial Optimization (4)
- Galerkin-Methode (4)
- Homogenisierung <Mathematik> (4)
- Hysterese (4)
- Kugel (4)
- Multicriteria optimization (4)
- NURBS (4)
- Optionspreistheorie (4)
- Portfolio-Optimierung (4)
- Simulation (4)
- Sphäre (4)
- hub location (4)
- linear algebra (4)
- mathematical education (4)
- multiscale model (4)
- network flows (4)
- neural network (4)
- numerics (4)
- praxis orientated (4)
- Analysis (3)
- Bewertung (3)
- Brownian motion (3)
- CHAMP <Satellitenmission> (3)
- Cauchy-Navier equation (3)
- Cauchy-Navier-Gleichung (3)
- Combinatorial optimization (3)
- Computeralgebra (3)
- Elastoplasticity (3)
- Erwarteter Nutzen (3)
- Finite-Volumen-Methode (3)
- Geodäsie (3)
- Geometric Ergodicity (3)
- Gravimetrie (3)
- Gröbner bases (3)
- Gröbner-Basis (3)
- Harmonische Spline-Funktion (3)
- Hysteresis (3)
- Intensity modulated radiation therapy (3)
- Kugelflächenfunktion (3)
- Lineare Optimierung (3)
- Monte-Carlo-Simulation (3)
- Mosco convergence (3)
- Multicriteria Optimization (3)
- Multiobjective optimization (3)
- Multiresolution Analysis (3)
- Numerische Strömungssimulation (3)
- Partial Differential Equations (3)
- Poisson-Gleichung (3)
- Portfolio Optimization (3)
- Portfoliomanagement (3)
- Randwertproblem / Schiefe Ableitung (3)
- Risikomanagement (3)
- Simplex (3)
- Sobolev-Raum (3)
- Spherical Wavelets (3)
- Spline (3)
- Spline-Approximation (3)
- Standortplanung (3)
- Stücklisten (3)
- Timetabling (3)
- Transaction Costs (3)
- Tropische Geometrie (3)
- Vektorwavelets (3)
- Wavelet-Analyse (3)
- autoregressive process (3)
- average density (3)
- combinatorial optimization (3)
- consecutive ones property (3)
- consistency (3)
- domain decomposition (3)
- facets (3)
- harmonic density (3)
- heuristic (3)
- kinetic equations (3)
- lattice Boltzmann method (3)
- low Mach number limit (3)
- optimales Investment (3)
- radiotherapy (3)
- tangent measure distributions (3)
- well-posedness (3)
- Algebraic Optimization (2)
- Algebraic dependence of commuting elements (2)
- Algebraische Abhängigkeit der kommutierende Elementen (2)
- Approximation Algorithms (2)
- Asymptotic Analysis (2)
- Asymptotic Expansion (2)
- Asymptotik (2)
- B-Spline (2)
- B-splines (2)
- Beam-on time (2)
- Biorthogonalisation (2)
- Biot-Savart Operator (2)
- Biot-Savart operator (2)
- CFD (2)
- CHAMP (2)
- Change analysis (2)
- Computer Algebra System (2)
- Computeralgebra System (2)
- Decomposition and Reconstruction Schemes (2)
- Decomposition cardinality (2)
- Delay Management (2)
- Derivat <Wertpapier> (2)
- Diffusionsprozess (2)
- Diskrete Fourier-Transformation (2)
- EM algorithm (2)
- Elastic BVP (2)
- Elasticity (2)
- Elastische Deformation (2)
- Elastisches RWP (2)
- Elastoplastisches RWP (2)
- Endliche Geometrie (2)
- Erdmagnetismus (2)
- FFT (2)
- Fatigue (2)
- Field splitting (2)
- Filtergesetz (2)
- Filtration (2)
- Finite Pointset Method (2)
- GOCE <Satellitenmission> (2)
- GRACE <Satellitenmission> (2)
- Geometrical Algorithms (2)
- Geothermal Flow (2)
- Graph Theory (2)
- Graphentheorie (2)
- Gruppentheorie (2)
- Hamilton-Jacobi-Differentialgleichung (2)
- Hochskalieren (2)
- Hypervolume (2)
- IMRT (2)
- Integer-valued time series (2)
- Inverse Problem (2)
- Isogeometrische Analyse (2)
- Jiang's model (2)
- Jiang-Modell (2)
- Konvergenz (2)
- Kreditrisiko (2)
- Langevin equation (2)
- Laplace transform (2)
- Lebensversicherung (2)
- Level-Set-Methode (2)
- Line Planning (2)
- Line Pool Generation (2)
- Lineare Elastizitätstheorie (2)
- Lineare partielle Differentialgleichung (2)
- Local smoothing (2)
- Logik (2)
- Lokalisation (2)
- Markov Chain (2)
- Mathematik (2)
- Mehrkriterielle Optimierung (2)
- Mehrskalenmodell (2)
- Mikrostruktur (2)
- Mixture Models (2)
- Modellbildung (2)
- Modulraum (2)
- Multileaf collimator sequencing (2)
- Multiobjective programming (2)
- Multiset Multicover (2)
- Multivariate Approximation (2)
- Neural networks (2)
- Numerisches Verfahren (2)
- Open Source Library (2)
- Optimal Control (2)
- Optimierung (2)
- Optimization in Public Transportation (2)
- POD (2)
- Palm distributions (2)
- Parallel volume (2)
- Particle Methods (2)
- Partielle Differentialgleichung (2)
- Poröser Stoff (2)
- Public Transport Planning (2)
- Rarefied Gas Dynamics (2)
- Ratenunabhängigkeit (2)
- Regressionsanalyse (2)
- Regularization (2)
- Robust Optimization (2)
- Scheduling (2)
- Schnitttheorie (2)
- Singularity theory (2)
- Sobolev spaces (2)
- Software for Public Transport Planning (2)
- Split Operator (2)
- Statistisches Modell (2)
- Stochastic Control (2)
- Stochastische Differentialgleichung (2)
- Stop Location (2)
- Subset selection (2)
- Theorie schwacher Lösungen (2)
- Time Series (2)
- Transaktionskosten (2)
- Up Functions (2)
- Upscaling (2)
- Value-at-Risk (2)
- Variationsungleichungen (2)
- Vehicle Scheduling (2)
- Vektorkugelfunktionen (2)
- Volatilität (2)
- Weißes Rauschen (2)
- White Noise Analysis (2)
- Wills functional (2)
- algebraic geometry (2)
- algorithmic game theory (2)
- approximate identity (2)
- asymptotic analysis (2)
- asymptotic behavior (2)
- average densities (2)
- cancer cell invasion (2)
- changepoint test (2)
- competitive analysis (2)
- connectedness (2)
- convergence (2)
- convex optimization (2)
- coset enumeration (2)
- curve singularity (2)
- degenerate diffusion (2)
- delay (2)
- density distribution (2)
- duality (2)
- dynamische Systeme (2)
- elastoplasticity (2)
- equilibrium strategies (2)
- evolutionary spectrum (2)
- finite volume method (2)
- geomagnetism (2)
- geometric ergodicity (2)
- global existence (2)
- harmonische Dichte (2)
- heat equation (2)
- hidden variables (2)
- homogenization (2)
- hub covering (2)
- hysteresis (2)
- illiquidity (2)
- image denoising (2)
- incompressible Navier-Stokes equations (2)
- interface problem (2)
- inverse optimization (2)
- inverse problem (2)
- inverse problems (2)
- k-link shortest path (2)
- lacunarity distribution (2)
- level set method (2)
- limit and jump relations (2)
- linear optimization (2)
- localizing basis (2)
- location theory (2)
- mathematical modeling (2)
- mesh generation (2)
- mixture (2)
- modal derivatives (2)
- moment realizability (2)
- monotropic programming (2)
- multicriteria optimization (2)
- multileaf collimator (2)
- multiplicative noise (2)
- nonlinear diffusion (2)
- nonlinear diffusion filtering (2)
- occupation measure (2)
- online optimization (2)
- optimal control (2)
- optimal investment (2)
- optimization (2)
- order-two densities (2)
- pH-taxis (2)
- parabolic system (2)
- particle method (2)
- particle methods (2)
- polynomial algorithms (2)
- poroelasticity (2)
- porous media (2)
- pyramid scheme (2)
- rate of convergence (2)
- regression analysis (2)
- regular surface (2)
- regularization (2)
- regularization wavelets (2)
- reproducing kernel (2)
- reproduzierender Kern (2)
- satellite gravity gradiometry (2)
- scale-space (2)
- series-parallel graphs (2)
- simplex (2)
- singularities (2)
- spherical approximation (2)
- splines (2)
- stationarity (2)
- stationary radiative transfer equation (2)
- subgroup problem (2)
- uniqueness (2)
- universal objective function (2)
- valid inequalities (2)
- variational inequalities (2)
- vector spherical harmonics (2)
- vectorial wavelets (2)
- weak solution (2)
- worst-case scenario (2)
- "Slender-Body"-Theorie (1)
- (Joint) chance constraints (1)
- (dynamic) network flows (1)
- 2-d kernel regression (1)
- 3D image analysis (1)
- A-infinity-bimodule (1)
- A-infinity-category (1)
- A-infinity-functor (1)
- ALE-Methode (1)
- AR-ARCH (1)
- Abel integral equations (1)
- Abelian groups (1)
- Abgeschlossenheit (1)
- Ableitung höherer Ordnung (1)
- Ableitungsfreie Optimierung (1)
- Abstract ODE (1)
- Abstract linear systems theory (1)
- Acid-mediated tumor invasion (1)
- Adjazenz-Beziehungen (1)
- Adjoint method (1)
- Adjoint system (1)
- Advanced Encryption Standard (1)
- Aggregation (1)
- Agriculture Loan (1)
- Algebraic Geometry (1)
- Algebraic geometry (1)
- Algebraic groups (1)
- Algebraic optimization (1)
- Algebraischer Funktionenkörper (1)
- Algorithmics (1)
- Alter (1)
- Analytic semigroup (1)
- Angewandte Mathematik (1)
- Anisotropic smoothness classes (1)
- Annulus (1)
- Anti-diffusion (1)
- Antidiffusion (1)
- Applications (1)
- Approximationsalgorithmus (1)
- Approximative Identität (1)
- Arbitrage (1)
- Arc distance (1)
- Archimedische Kopula (1)
- Arduino (1)
- Asiatische Option (1)
- Asset allocation (1)
- Asset-liability management (1)
- Associative Memory Problem (1)
- Asympotic Analysis (1)
- Asymptotische Entwicklung (1)
- Ausfallrisiko (1)
- Automatische Differentiation (1)
- Automatische Spracherkennung (1)
- Automorphismengruppe (1)
- Autoregression (1)
- Autoregressive Hilbertian model (1)
- Autoregressive time series (1)
- Balance sheet (1)
- Banach lattice (1)
- Barriers (1)
- Basic Scheme (1)
- Basis Risk (1)
- Basket Option (1)
- Baum <Mathematik> (1)
- Bayes risk (1)
- Bayes-Entscheidungstheorie (1)
- Bayesrisiko (1)
- Beam models (1)
- Beam orientation (1)
- Behinderter (1)
- Bell Number (1)
- Berechnungskomplexität (1)
- Bernstein Kern (1)
- Bernstein–Gelfand–Gelfand construction (1)
- Bernstejn-Polynom (1)
- Beschichtungsprozess (1)
- Beschränkte Krümmung (1)
- Bessel functions (1)
- Betrachtung des Schlimmstmöglichen Falles (1)
- Betriebsfestigkeit (1)
- Bilanzstrukturmanagement (1)
- Bildsegmentierung (1)
- Binary (1)
- Binomialbaum (1)
- Biot Poroelastizitätgleichung (1)
- Bisector (1)
- Black-Scholes model (1)
- Bondindizes (1)
- Bootstrap (1)
- Boundary Value Problem (1)
- Boundary Value Problem / Oblique Derivative (1)
- Boundary Value Problems (1)
- Box Algorithms (1)
- Box-Algorithm (1)
- Brinkman (1)
- Brownian Diffusion (1)
- Brownsche Bewegung (1)
- CAQ (1)
- CDO (1)
- CDS (1)
- CDSwaption (1)
- CFL type conditions (1)
- CHAMP-Mission (1)
- CPDO (1)
- CUSUM statistic (1)
- CWGAN (1)
- Cantor sets (1)
- Capacity (1)
- Capital-at-Risk (1)
- Carreau law (1)
- Castelnuovo Funktion (1)
- Castelnuovo function (1)
- Cauchy-Navier scaling function and wavelet (1)
- Cauchy-Navier-Equation (1)
- Censoring (1)
- Center Location (1)
- Change Point Analysis (1)
- Change Point Test (1)
- Change-point Analysis (1)
- Change-point estimator (1)
- Change-point test (1)
- Charakter <Gruppentheorie> (1)
- Chi-Quadrat-Test (1)
- Cholesky-Verfahren (1)
- Chorin's projection scheme (1)
- Chow Quotient (1)
- Circle Location (1)
- Classification (1)
- Cluster-Analyse (1)
- Coarse graining (1)
- Cohen-Lenstra heuristic (1)
- Collision Operator (1)
- Collocation Method plus (1)
- Commodity Index (1)
- Competitive Analysis (1)
- Complex Structures (1)
- Complexity (1)
- Complexity and performance of numerical algorithms (1)
- Composite Materials (1)
- Computer Algebra (1)
- Computer algebra (1)
- Conditional Value-at-Risk (1)
- Connectivity (1)
- Consistencyanalysis (1)
- Consistent Price Processes (1)
- Constraint Generation (1)
- Construction of hypersurfaces (1)
- Container (1)
- Continuum mechanics (1)
- Convergence Rate (1)
- Convex Analysis (1)
- Convex geometry (1)
- Convex sets (1)
- Convexity (1)
- Copula (1)
- Core (1)
- Cosine function (1)
- Coupled PDEs (1)
- Covid-19 (1)
- Coxeter groups (1)
- Coxeter-Freudenthal-Kuhn triangulation (1)
- Crane (1)
- Crash (1)
- Crash Hedging (1)
- Crash modelling (1)
- Crashmodellierung (1)
- Credit Default Swap (1)
- Credit Risk (1)
- Curvature (1)
- Curved viscous fibers (1)
- Cut (1)
- Cutting and Packing (1)
- Cycle Decomposition (1)
- DSMC (1)
- Darstellungstheorie (1)
- Das Urbild von Ideal unter einen Morphismus der Algebren (1)
- Debt Management (1)
- Decision Making (1)
- Decision support (1)
- Decomposition of integer matrices (1)
- Defaultable Options (1)
- Deformationstheorie (1)
- Degenerate Diffusion Semigroups (1)
- Delaunay (1)
- Delaunay triangulation (1)
- Delaunay triangulierung (1)
- Delay Differential Equations (1)
- Dense gas (1)
- Derivatives (1)
- Didaktik (1)
- Differential Cross-Sections (1)
- Differential forms (1)
- Differentialgleichung mit nacheilendem Argument (1)
- Differentialinklusionen (1)
- Differenzenverfahren (1)
- Differenzierbare Mannigfaltigkeit (1)
- Differenzmenge (1)
- Diffusion (1)
- Diffusion processes (1)
- Dirichlet series (1)
- Dirichlet-Problem (1)
- Discrete Bicriteria Optimization (1)
- Discrete decision problems (1)
- Discrete velocity models (1)
- Discriminatory power (1)
- Diskrete Mathematik (1)
- Dispersionsrelation (1)
- Dissertation (1)
- Diversifikation (1)
- Domain Decomposition (1)
- Doppelbarriereoption (1)
- Double Barrier Option (1)
- Druckkorrektur (1)
- Dynamic Network Flow Problem (1)
- Dynamic Network Flows (1)
- Dynamic capillary pressure (1)
- Dynamic cut (1)
- Dynamische Systeme (1)
- Dynamische Topographie (1)
- Dynasys (1)
- Dünnfilmapproximation (1)
- EDF observation models (1)
- EGM96 (1)
- EM algorith (1)
- Earliest arrival augmenting path (1)
- Earth' (1)
- Earth's disturbing potential (1)
- Education (1)
- Edwards Model (1)
- Effective Conductivity (1)
- Efficiency (1)
- Efficient Reliability Estimation (1)
- Effizienter Algorithmus (1)
- Effizienz (1)
- Eigenschwingung (1)
- Eikonal equation (1)
- Elastoplastic BVP (1)
- Electricity consumption (1)
- Elektromagnetische Streuung (1)
- Elektronik (1)
- Elementare Zahlentheorie (1)
- Eliminationsverfahren (1)
- Elliptic-parabolic equation (1)
- Elliptische Verteilung (1)
- Elliptisches Randwertproblem (1)
- Endliche Gruppe (1)
- Endliche Lie-Gruppe (1)
- Energy markets (1)
- Enskog equation (1)
- Entscheidungsbaum (1)
- Entscheidungsunterstützung (1)
- Enumerative Geometrie (1)
- Epidemiologie (1)
- Epidemiology (1)
- Erdöl Prospektierung (1)
- Ergodic (1)
- Erwartungswert-Varianz-Ansatz (1)
- Essential m-dissipativity (1)
- Euler's equation of motion (1)
- Evacuation Planning (1)
- Evakuierung (1)
- Evolution Equations (1)
- Evolutionary Integral Equations (1)
- Exogenous (1)
- Expected shortfall (1)
- Experimental Data (1)
- Exponential Utility (1)
- Exponentieller Nutzen (1)
- Extrapolation (1)
- Extreme Events (1)
- Extreme value theory (1)
- FEM (1)
- FEM-FCT stabilization (1)
- FPM (1)
- FPTAS (1)
- Faden (1)
- Faltung (1)
- Faltung <Mathematik> (1)
- Families of Probability Measures (1)
- Fast Pseudo Spectral Algorithm (1)
- Fast Wavelet Transform (1)
- Feed-forward Networks (1)
- Feedfoward Neural Networks (1)
- Feynman Integrals (1)
- Feynman path integrals (1)
- Fiber spinning (1)
- Fiber suspension flow (1)
- Filippov theory (1)
- Filippov-Theorie (1)
- Financial Engineering (1)
- Finanzkrise (1)
- Finanznumerik (1)
- Finanzzeitreihe (1)
- Finite-Elemente-Methode (1)
- Finite-Punktmengen-Methode (1)
- Firmwertmodell (1)
- First Order Optimality System (1)
- First--order optimality system (1)
- Flachwasser (1)
- Flachwassergleichungen (1)
- FlowLoc (1)
- Fluid dynamics (1)
- Fluid-Feststoff-Strömung (1)
- Fluid-Struktur-Kopplung (1)
- Fluid-Struktur-Wechselwirkung (1)
- Foam decay (1)
- Fokker-Planck equation (1)
- Fokker-Planck-Gleichung (1)
- Forbidden Regions (1)
- Forecasting (1)
- Forward-Backward Stochastic Differential Equation (1)
- Fourier-Transformation (1)
- Fredholm integral equation of the second kind (1)
- Fredholmsche Integralgleichung (1)
- Frequency Averaging (1)
- Function of bounded variation (1)
- Functional autoregression (1)
- Functional time series (1)
- Funktionalanalysis (1)
- Funktionenkörper (1)
- Fuzzy Programming (1)
- GARCH (1)
- GARCH Modelle (1)
- GOCE <satellite mission> (1)
- GPS-satellite-to-satellite tracking (1)
- GRACE (1)
- GRACE <satellite mission> (1)
- Galerkin Approximation (1)
- Gamma-Konvergenz (1)
- Garantiezins (1)
- Garbentheorie (1)
- Gauge Distances (1)
- Gauss-Manin connection (1)
- Gaussian random noise (1)
- Gebietszerlegung (1)
- Gebietszerlegungsmethode (1)
- Gebogener viskoser Faden (1)
- Generative adversarial networks (1)
- Geo-referenced data (1)
- Geodesie (1)
- Geodätischer Satellit (1)
- Geomagnetic Field Modelling (1)
- Geomagnetismus (1)
- Geomathematik (1)
- Geometrical algorithms (1)
- Geometrische Ergodizität (1)
- Geostrophic flow (1)
- Geostrophisches Gleichgewicht (1)
- Geothermal Systems (1)
- Geothermischer Fluss (1)
- Gewichteter Sobolev-Raum (1)
- Gewichtung (1)
- Gittererzeugung (1)
- Gleichgewichtsstrategien (1)
- Gleichmäßige Approximation (1)
- Global Optimization (1)
- Global optimization (1)
- Globale nichtlineare Analysis (1)
- Glättung (1)
- Glättungsparameterwahl (1)
- Grad expansion (1)
- Gradient based optimization (1)
- Granular flow (1)
- Granulat (1)
- Graph coloring (1)
- Gravimetry (1)
- Gravitation (1)
- Gravitational Field (1)
- Gravitationsmodell (1)
- Greedy Heuristic (1)
- Greedy algorithm (1)
- Green’s function (1)
- Gromov Witten (1)
- Gromov-Witten-Invariante (1)
- Große Abweichung (1)
- Gruppenoperation (1)
- Gröbner base (1)
- Gröbner bases in monoid and group rings (1)
- Gröbner-basis (1)
- Gyroscopic (1)
- Hadamard manifold (1)
- Hadamard space (1)
- Hadamard-Mannigfaltigkeit (1)
- Hadamard-Raum (1)
- Hamiltonian (1)
- Hamiltonian Path Integrals (1)
- Hamiltonian groups (1)
- Handelsstrategien (1)
- Hardy space (1)
- Harmonische Analyse (1)
- Harmonische Dichte (1)
- Harmonische Funktion (1)
- Hazard Functions (1)
- Heavy-tailed Verteilung (1)
- Hedging (1)
- Helmholtz Type Boundary Value Problems (1)
- Helmholtz decomposition (1)
- Helmholtz-Decomposition (1)
- Helmholtz-Zerlegung (1)
- Heston-Modell (1)
- Heuristic (1)
- Heuristik (1)
- Hidden Markov models for Financial Time Series (1)
- Hierarchies (1)
- Hierarchische Matrix (1)
- Higher Order Differentials as Boundary Data (1)
- Hilbert complexes (1)
- Hochschild homology (1)
- Hochschild-Homologie (1)
- Homogeneous Relaxation (1)
- Homogenization (1)
- Homologietheorie (1)
- Homologische Algebra (1)
- Homotopie (1)
- Homotopiehochhebungen (1)
- Homotopy (1)
- Homotopy lifting (1)
- Hub Location Problem (1)
- Hub-and-Spoke-System (1)
- Hybrid Codes (1)
- Hydrological Gravity Variations (1)
- Hydrologie (1)
- Hydrostatischer Druck (1)
- Hyperbolic Conservation (1)
- Hyperelastizität (1)
- Hyperelliptische Kurve (1)
- Hyperflächensingularität (1)
- Hyperspektraler Sensor (1)
- Hypocoercivity (1)
- INGARCH (1)
- ITSM (1)
- Idealklassengruppe (1)
- Identifiability (1)
- Ill-Posed Problems (1)
- Ill-posed Problems (1)
- Ill-posed problem (1)
- Illiquidität (1)
- Image restoration (1)
- Immiscible lattice BGK (1)
- Immobilienaktie (1)
- Improperly posed problems (1)
- Impulse control (1)
- Incompressible Navier-Stokes (1)
- Index Insurance (1)
- Industrial Applications (1)
- Infectious Diseases (1)
- Inflation (1)
- Information Theory (1)
- Infrarotspektroskopie (1)
- Injectivity of mappings (1)
- Injektivität von Abbildungen (1)
- Inkompressibel Navier-Stokes (1)
- Inkorrekt gestelltes Problem (1)
- Insurance (1)
- Integral (1)
- Integral Equations (1)
- Integral transform (1)
- Integration (1)
- Intensität (1)
- Interdisziplinärer Projektunterricht (1)
- Internationale Diversifikation (1)
- Interpolation Algorithm (1)
- Inverse Problems (1)
- Inverse problems in Banach spaces (1)
- Irreduzibler Charakter (1)
- Isogeometric Analysis (1)
- Isotropy (1)
- Iterative Methods (1)
- Ito (1)
- Jacobigruppe (1)
- Jeffreys' prior (1)
- Jiang's constitutive model (1)
- Jiangsches konstitutives Gesetz (1)
- K-best solution (1)
- K-cardinality trees (1)
- Kaktusgraph (1)
- Kalkül (1)
- Kalkül des natürlichen Schließens (1)
- Kallianpur-Robbins law (1)
- Kanalcodierung (1)
- Karhunen-Loève expansion (1)
- Kategorientheorie (1)
- Kelvin Transformation (1)
- Kernschätzer (1)
- Kinetic Schems (1)
- Kinetic Theory of Gases (1)
- Kinetic theory (1)
- Kirchhoff-Love shell (1)
- Kiyoshi (1)
- Knapsack (1)
- Knapsack problem (1)
- Kohonen's SOM (1)
- Kombinatorik (1)
- Kombinatorische Optimierung (1)
- Kommutative Algebra (1)
- Kompakter Träger <Mathematik> (1)
- Konjugierte Dualität (1)
- Konstruktion von Hyperflächen (1)
- Konstruktive Approximation (1)
- Kontinuum <Mathematik> (1)
- Kontinuumsmechanik (1)
- Kontinuumsphysik (1)
- Konvergenzrate (1)
- Konvergenzverhalten (1)
- Konvexe Mengen (1)
- Konvexe Optimierung (1)
- Kopplungsmethoden (1)
- Kopplungsproblem (1)
- Kopula <Mathematik> (1)
- Kreitderivaten (1)
- Kristallmathematik (1)
- Kryptoanalyse (1)
- Kryptologie (1)
- Krümmung (1)
- Kugelfunktion (1)
- Kullback Leibler distance (1)
- Kullback-Leibler divergence (1)
- Kurvenschar (1)
- L-curve Methode (1)
- L2-Approximation (1)
- LIBOR (1)
- Label correcting algorithm (1)
- Label setting algorithm (1)
- Lagrange (1)
- Lagrangian Functions (1)
- Lagrangian relaxation (1)
- Large-Scale Problems (1)
- Lattice Boltzmann (1)
- Lattice-BGK (1)
- Lattice-Boltzmann (1)
- Lavrentiev regularization (1)
- Lavrentiev regularization for equations with monotone operators (1)
- Leading-Order Optimality (1)
- Learnability (1)
- Learning systems (1)
- Least-squares Monte Carlo method (1)
- Lebesque-Integral (1)
- Legendre Wavelets (1)
- Lehrkräfte (1)
- Lehrmittel (1)
- Level set methods (1)
- Level sets (1)
- Levy process (1)
- Lexicographic Order (1)
- Lexicographic max-ordering (1)
- Lie algebras (1)
- Lie-Typ-Gruppe (1)
- Linear Integral Equations (1)
- Linear kinematic hardening (1)
- Linear kinematische Verfestigung (1)
- Linear membership function (1)
- Lineare Integralgleichung (1)
- Lippmann-Schwinger Equation (1)
- Lippmann-Schwinger equation (1)
- Liquidität (1)
- Local completeness (1)
- Local existence uniqueness (1)
- Locally Supported Radial Basis Functions (1)
- Locally Supported Zonal Kernels (1)
- Locally stationary processes (1)
- Location (1)
- Location problems (1)
- Location theory (1)
- Locational Planning (1)
- Lokalkompakte Kerne (1)
- Low-discrepancy sequences (1)
- Lucena (1)
- MBS (1)
- MKS (1)
- ML-estimation (1)
- MLC (1)
- MLE (1)
- MOCO (1)
- Macaulay’s inverse system (1)
- Machine Scheduling (1)
- Magneto-Elastic Coupling (1)
- Magnetoelastic coupling (1)
- Magnetoelasticity (1)
- Magnetostriction (1)
- Marangoni-Effekt (1)
- Market Equilibrium (1)
- Markov Kette (1)
- Markov process (1)
- Markov switching (1)
- Markov-Ketten-Monte-Carlo-Verfahren (1)
- Markov-Prozess (1)
- Marktmanipulation (1)
- Marktrisiko (1)
- Martingaloptimalitätsprinzip (1)
- Maschinelles Lernen (1)
- Massendichte (1)
- Math-Talent-School (1)
- Mathematical Epidemiology (1)
- Mathematical Finance (1)
- Mathematical Modeling (1)
- Mathematics (1)
- Mathematisches Modell (1)
- Matrixkompression (1)
- Matrizenfaktorisierung (1)
- Matrizenzerlegung (1)
- Matroids (1)
- Max-Ordering (1)
- Maximal Cohen-Macaulay modules (1)
- Maximale Cohen-Macaulay Moduln (1)
- Maximum Likelihood Estimation (1)
- Maximum-Likelihood-Schätzung (1)
- Maxwell's equations (1)
- McKay conjecture (1)
- McKay-Conjecture (1)
- McKay-Vermutung (1)
- Medical Physics (1)
- Mehrdimensionale Bildverarbeitung (1)
- Mehrdimensionale Spline-Funktion (1)
- Mehrdimensionales Variationsproblem (1)
- Mehrskalen (1)
- Methode der Fundamentallösungen (1)
- Microstructure (1)
- Mie representation (1)
- Mie- and Helmholtz-Representation (1)
- Mie- und Helmholtz-Darstellung (1)
- Mie-Darstellung (1)
- Mie-Representation (1)
- Mikroelektronik (1)
- Minimal spannender Baum (1)
- Minimum Cost Network Flow Problem (1)
- Minimum Principle (1)
- Minkowski space (1)
- Mixed Connectivity (1)
- Mixed integer programming (1)
- Mixed method (1)
- Model-Dynamics (1)
- Moduli Spaces (1)
- Molekulardynamik (1)
- Molodensky Problem (1)
- Molodensky problem (1)
- Moment sequence (1)
- Momentum and Mas Transfer (1)
- Monoid and group rings (1)
- Monotone dynamical systems (1)
- Monte Carlo (1)
- Monte Carlo method (1)
- Moreau-Yosida regularization (1)
- Morphismus (1)
- Mosaike (1)
- Motion Capturing (1)
- Multi Primary and One Second Particle Method (1)
- Multi-Asset Option (1)
- Multi-Variant Model (1)
- Multi-dimensional systems (1)
- Multicriteria Location (1)
- Multileaf Collimator (1)
- Multileaf collimator (1)
- Multiperiod planning (1)
- Multiphase Flows (1)
- Multiple Criteria (1)
- Multiple Objective Programs (1)
- Multiple criteria analysis (1)
- Multiple criteria optimization (1)
- Multiple objective combinatorial optimization (1)
- Multiple objective optimization (1)
- Multiplicative Schwarz Algorithm (1)
- Multiresolution analysis (1)
- Multiscale Methods (1)
- Multiscale model (1)
- Multiscale modelling (1)
- Multiskalen-Entrauschen (1)
- Multiskalenapproximation (1)
- Multispektralaufnahme (1)
- Multispektralfotografie (1)
- Multisresolution Analysis (1)
- Multivariate (1)
- Multivariate Analyse (1)
- Multivariate Wahrscheinlichkeitsverteilung (1)
- Multivariates Verfahren (1)
- NP (1)
- NP-completeness (1)
- Nash equilibria (1)
- Navier Stokes equation (1)
- Navier-Stokes (1)
- Network flows (1)
- Networks (1)
- Netzwerksynthese (1)
- Neumann Wavelets (1)
- Neumann wavelets (1)
- Neumann-Problem (1)
- Neural Networks (1)
- Neuronales Netz (1)
- Newtonsches Potenzial (1)
- Nicht-Desarguessche Ebene (1)
- Nichtglatte Optimierung (1)
- Nichtkommutative Algebra (1)
- Nichtkonvexe Optimierung (1)
- Nichtkonvexes Variationsproblem (1)
- Nichtlineare Approximation (1)
- Nichtlineare Diffusion (1)
- Nichtlineare Optimierung (1)
- Nichtlineare Zeitreihenanalyse (1)
- Nichtlineare partielle Differentialgleichung (1)
- Nichtlineare/große Verformungen (1)
- Nichtlineares Galerkinverfahren (1)
- Nichtparametrische Regression (1)
- Nichtpositive Krümmung (1)
- Niederschlag (1)
- Nilpotent elements (1)
- No-Arbitrage (1)
- Non-commutative Computer Algebra (1)
- Non-convex body (1)
- Non-linear wavelet thresholding (1)
- Nonlinear Galerkin Method (1)
- Nonlinear Optimization (1)
- Nonlinear dynamics (1)
- Nonlinear regression (1)
- Nonlinear time series analysis (1)
- Nonlinear/large deformations (1)
- Nonparametric AR-ARCH (1)
- Nonparametric regression (1)
- Nonparametric time series (1)
- Nonsmooth contact dynamics (1)
- Nonstationary processes (1)
- Nulldimensionale Schemata (1)
- Numerical Analysis (1)
- Numerical Flow Simulation (1)
- Numerical methods (1)
- Numerics (1)
- Numerische Mathematik / Algorithmus (1)
- Oberflächenmaße (1)
- Oberflächenspannung (1)
- On-line algorithm (1)
- One-dimensional systems (1)
- Online Algorithms (1)
- Optimal Prior Distribution (1)
- Optimal control (1)
- Optimal portfolios (1)
- Optimal semiconductor design (1)
- Optimale Kontrolle (1)
- Optimale Portfolios (1)
- Optimization Algorithms (1)
- Option (1)
- Option Valuation (1)
- Optionsbewertung (1)
- Order (1)
- Orthonormalbasis (1)
- Ovoid (1)
- PDE-Constrained Optimization, Robust Design, Multi-Objective Optimization (1)
- Palm distribution (1)
- Panel clustering (1)
- Papiermaschine (1)
- Parallel Algorithms (1)
- Paralleler Algorithmus (1)
- Parameter identification (1)
- Parameteridentifikation (1)
- Pareto Optimality (1)
- Pareto Points (1)
- Pareto optimality (1)
- Parkette (1)
- Partikel Methoden (1)
- Patchworking Methode (1)
- Patchworking method (1)
- Pathwise Optimality (1)
- Pedestrian FLow (1)
- Perceptron (1)
- Periodic Homogenization (1)
- Perona-Malik filter (1)
- Pfadintegral (1)
- Planares Polynom (1)
- Poisson autoregression (1)
- Poisson noise (1)
- Poisson regression (1)
- PolyBoRi (1)
- Polyhedron (1)
- Polynomapproximation (1)
- Polynomial Eigenfunctions (1)
- Pontrjagin (1)
- Population Balance Equation (1)
- Poroelastizität (1)
- Porous flow (1)
- Portfolio Optimierung (1)
- Portfoliooptimierung (1)
- Potential transform (1)
- Preimage of an ideal under a morphism of algebras (1)
- Probust optimization (1)
- Project prioritization (1)
- Project selection (1)
- Projektarbeit (1)
- Projektionsoperator (1)
- Projektive Fläche (1)
- Projektunterricht (1)
- Prox-Regularisierung (1)
- Pseudopolynomial-Time Algorithm (1)
- Punktmengen (1)
- Punktprozess (1)
- QMC (1)
- QVIs (1)
- Quadratischer Raum (1)
- Quantile autoregression (1)
- Quantization (1)
- Quasi-Variational Inequalities (1)
- Quasi-identities (1)
- RCGAN (1)
- RCWGAN (1)
- RKHS (1)
- Radial Basis Functions (1)
- Radiation Therapy (1)
- Radiative Heat Trasfer (1)
- Radiative heat transfer (1)
- Radiotherapy (1)
- Rainflow (1)
- Random Errors (1)
- Random body (1)
- Random differential equations (1)
- Random number generation (1)
- Randwertproblem (1)
- Rank test (1)
- Rarefied Gas Flows (1)
- Rarefied Gsa Dynamics (1)
- Rarefied Polyatomic Gases (1)
- Rarefied gas (1)
- Raspberry Pi (1)
- Rate-independency (1)
- Ray-Knight Theorem (1)
- Rayleigh Number (1)
- Reaction-diffusion equations (1)
- Rectifiability (1)
- Recurrent Networks (1)
- Recurrent neural networks (1)
- Reflection (1)
- Reflexionsspektroskopie (1)
- Regime Shifts (1)
- Regime-Shift Modell (1)
- Regularisierung / Stoppkriterium (1)
- Regularization / Stop criterion (1)
- Regularization Wavelets (1)
- Regularization methods (1)
- Rehabilitation clinics (1)
- Relaxation (1)
- Reliability (1)
- Representation (1)
- Resolvent Estimate (1)
- Resonant tunneling diode (1)
- Restricted Regions (1)
- Restricted Shortest Path (1)
- Richtungsableitung (1)
- Riemann-Siegel formula (1)
- Riemannian manifolds (1)
- Riemannsche Mannigfaltigkeiten (1)
- Riemannsche Summen (1)
- Riesz Transform (1)
- Rigid Body Motion (1)
- Risikoanalyse (1)
- Risikomaße (1)
- Risikotheorie (1)
- Risk Management (1)
- Risk Measures (1)
- Risk Sharing (1)
- Robust smoothing (1)
- Rohstoffhandel (1)
- Rohstoffindex (1)
- Räumliche Statistik (1)
- SARS-CoV-2 (1)
- SAW filters (1)
- SGG (1)
- SPn-approximation (1)
- SST (1)
- SWARM (1)
- Saddle Points (1)
- Sandwiching algorithm (1)
- Satellitendaten (1)
- Satellitengeodäsie (1)
- Satellitengradiogravimetrie (1)
- Satellitengradiometrie (1)
- Scalar type operator (1)
- Scalar-type operator (1)
- Scale function (1)
- Scattered-Data-Interpolation (1)
- Schaum (1)
- Schaumzerfall (1)
- Schiefe Ableitung (1)
- Schnelle Fourier-Transformation (1)
- Schnitt <Mathematik> (1)
- Schwache Formulierung (1)
- Schwache Konvergenz (1)
- Schwache Lösu (1)
- Second Order Conditions (1)
- Seismic Modeling (1)
- Seismische Tomographie (1)
- Seismische Welle (1)
- Semantik (1)
- Semi-Markov-Kette (1)
- Semi-infinite optimization (1)
- Semigroups (1)
- Sensitivitäten (1)
- Sequential test (1)
- Sequenzieller Algorithmus (1)
- Serre functor (1)
- Shallow Water Equations (1)
- Shannon capacity (1)
- Shannon optimal priors (1)
- Shannon-Capacity (1)
- Shape optimization (1)
- Shapley value (1)
- Shapleywert (1)
- Shearlets (1)
- Sheaves (1)
- Shock Wave Problem (1)
- Shortest path problem (1)
- Signalanalyse (1)
- Similarity measures (1)
- Singular <Programm> (1)
- Singularität (1)
- Singularitätentheorie (1)
- Skalierungsfunktion (1)
- Slender body theory (1)
- Slender-Body Approximations (1)
- Smoothed Particle Hydrodynamics (1)
- Sobolevräume (1)
- Solvency II (1)
- Solvency-II-Richtlinie (1)
- Spannungs-Dehn (1)
- Spatial Statistics (1)
- Spectral Analysis (1)
- Spectral Method (1)
- Spectral theory (1)
- Spektralanalyse <Stochastik> (1)
- Spherical (1)
- Spherical Fast Wavelet Transform (1)
- Spherical Harmonics (1)
- Spherical Location Problem (1)
- Spherical Multiresolution Analysis (1)
- Sphärische Approximation (1)
- Sphärische Wavelets (1)
- Spieltheorie (1)
- Spline-Interpolation (1)
- Spline-Wavelets (1)
- Splines (1)
- Split-Operator (1)
- Splitoperator (1)
- Sprung-Diffusions-Prozesse (1)
- Square-mean Convergence (1)
- Stabile Vektorbundle (1)
- Stable vector bundles (1)
- Standard basis (1)
- Standortprobleme (1)
- Standorttheorie (1)
- Statistical Experiments (1)
- Statistics (1)
- Steuer (1)
- Stieltjes transform (1)
- Stochastic Impulse Control (1)
- Stochastic Processes (1)
- Stochastic Volatility (1)
- Stochastic optimization (1)
- Stochastische Inhomogenitäten (1)
- Stochastische Processe (1)
- Stochastische Volatilität (1)
- Stochastische Zinsen (1)
- Stochastische optimale Kontrolle (1)
- Stochastischer Prozess (1)
- Stochastisches Feld (1)
- Stochastisches Modell (1)
- Stokes Flow (1)
- Stokes Wavelets (1)
- Stokes wavelets (1)
- Stokes-Gleichung (1)
- Stop- and Play-Operators (1)
- Stop- und Play-Operator (1)
- Stop- und Spieloperator (1)
- Stop-und Play-Operator (1)
- Stornierung (1)
- Stoßdämpfer (1)
- Strahlentherapie (1)
- Strahlungstransport (1)
- Strain-Life Approach (1)
- Stratifaltigkeiten (1)
- Structural Reliability (1)
- Structure Theory (1)
- Strukturiertes Finanzprodukt (1)
- Strukturoptimierung (1)
- Strömungsdynamik (1)
- Strömungsmechanik (1)
- Subset Simulationen (1)
- Success Run (1)
- Survival Analysis (1)
- Synthesizer (1)
- Synthetic data generation (1)
- Systemidentifikation (1)
- Sägezahneffekt (1)
- TDTSP (1)
- TSP (1)
- Tabellenkalkulation (1)
- Tableau-Kalkül (1)
- Tail Dependence Koeffizient (1)
- Temporal Variational Autoencoders (1)
- Tensor Spherical Harmonics (1)
- Tensorfeld (1)
- Test for Changepoint (1)
- Theorem of Plemelj-Privalov (1)
- Thermophoresis (1)
- Thin film approximation (1)
- Tichonov-Regularisierung (1)
- Tiefengeothermie (1)
- Time-Series (1)
- Time-Space Multiresolution Analysis (1)
- Time-delay-Netz (1)
- TimeGAN (1)
- Titration (1)
- Topologieoptimierung (1)
- Topology optimization (1)
- Traffic flow (1)
- Train Rearrangement (1)
- Transaction costs (1)
- Translation planes (1)
- Transportation Problem (1)
- Tree (1)
- Trennschärfe <Statistik> (1)
- Trennverfahren (1)
- Treppenfunktionen (1)
- Triangular fuzzy number (1)
- Tropical Grassmannian (1)
- Tropical Intersection Theory (1)
- Tube Drawing (1)
- Two-Scale Convergence (1)
- Two-phase flow (1)
- Uniform matroids (1)
- Universal objective function (1)
- Unreinheitsfunktion (1)
- Unschärferelation (1)
- Unsupervised learning (1)
- Untermannigfaltigkeit (1)
- Upwind-Verfahren (1)
- Usage modeling (1)
- Utility (1)
- Value at Risk (1)
- Value at risk (1)
- Van Neumann-Kakutani transformation (1)
- Variational autoencoders (1)
- Variational inequalities (1)
- Variationsrechnung (1)
- Variationsungleichugen (1)
- Vector Spherical Harmonics (1)
- Vector-valued holomorphic function (1)
- Vectorfield approximation (1)
- Vectorial Wavelets (1)
- Vektor-Wavelets (1)
- Vektorfeld (1)
- Vektorfeldapproximation (1)
- Verkehsplanung (1)
- Verschlüsselung (1)
- Verschwindungsatz (1)
- Versicherung (1)
- Vetor optimization (1)
- Vigenere (1)
- Virus Variants (1)
- Viskoelastische Flüssigkeiten (1)
- Viskose Transportschemata (1)
- Volatilitätsarbitrage (1)
- Vollständigkeit (1)
- Vorkonditionierer (1)
- Vorlesungsskript (1)
- Voronoi diagram (1)
- Vorwärts-Rückwärts-Stochastische-Differentialgleichung (1)
- Water reservoir management (1)
- Wave Based Method (1)
- Wavelet Analysis auf regulären Flächen (1)
- Wavelet-Theorie (1)
- Wavelet-Theory (1)
- Wavelet-Transformation (1)
- Wavelets (1)
- Wavelets auf der Kugel und der Sphäre (1)
- Weak Solution Theory (1)
- Wellengeschwindigkeit (1)
- White Noise (1)
- Wirbelabtrennung (1)
- Wirbelströmung (1)
- Wirkungsnetz (1)
- Wissenschaftliches Rechnen (1)
- Word problem (1)
- Worst-Case (1)
- Wärmeleitfähigkeit (1)
- Yaglom limits (1)
- Zeitabhängigkeit (1)
- Zeitintegrale Modelle (1)
- Zeitliche Veränderungen (1)
- Zeitreihe (1)
- Zeitreihen (1)
- Zellulärer Automat (1)
- Zentrenprobleme (1)
- Zerlegungen (1)
- Zero-dimensional schemes (1)
- Zonal Kernel Functions (1)
- Zopfgruppe (1)
- Zufälliges Feld (1)
- Zweiphasenströmung (1)
- Zyklische Homologie (1)
- abgeleitete Kategorie (1)
- acid-mediated tumor invasion (1)
- activity-based model (1)
- adaptive algorithm (1)
- adaptive estimation (1)
- adaptive grid generation (1)
- additive Gaussian noise (1)
- adjacency (1)
- adjacency relations (1)
- adjoint approach (1)
- adjoints (1)
- aggressive space mapping (1)
- aleph (1)
- algebraic attack (1)
- algebraic correspondence (1)
- algebraic function fields (1)
- algebraic number fields (1)
- algebraic topology (1)
- algebraische Korrespondenzen (1)
- algebraische Topologie (1)
- algebroid curve (1)
- algorithm (1)
- alternating minimization (1)
- alternating optimization (1)
- analoge Mikroelektronik (1)
- angewandte Mathematik (1)
- angewandte Topologie (1)
- anisotropen Viskositätsmodell (1)
- anisotropic diffusion (1)
- anisotropic viscosity (1)
- anisotropy (1)
- applied mathematics (1)
- approximation methods (1)
- approximative Identität (1)
- arbitrary Lagrangian-Eulerian methods (ALE) (1)
- arbitrary function (1)
- archimedean copula (1)
- area loss (1)
- asian option (1)
- associated Legendre functions (1)
- asymptotic expansions (1)
- asymptotic preserving numerical scheme (1)
- asymptotic-preserving (1)
- auto-pruning (1)
- automatic differentiation (1)
- ball (1)
- basic systems theoretic properties (1)
- basket option (1)
- benders decomposition (1)
- bending strip method (1)
- best basis (1)
- bicriteria shortest path problem (1)
- bicriterion path problems (1)
- bills of material (1)
- bills of materials (1)
- bin coloring (1)
- binomial tree (1)
- biorthogonal bases of L^2 (1)
- bipolar quantum drift diffusion model (1)
- blackout period (1)
- bocses (1)
- body wave velocity (1)
- bootstrap (1)
- bottleneck (1)
- boundary conditions (1)
- boundary value problem (1)
- boundary-value problems of potent (1)
- branch and cut (1)
- branching process (1)
- bus bunching (1)
- cactus graph (1)
- cancer radiation therapy (1)
- canonical ideal (1)
- canonical module (1)
- cardinality constraint combinatorial optimization (1)
- cash management (1)
- center hyperplane (1)
- centrally symmetric polytope (1)
- change analysis (1)
- change point (1)
- changing market coefficients (1)
- characteristic polynomial (1)
- charged fluids (1)
- chemotaxis (1)
- chemotherapy (1)
- classical solutions (1)
- clo (1)
- closure approximation (1)
- clustering (1)
- clustering methods (1)
- combinatorics (1)
- common transversal (1)
- compact operator equation (1)
- complete presentations (1)
- complexity (1)
- composites (1)
- computational complexity (1)
- computational finance (1)
- computer algebra (1)
- computeralgebra (1)
- conditional quantile (1)
- conditional quantiles (1)
- confluence (1)
- consecutive ones matrix (1)
- consecutive ones polytopes (1)
- constructive approximation (1)
- control theory (1)
- convergence behaviour (1)
- convex constraints (1)
- convex distance funtion (1)
- convex operator (1)
- cooling processes (1)
- cooperative game (1)
- core (1)
- correlated errors (1)
- count data (1)
- coupling methods (1)
- coverage error (1)
- crash (1)
- crash hedging (1)
- crash modelling (1)
- credit risk (1)
- curvature (1)
- cusp forms (1)
- cut (1)
- cut basis problem (1)
- cuts (1)
- cyclic homology (1)
- da (1)
- data structure (1)
- data-adaptive bandwidth choice (1)
- decision support (1)
- decision support systems (1)
- decisions (1)
- decoding (1)
- decrease direction (1)
- default time (1)
- deficiency (1)
- deflections of the vertical (1)
- degenerations of an elliptic curve (1)
- delay management (1)
- delay management problem (1)
- denoising (1)
- dense univariate rational interpolation (1)
- density gradient equation (1)
- derivative-free iterative method (1)
- derived category (1)
- descent algorithm (1)
- determinant (1)
- differential inclusions (1)
- diffusion models (1)
- diffusive scaling (1)
- direct product (1)
- directional derivative (1)
- discrepancy (1)
- discrete element method (1)
- discrete measure (1)
- discrete time setting (1)
- discrete velocity models (1)
- discretization (1)
- diskrete Systeme (1)
- displacement problem (1)
- distribution (1)
- diversification (1)
- domain decomposition methods (1)
- domain parametrization (1)
- double exponential distribution (1)
- downward continuation (1)
- drift diffusion (1)
- drift-diffusion limit (1)
- durability (1)
- dynamic holding (1)
- dynamic network flows (1)
- dynamical topography (1)
- earliest arrival flow (1)
- earliest arrival flows (1)
- efficiency loss (1)
- efficient solution (1)
- eigenvalue problems (1)
- elasticity problem (1)
- elastoplastic BVP (1)
- elliptical distribution (1)
- endomorphism ring (1)
- energy transport (1)
- entropy (1)
- enumerative geometry (1)
- epsilon-constraint method (1)
- equilibrium state (1)
- equisingular families (1)
- estimate (1)
- estimation (1)
- estimator (1)
- exact fully discrete vectorial wavelet transform (1)
- exact solution (1)
- exchange rate (1)
- explicit representation (1)
- explicit representations (1)
- explizite Darstellung (1)
- exponential rate (1)
- extreme equilibria (1)
- f-dissimilarity (1)
- face value (1)
- facility location (1)
- fast approximation (1)
- fatigue (1)
- fiber reinforced silicon carbide (1)
- fibre lay-down dynamics (1)
- film casting (1)
- filtration (1)
- final prediction error (1)
- financial mathematics (1)
- finite biodiversity (1)
- finite difference method (1)
- finite difference schemes (1)
- finite element method (1)
- finite groups of Lie type (1)
- finite pointset method (1)
- finite spin group (1)
- finite volume methods (1)
- finite-difference methods (1)
- first hitting time (1)
- fixpoint theorem (1)
- float glass (1)
- flood risk (1)
- fluid dynamic equations (1)
- fluid structure (1)
- fluid structure interaction (1)
- fluid-structure interaction (FSI) (1)
- formale Logik (1)
- formants (1)
- formulation as integral equation (1)
- forward-shooting grid (1)
- fptas (1)
- fractals (1)
- free boundary (1)
- free surface (1)
- freie Oberfläche (1)
- frequency bands (1)
- freqzency bands (1)
- function of bounded variation (1)
- functional data (1)
- functional time series (1)
- fundamental cut (1)
- fundamental systems (1)
- gas dynamics (1)
- gauge (1)
- gebietszerlegung (1)
- general multidimensional moment problem (1)
- generalized Gummel itera (1)
- generalized inverse Gaussian diffusion (1)
- generic character table (1)
- geodetic (1)
- geomagnetic field modelling from MAGSAT data (1)
- geomathematics (1)
- geometric measure theory (1)
- geometrical algorithms (1)
- geometry of measures (1)
- geopotential determination (1)
- gitter (1)
- glioblastoma (1)
- global optimization (1)
- go-or-grow (1)
- go-or-grow dichotomy (1)
- good semigroup (1)
- gradient descent reprojection (1)
- granular flow (1)
- graph and network algorithm (1)
- graph p-Laplacian (1)
- gravimetry (1)
- gravitation (1)
- gravitational field recovery (1)
- grenzwert (1)
- group action (1)
- groups of Lie type (1)
- growing sub-quadratically (1)
- growth optimal portfolios (1)
- großer Investor (1)
- harmonic WFT (1)
- harmonic balance (1)
- harmonic scaling functions and wavelets (1)
- harmonic wavelets (1)
- headway prediction (1)
- heat radiation (1)
- hedging (1)
- hidden Markov (1)
- hierarchical matrix (1)
- higher order (1)
- higher-order moments (1)
- homological algebra (1)
- hybrid method (1)
- hyper-quasi-identities (1)
- hyperbolic conservation laws (1)
- hyperbolic systems (1)
- hyperbolic systems of conservation laws (1)
- hyperelliptic function field (1)
- hyperelliptische Funktionenkörper (1)
- hypergeometric functions (1)
- hyperplane transversal (1)
- hyperquasivarieties (1)
- hyperspectal unmixing (1)
- hypocoercivity (1)
- idealclass group (1)
- image analysis (1)
- image enhancement (1)
- image processing (1)
- image restoration (1)
- impulse control (1)
- impurity functions (1)
- incident wave (1)
- incompressible Euler equation (1)
- incompressible elasticity (1)
- incompressible limit (1)
- infinite-dimensional analysis (1)
- infinite-dimensional manifold (1)
- inflation-linked product (1)
- information (1)
- inhibitory synaptic transmission (1)
- initial temperature (1)
- initial temperature reconstruction (1)
- instantaneous phase (1)
- integer GARCH (1)
- integer-valued time series (1)
- integral constitutive equations (1)
- intensity (1)
- intensity map segmentation (1)
- interest oriented portfolios (1)
- internal approximation (1)
- intersection local time (1)
- interval graphs (1)
- intra- and extracellular proton dynamics (1)
- invariant theory (1)
- inverse Fourier transform (1)
- inversion method (1)
- isogeometric analysis (IGA) (1)
- iterative bandwidth choice (1)
- jump diffusion (1)
- jump-diffusion process (1)
- junction (1)
- k-cardinality minimum cut (1)
- k-max (1)
- kardinalzahl (1)
- kernel (1)
- kernel estimate (1)
- kernel estimates (1)
- kinetic approach (1)
- kinetic models (1)
- kinetic semiconductor equations (1)
- kinetic theory (1)
- knapsack (1)
- kombinatorische Optimierung (1)
- konvexe Analysis (1)
- kooperative Spieltheorie (1)
- label setting algorithm (1)
- large deviations (1)
- large investor (1)
- large scale integer programming (1)
- lattice Boltzmann (1)
- level K-algebras (1)
- life insurance (1)
- limes (1)
- limit models (1)
- limit theorems (1)
- linear code (1)
- linear programming (1)
- linear systems (1)
- linear transport equation (1)
- local approximation of sea surface topography (1)
- local bandwidths (1)
- local multiscale (1)
- local orientation (1)
- local search algorithm (1)
- local stationarity (1)
- local support (1)
- local trigonometric packets (1)
- local-global conjectures (1)
- localization (1)
- locally compact (1)
- locally compact kernels (1)
- locally maximal clone (1)
- locally stationary process (1)
- locally supported (Green's) vector wavelets (1)
- locally supported (Green’s) vector wavelets (1)
- locally supported wavelets (1)
- location (1)
- location problem (1)
- locational planning (1)
- log averaging methods (1)
- log-utility (1)
- logarithmic average (1)
- logarithmic averages (1)
- logarithmic utility (1)
- logical analysis (1)
- logische Analyse (1)
- lokal kompakt (1)
- lokaler Träger (1)
- lokalisierende Basis (1)
- lokalisierende Kerne (1)
- longevity bonds (1)
- loss analysis (1)
- low discrepancy (1)
- low-rank approximation (1)
- machine learning (1)
- macro derivative (1)
- magnetic field (1)
- market crash (1)
- market manipulation (1)
- markov model (1)
- martingale measu (1)
- martingale optimality principle (1)
- mathematica education (1)
- mathematical modelling (1)
- mathematical morphology (1)
- mathematische Modellierung (1)
- matrix decomposition (1)
- matrix problems (1)
- matroid flows (1)
- maximal dynamic flow (1)
- maximum a posteriori estimation (1)
- maximum capacity path (1)
- maximum entropy (1)
- maximum entropy moment (1)
- maximum flows (1)
- maximum likelihood estimation (1)
- maximum-entropy (1)
- mean-variance approach (1)
- mechanism design (1)
- mehrwertig (1)
- mesh deformation (1)
- mesh-free method (1)
- method of fundamental solutions (1)
- micromechanics (1)
- minimal paths (1)
- minimal polynomial (1)
- minimal spanning tree (1)
- minimaler Schnittbaum (1)
- minimax estimation (1)
- minimax rate (1)
- minimax risk (1)
- minimum cost flows (1)
- minimum cut (1)
- minimum cut tree (1)
- minimum fundamental cut basis (1)
- mixed convection (1)
- mixed methods (1)
- mixed multiscale finite element methods (1)
- mixing (1)
- mixture models (1)
- mixture of quantum fluids and classical fluids (1)
- model order reduction (1)
- model reduction (1)
- moduli space (1)
- moduli spaces (1)
- moment methods (1)
- monlinear vibration (1)
- monodromy (1)
- monogenic signals (1)
- monoid- and group-presentations (1)
- monotone Konvergenz (1)
- monotone consecutive arrangement (1)
- moving contact line (1)
- multi scale (1)
- multi-asset option (1)
- multi-class image segmentation (1)
- multi-level Monte Carlo (1)
- multi-phase flow (1)
- multi-scale model (1)
- multicategory (1)
- multicriteria minimal path problem is presented (1)
- multidimensional Kohonen algorithm (1)
- multifilament superconductor (1)
- multigrid method (1)
- multigrid methods (1)
- multileaf collimator sequencing (1)
- multiliead collimator sequencing (1)
- multiobjective optimization (1)
- multipatch (1)
- multiple collision frequencies (1)
- multiple objective (1)
- multiple objective optimization (1)
- multiresolution analysis (1)
- multiscale analysis (1)
- multiscale approximation (1)
- multiscale approximation on regular telluroidal surfaces (1)
- multiscale denoising (1)
- multiscale methods (1)
- multiscale modeling (1)
- multiscale models (1)
- multivariate chi-square-test (1)
- mutiresolution (1)
- naive diversification (1)
- neighborhood search (1)
- network congestion game (1)
- network flow (1)
- network location (1)
- network synthesis (1)
- netzgenerierung (1)
- neural networks (1)
- never-meet property (1)
- nicht-newtonsche Strömungen (1)
- nichtlineare Druckkorrektor (1)
- nichtlineare Modellreduktion (1)
- nichtlineare Netzwerke (1)
- nichtparametrisch (1)
- non square linear system solving (1)
- non-Gaussia non-i.i.d. errors (1)
- non-commutative geometry (1)
- non-convex body (1)
- non-convex optimization (1)
- non-desarguesian plane (1)
- non-linear wavelet thresholding (1)
- non-local filtering (1)
- non-newtonian flow (1)
- non-parametric regression (1)
- non-stationary time series (1)
- nonconvex optimization (1)
- noninformative prior (1)
- nonlinear circuits (1)
- nonlinear elasticity (1)
- nonlinear finite element method (1)
- nonlinear heat equation (1)
- nonlinear inverse problem (1)
- nonlinear model reduction (1)
- nonlinear pressure correction (1)
- nonlinear term structure dependence (1)
- nonlinear thresholding (1)
- nonlinear vibration analysis (1)
- nonlinear wavelet thresholding (1)
- nonlocal filtering (1)
- nonlocal sample dependence (1)
- nonnegative matrix factorization (1)
- nonparametric (1)
- nonparametric regression and (spectral) density estimation (1)
- nonwovens (1)
- norm (1)
- normal cone (1)
- normal mode (1)
- normality (1)
- normalization (1)
- normed residuum (1)
- number fields (1)
- number of objectives (1)
- numeraire portfolios (1)
- numerical integration (1)
- numerical irreducible decomposition (1)
- numerical methods (1)
- numerical methods for stiff equations (1)
- numerics for pdes (1)
- numerische Strömungssimulation (1)
- numerisches Verfahren (1)
- oblique derivative (1)
- one-dimensional self-organization (1)
- operator splitting (1)
- optimal capital structure (1)
- optimal consumption and investment (1)
- optimal portfolios (1)
- optimal rate of convergence (1)
- optiman stopping (1)
- option pricing (1)
- option valuation (1)
- order selection (1)
- order-three density (1)
- order-two density (1)
- orthogonal bandlimited and non-bandlimited wavelets (1)
- ovoids (1)
- parallel numerical algorithms (1)
- parameter choice (1)
- parameter identification (1)
- partial differential equation (1)
- partial differential equations (1)
- partial differential-algebraic equations (1)
- partial information (1)
- partition of unity (1)
- path-dependent options (1)
- pattern (1)
- penalization (1)
- penalty methods (1)
- penalty-free formulation (1)
- personnel scheduling (1)
- petroleum exploration (1)
- physicians (1)
- piezoelectric periodic surface acoustic wave filters (1)
- planar Brownian motion (1)
- planar polynomial (1)
- polycyclic group rings (1)
- polyhedral analysis (1)
- polyhedral norm (1)
- polynomial weight functions (1)
- porous media flow (1)
- portfolio (1)
- portfolio decision (1)
- portfolio optimisation (1)
- portfolio-optimization (1)
- poröse Medien (1)
- positivity preserving time integration (1)
- posterior collapse (1)
- potential (1)
- potential operators (1)
- preconditioners (1)
- predictive control (1)
- prefix reduction (1)
- prefix string rewriting (1)
- prefix-rewriting (1)
- preservation of relations (1)
- pressure correction (1)
- price of anarchy (1)
- price of stability (1)
- primal-dual algorithm (1)
- probability distribution (1)
- projected quasi-gradient method (1)
- projection method (1)
- projective surfaces (1)
- properly efficient solution (1)
- proximation (1)
- proxy modeling (1)
- pseudospectral methods (1)
- public transport (1)
- public transportation (1)
- pyramid schemes (1)
- pyramids (1)
- quadratic forms (1)
- quadrinomial tree (1)
- qualitative threshold model (1)
- quantile autoregression (1)
- quasi-Monte Carlo (1)
- quasi-P (1)
- quasi-SH (1)
- quasi-SV (1)
- quasi-variational inequalities (1)
- quasihomogeneity (1)
- quasiregular group (1)
- quasireguläre Gruppe (1)
- quasivarieties (1)
- quickest path (1)
- radiation therapy (1)
- radiative heat transfer (1)
- rainflow (1)
- random noise (1)
- rare disasters (1)
- rarefied gas flows (1)
- rate-independency (1)
- ratio ergodic theorem (1)
- raum-zeitliche Analyse (1)
- reaction-diffusion-taxis equations (1)
- reaction-diffusion-transport equations (1)
- real quadratic number fields (1)
- reconstruction formula (1)
- reconstructions (1)
- redundant constraint (1)
- reference prior (1)
- reflectionless boundary condition (1)
- reflexionslose Randbedingung (1)
- refraction (1)
- regime-shift model (1)
- regularization by wavelets (1)
- regularization methods (1)
- reguläre Fläche (1)
- reinitialization (1)
- rela (1)
- representative systems (1)
- residual based error formula (1)
- resource constrained shortest path problem (1)
- rewriting (1)
- rheology (1)
- risk analysis (1)
- risk measures (1)
- risk reduction (1)
- robust network flows (1)
- robustness (1)
- rostering (1)
- s external gravitational field (1)
- sampling (1)
- satellite gradiometry (1)
- satellite-to-satellite tracking (1)
- sawtooth effect (1)
- scalar and vectorial wavelets (1)
- scalar conservation laws (1)
- scalarization (1)
- scale discrete spherical vector wavelets (1)
- scaled boundary isogeometric analysis (1)
- scaled boundary parametrizations (1)
- scaled translates (1)
- scaling functions (1)
- scheduling (1)
- scheduling theory (1)
- schlecht gestellt (1)
- schnelle Approximation (1)
- second class group (1)
- second order upwind discretization (1)
- seismic tomography (1)
- seismic wave (1)
- selfish routing (1)
- semi-classical limits (1)
- semigroup of values (1)
- semisprays (1)
- sensitivities (1)
- separation problem (1)
- sequential test (1)
- set covering (1)
- severely ill-posed inverse problems (1)
- shape optimization (1)
- sheaf theory (1)
- shear flow (1)
- shock wave (1)
- short-time periodogram (1)
- shortest path problem (1)
- sieve estimate (1)
- similarity measures (1)
- simulierte Finanzzeitreihe (1)
- single layer kernel (1)
- singular fluxes (1)
- singular optimal control (1)
- singular spaces (1)
- singuläre Räume (1)
- sink location (1)
- slope limiter (1)
- smoothing (1)
- solution formula (1)
- sparse interpolation of multivariate rational functions (1)
- sparse multivariate polynomial interpolation (1)
- sparsity (1)
- special entropies (1)
- spectral sequences (1)
- spectrogram (1)
- speech recognition (1)
- sphere (1)
- spherical decomposition (1)
- spherical splines (1)
- spline (1)
- spline and wavelet based determination of the geoid and the gravitational potential (1)
- spline-wavelets (1)
- splitting function (1)
- sputtering process (1)
- squares (1)
- stability (1)
- stability uniformly in the mean free path (1)
- star-shaped domain (1)
- stationary solutions (1)
- statistical experiment (1)
- steady Boltzmann equation (1)
- stimulus response data (1)
- stochastic arbitrage (1)
- stochastic coefficient (1)
- stochastic differential equations (1)
- stochastic interest rate (1)
- stochastic optimal control (1)
- stochastic processes (1)
- stochastic stability (1)
- stochastische Arbitrage (1)
- stop location (1)
- stop- and play-operator (1)
- stop- and play-operators (1)
- stratifolds (1)
- strictly quasi-convex functions (1)
- strong equilibria (1)
- strong theorems (1)
- strongly polynomial-time algorithm (1)
- structure tensor (1)
- subgradient (1)
- subgroup presentation problem (1)
- superposed fluids (1)
- superstep cycles (1)
- surface measures (1)
- surrender options (1)
- surrogate algorithm (1)
- systems (1)
- syzygies (1)
- tail dependence coefficient (1)
- tax (1)
- technische Analyse (1)
- tension problems (1)
- tensions (1)
- tensor product basis (1)
- test (1)
- texture (1)
- thermal equilibrium state (1)
- threshold choice (1)
- time delays (1)
- time-delayed carrying capacities (1)
- time-dependent shortest path problem (1)
- time-frequency plan (1)
- time-varying autoregression (1)
- time-varying covariance (1)
- topological asymptotic expansion (1)
- toric geometry (1)
- torische Geometrie (1)
- total latency (1)
- total variation (1)
- total variation spatial regularization (1)
- traffic planning (1)
- transit operations (1)
- translation invariant spaces (1)
- translinear circuits (1)
- translineare Schaltungen (1)
- transmission conditions (1)
- trial systems (1)
- triclinic medium (1)
- tropical geometry (1)
- tumor acidity (1)
- tumor cell invasion (1)
- tumor cell migration (1)
- two-scale expansion (1)
- unbeschränktes Potential (1)
- unbounded potential (1)
- uncapacitated facility location (1)
- uncertainty principle (1)
- unendlich (1)
- uniform central limit theorem (1)
- uniform consistency (1)
- uniform ergodicity (1)
- unimodular certification (1)
- unimodularity (1)
- value preserving portfolios (1)
- value semigroup (1)
- value-at-risk (1)
- valuing contracts (1)
- variable selection (1)
- variational methods (1)
- variational model (1)
- vector bundles (1)
- vector measure (1)
- vector wavelets (1)
- vectorial multiresolution analysis (1)
- vehicular traffic (1)
- verication theorem (1)
- vertical velocity (1)
- vertikale Geschwindigkeiten (1)
- viscoelastic fluids (1)
- viscosity solutions (1)
- volatility arbitrage (1)
- vortex seperation (1)
- wave propagation (1)
- wavelet estimators (1)
- wavelet packets (1)
- wavelet thresholding (1)
- wavelet transform (1)
- weak dependence (1)
- weak solution theory (1)
- weak solutions (1)
- weakly/ strictly pareto optima (1)
- weight optimization (1)
- windowed Fourier transform (1)
- winner definition (1)
- worst-case (1)
- Äquisingularität (1)
- Überflutung (1)
- Überflutungsrisiko (1)
- Übergangsbedingungen (1)
Faculty / Organisational entity
Diese Diplomarbeit gibt eine kurze Einführung in das Gebiet der Diffusionsprozesse (beschrieben als Lösungen stochastischer Differentialgleichungen) und der großen Abweichungen. Mit Methoden aus dem Gebiet der großen Abweichungen wird dann das asymptotische Verhalten des Bayesrisikos für die unterscheidung zweier Diffusionsprozesse untersucht.
Anhand des vom Gutachterausschuß der Stadt Kaiserlautern zur Verfügung gestellten Datenmaterials soll untersucht werden, welche Faktoren den Verkehrswert eines bebauten Grundstücks beeinflussen. Mit diesen Erkenntnissen soll eine möglichst einfache Formel ermittelt werden, die eine Schätzung für den Verkehrswert liefert, und die dabei die in der Vergangenheit erzielten Kaufpreise berücksichtigt. Für die Lösung dieser Aufgabe bietet sich das Verfahren der multiplen linearen Regression an. Auf die theoretischen Grundlagen soll hier nicht näher eingegangen werden, man findet sie in jedem Buch über mathematische Statistik, oder in [1]. Bei der Analyse der Daten wurde im großen und ganzen der Weg eingeschlagen, den Angelika Schwarz in [1] beschreibt. Ihre Ergebnisse lassen sich jedoch nicht direkt übertragen, da die dort betrachteten Grundstücke unbebaut waren. Da bei der statistischen Auswertung großer Datenmengen ein immenser Rechenaufwand anfällt, ist es unverzichtbar, professionelle statistische Software einzusetzen. Es stand das Programm S-Plus 2.0 (PC-Version für Windows) zur Verfügung. Sämtliche Berechnungen und alle Grafiken in diesem Bericht wurden in S-Plus erstellt.
We consider the problem to evacuate several regions due to river flooding, where sufficient time is given to plan ahead. To ensure a smooth evacuation procedure, our model includes the decision which regions to assign to which shelter, and when evacuation orders should be issued, such that roads do not become congested.
Due to uncertainty in weather forecast, several possible scenarios are simultaneously considered in a robust optimization framework. To solve the resulting integer program, we apply a Tabu search algorithm based on decomposing the problem into better tractable subproblems. Computational experiments on random instances and an instance based on Kulmbach, Germany, data show considerable improvement compared to an MIP solver provided with a strong starting solution.
Das zinsoptimierte Schuldenmanagement hat zum Ziel, eine möglichst effiziente Abwägung zwischen den erwarteten Finanzierungskosten einerseits und den Risiken für den Staatshaushalt andererseits zu finden. Um sich diesem Spannungsfeld zu nähern, schlagen wir erstmals die Brücke zwischen den Problemstellungen des Schuldenmanagements und den Methoden der zeitkontinuierlichen, dynamischen Portfoliooptimierung.
Das Schlüsselelement ist dabei eine neue Metrik zur Messung der Finanzierungskosten, die Perpetualkosten. Diese spiegeln die durchschnittlichen zukünftigen Finanzierungskosten wider und beinhalten sowohl die bereits bekannten Zinszahlungen als auch die noch unbekannten Kosten für notwendige Anschlussfinanzierungen. Daher repräsentiert die Volatilität der Perpetualkosten auch das Risiko einer bestimmten Strategie; je langfristiger eine Finanzierung ist, desto kleiner ist die Schwankungsbreite der Perpetualkosten.
Die Perpetualkosten ergeben sich als Produkt aus dem Barwert eines Schuldenportfolios und aus der vom Portfolio unabhängigen Perpetualrate. Für die Modellierung des Barwertes greifen wir auf das aus der dynamischen Portfoliooptimierung bekannte Konzept eines selbstfinanzierenden Bondportfolios zurück, das hier auf einem mehrdimensionalen affin-linearen Zinsmodell basiert. Das Wachstum des Schuldenportfolios wird dabei durch die Einbeziehung des Primärüberschusses des Staates gebremst bzw. verhindert, indem wir diesen als externen Zufluss in das selbstfinanzierende Modell aufnehmen.
Wegen der Vielfältigkeit möglicher Finanzierungsinstrumente wählen wir nicht deren Wertanteile als Kontrollvariable, sondern kontrollieren die Sensitivitäten des Portfolios gegenüber verschiedenen Zinsbewegungen. Aus optimalen Sensitivitäten können in einem nachgelagerten Schritt dann optimale Wertanteile für verschiedenste Finanzierungsinstrumente abgeleitet werden. Beispielhaft demonstrieren wir dies mittels Rolling-Horizon-Bonds unterschiedlicher Laufzeit.
Schließlich lösen wir zwei Optimierungsprobleme mit Methoden der stochastischen Kontrolltheorie. Dabei wird stets der erwartete Nutzen der Perpetualkosten maximiert. Die Nutzenfunktionen sind jeweils an das Schuldenmanagement angepasst und zeichnen sich insbesondere dadurch aus, dass höhere Kosten mit einem niedrigeren Nutzen einhergehen. Im ersten Problem betrachten wir eine Potenznutzenfunktion mit konstanter relativer Risikoaversion, im zweiten wählen wir eine Nutzenfunktion, welche die Einhaltung einer vorgegebenen Schulden- bzw. Kostenobergrenze garantiert.
Zeitreihen und Modalanalyse
(1987)
Die Arbeit ist zu verstehen als ein Teil im großen Projekt der Universität Kaiserslautern, das sich unter dem Namen Technomathematik um die dringend erforderliche Verständigung zwischen Technik und Mathematik bemüht.; Der große Leitfaden war das Buch von Natke: Einführung in Theorie und Praxis der Zeitreihen- und Modalanalyse, Schilderung der wesentlichen dort verwendeten Ideen der indirekten Systemidentifikation sowie des wahrscheinlichkeitstheoretischen und physikalisch-technischen Hintergrundes.
Yield Curves and Chance-Risk Classification: Modeling, Forecasting, and Pension Product Portfolios
(2021)
This dissertation consists of three independent parts: The yield curve shapes generated by interest rate models, the yield curve forecasting, and the application of the chance-risk classification to a portfolio of pension products. As a component of the capital market model, the yield curve influences the chance-risk classification which was introduced to improve the comparability of pension products and strengthen consumer protection. Consequently, all three topics have a major impact on this essential safeguard.
Firstly, we focus on the obtained yield curve shapes of the Vasicek interest rate models. We extend the existing studies on the attainable yield curve shapes in the one-factor Vasicek model by analysis of the curvature. Further, we show that the two-factor Vasicek model can explain significantly more effects that are observed at the market than its one-factor variant. Among them is the occurrence of dipped yield curves.
We further introduce a general change of measure framework for the Monte Carlo simulation of the Vasicek model under a subjective measure. This can be used to avoid the occurrence of a far too high frequency of inverse yield curves with growing time.
Secondly, we examine different time series models including machine learning algorithms forecasting the yield curve. For this, we consider statistical time series models such as autoregression and vector autoregression. Their performances are compared with the performance of a multilayer perceptron, a fully connected feed-forward neural network. For this purpose, we develop an extended approach for the hyperparameter optimization of the perceptron which is based on standard procedures like Grid and Random Search but allows to search a larger hyperparameter space. Our investigation shows that multilayer perceptrons outperform statistical models for long forecast horizons.
The third part deals with the chance-risk classification of state-subsidized pension products in Germany as well as its relevance for customer consulting. To optimize the use of the chance-risk classes assigned by Produktinformationsstelle Altersvorsorge gGmbH, we develop a procedure for determining the chance-risk class of different portfolios of state-subsidized pension products under the constraint that the portfolio chance-risk class does not exceed the customer's risk preference. For this, we consider a portfolio consisting of two new pension products as well as a second one containing a product already owned by the customer as well as the offer of a new one. This is of particular interest for customer consulting and can include other assets of the customer. We examine the properties of various chance and risk parameters as well as their corresponding mappings and show that a diversification effect exists. Based on the properties, we conclude that the average final contract values have to be used to obtain the upper bound of the portfolio chance-risk class. Furthermore, we develop an approach for determining the chance-risk class over the contract term since the chance-risk class is only assigned at the beginning of the accumulation phase. On the one hand, we apply the current legal situation, but on the other hand, we suggest an approach that requires further simulations. Finally, we translate our results into recommendations for customer consultation.
Gegenstand dieser Arbeit ist die Entwicklung eines Wärmetransportmodells für tiefe geothermische (hydrothermale) Reservoire. Existenz- und Eindeutigkeitsaussagen bezüglich einer schwachen Lösung des vorgestellten Modells werden getätigt. Weiterhin wird ein Verfahren zur Approximation dieser Lösung basierend auf einem linearen Galerkin-Schema dargelegt, wobei sowohl die Konvergenz nachgewiesen als auch eine Konvergenzrate erarbeitet werden.
Wreath product groups \(C_\ell \wr \mathfrak{S}_n\) have a rich combinatorial representation theory coming from the symmetric group case and involving partitions, Young tableaux, and Specht modules. To such a wreath product group \(W\), one can associate various algebras and geometric objects: Hecke algebras, quantum groups, Hilbert schemes, Calogero--Moser spaces, and (restricted) rational Cherednik algebras. Over the years, surprising connections have been made between a lot of these objects, with many of these connections having been traced back to combinatorial constructions and properties of the group \(W\) itself.
In this thesis, we have studied one of the algebras, namely the restricted rational Cherednik algebra \(\overline{\mathsf{H}}_\mathbf{c}(W)\), in order to find combinatorial models which describe certain representation theoretical phenomena around \(\overline{\mathsf{H}}_\mathbf{c}(W)\). In particular, we generalize a result by Gordon and describe the graded \(W\)-characters of the simple modules of \(\overline{\mathsf{H}}_\mathbf{c}(W)\) for generic parameter \(\mathbf{c}\) using Haiman's wreath Macdonald polynomials. These graded \(W\)-characters turn out to be specializations of Haiman's wreath Macdonald polynomials. In the non-generic parameter case, we use recent results by Maksimau to combinatorially express an inductive rule of \(\overline{\mathsf{H}}_\mathbf{c}(W)\)-modules first described by Bellamy. We use our results in type \(B\) to describe the (ungraded) \(B_n\)-character of simple \(\overline{\mathsf{H}}_\mathbf{c}(B_n)\)-modules associated to bipartitions with one empty part. Afterwards, we relate this combinatorial induction to various other algebras and families of \(W\)-characters found in the literature such as Lusztig's constructible characters, as well as detail some connections between generic and non-generic parameter using wreath Macdonald polynomials.
In this thesis we extend the worst-case modeling approach as first introduced by Hua and Wilmott (1997) (option pricing in discrete time) and Korn and Wilmott (2002) (portfolio optimization in continuous time) in various directions.
In the continuous-time worst-case portfolio optimization model (as first introduced by Korn and Wilmott (2002)), the financial market is assumed to be under the threat of a crash in the sense that the stock price may crash by an unknown fraction at an unknown time. It is assumed that only an upper bound on the size of the crash is known and that the investor prepares for the worst-possible crash scenario. That is, the investor aims to find the strategy maximizing her objective function in the worst-case crash scenario.
In the first part of this thesis, we consider the model of Korn and Wilmott (2002) in the presence of proportional transaction costs. First, we treat the problem without crashes and show that the value function is the unique viscosity solution of a dynamic programming equation (DPE) and then construct the optimal strategies. We then consider the problem in the presence of crash threats, derive the corresponding DPE and characterize the value function as the unique viscosity solution of this DPE.
In the last part, we consider the worst-case problem with a random number of crashes by proposing a regime switching model in which each state corresponds to a different crash regime. We interpret each of the crash-threatened regimes of the market as states in which a financial bubble has formed which may lead to a crash. In this model, we prove that the value function is a classical solution of a system of DPEs and derive the optimal strategies.
In 2002, Korn and Wilmott introduced the worst-case scenario optimal portfolio approach.
They extend a Black-Scholes type security market, to include the possibility of a
crash. For the modeling of the possible stock price crash they use a Knightian uncertainty
approach and thus make no probabilistic assumption on the crash size or the crash time distribution.
Based on an indifference argument they determine the optimal portfolio process
for an investor who wants to maximize the expected utility from final wealth. In this thesis,
the worst-case scenario approach is extended in various directions to enable the consideration
of stress scenarios, to include the possibility of asset defaults and to allow for parameter
uncertainty.
Insurance companies and banks regularly have to face stress tests performed by regulatory
instances. In the first part we model their investment decision problem that includes stress
scenarios. This leads to optimal portfolios that are already stress test prone by construction.
The solution to this portfolio problem uses the newly introduced concept of minimum constant
portfolio processes.
In the second part we formulate an extended worst-case portfolio approach, where asset
defaults can occur in addition to asset crashes. In our model, the strictly risk-averse investor
does not know which asset is affected by the worst-case scenario. We solve this problem by
introducing the so-called worst-case crash/default loss.
In the third part we set up a continuous time portfolio optimization problem that includes
the possibility of a crash scenario as well as parameter uncertainty. To do this, we combine
the worst-case scenario approach with a model ambiguity approach that is also based on
Knightian uncertainty. We solve this portfolio problem and consider two concrete examples
with box uncertainty and ellipsoidal drift ambiguity.
Aufgrund der vernetzten Strukturen und Wirkungszusammenhänge dynamischer Systeme werden die zugrundeliegenden mathematischen Modelle meist sehr komplex und erfordern ein hohes mathematisches Verständnis und Geschick. Bei Verwendung von spezieller Software können jedoch auch ohne tiefgehende mathematische oder informatorische Fachkenntnisse komplexe Wirkungsnetze dynamischer Systeme interaktiv erstellt werden. Als Beispiel wollen wir schrittweise das Modell einer Miniwelt entwerfen und Aussagen bezüglich ihrer Bevölkerungsentwicklung treffen.
Die Akustik liefert einen interessanten Hintergrund, interdisziplinären und fächerverbindenen Unterricht zwischen Mathematik, Physik und Musik durchzuführen. SchülerInnen können hierbei beispielsweise experimentell tätig sein, indem sie Audioaufnahmen selbst erzeugen und sich mit Computersoftware Frequenzspektren erzeugen lassen. Genauso können die Schüler auch Frequenzspektren vorgeben und daraus Klänge erzeugen. Dies kann beispielsweise dazu dienen, den Begriff der Obertöne im Musikunterricht physikalisch oder mathematisch greifbar zu machen oder in der Harmonielehre Frequenzverhältnisse von Intervallen und Dreiklängen näher zu untersuchen.
Der Computer ist hier ein sehr nützliches Hilfsmittel, da der mathematische Hintergrund dieser Aufgabe -- das Wechseln zwischen Audioaufnahme und ihrem Frequenzbild -- sich in der Fourier-Analysis findet, die für SchülerInnen äußerst anspruchsvoll ist. Indem man jedoch die Fouriertransformation als numerisches Hilfsmittel einführt, das nicht im Detail verstanden werden muss, lässt sich an anderer Stelle interessante Mathematik betreiben und die Zusammenhänge zwischen Akustik und Musik können spielerisch erfahren werden.
Im folgenden Beitrag wird eine Herangehensweise geschildert, wie wir sie bereits bei der Felix-Klein-Modellierungswoche umgesetzt haben: Die SchülerInnen haben den Auftrag erhalten, einen Synthesizer zu entwickeln, mit dem verschiedene Musikinstrumente nachgeahmt werden können. Als Hilfsmittel haben sie eine kurze Einführung in die Eigenschaften der Fouriertransformation erhalten, sowie Audioaufnahmen verschiedener Instrumente.
Using particle methods to solve the Boltzmann equation for rarefied gases numerically, in realistic streaming problems, huge differences in the total number of particles per cell arise. In order to overcome the resulting numerical difficulties the application of a weighted particle concept is well-suited. The underlying idea is to use different particle masses in different cells depending on the macroscopic density of the gas. Discrepance estimates and numerical results are given.
Weighted k-cardinality trees
(1992)
We consider the k -CARD TREE problem, i.e., the problem of finding in a given undirected graph G a subtree with k edges, having minimum weight. Applications of this problem arise in oil-field leasing and facility layout. While the general problem is shown to be strongly NP hard, it can be solved in polynomial time if G is itself a tree. We give an integer programming formulation of k-CARD TREE, and an efficient exact separation routine for a set of generalized subtour elimination constraints. The polyhedral structure of the convex huLl of the integer solutions is studied.
In an undirected graph G we associate costs and weights to each edge. The weight-constrained minimum spanning tree problem is to find a spanning tree of total edge weight at most a given value W and minimum total costs under this restriction. In this thesis a literature overview on this NP-hard problem, theoretical properties concerning the convex hull and the Lagrangian relaxation are given. We present also some in- and exclusion-test for this problem. We apply a ranking algorithm and the method of approximation through decomposition to our problem and design also a new branch and bound scheme. The numerical results show that this new solution approach performs better than the existing algorithms.
Given a finite set of points in the plane and a forbidden region R, we want to find a point X not an element of int(R), such that the weighted sum to all given points is minimized. This location problem is a variant of the well-known Weber Problem, where we measure the distance by polyhedral gauges and allow each of the weights to be positive or negative. The unit ball of a polyhedral gauge may be any convex polyhedron containing the origin. This large class of distance functions allows very general (practical) settings - such as asymmetry - to be modeled. Each given point is allowed to have its own gauge and the forbidden region R enables us to include negative information in the model. Additionally the use of negative and positive weights allows to include the level of attraction or dislikeness of a new facility. Polynomial algorithms and structural properties for this global optimization problem (d.c. objective function and a non-convex feasible set) based on combinatorial and geometrical methods are presented.
We introduce a class of models for time series of counts which include INGARCH-type models as well as log linear models for conditionally Poisson distributed data. For those processes, we formulate simple conditions for stationarity and weak dependence with a geometric rate. The coupling argument used in the proof serves as a role model for a similar treatment of integer-valued time series models based on other types of thinning operations.
By means of the limit and jump relations of classical potential theory the framework of a wavelet approach on a regular surface is established. The properties of a multiresolution analysis are verified, and a tree algorithm for fast computation is developed based on numerical integration. As applications of the wavelet approach some numerical examples are presented, including the zoom-in property as well as the detection of high frequency perturbations. At the end we discuss a fast multiscale representation of the solution of (exterior) Dirichlet's or Neumann's boundary-value problem corresponding to regular surfaces.
A wavelet technique, the wavelet-Mie-representation, is introduced for the analysis and modelling of the Earth's magnetic field and corresponding electric current distributions from geomagnetic data obtained within the ionosphere. The considerations are essentially based on two well-known geomathematical keystones, (i) the Helmholtz-decomposition of spherical vector fields and (ii) the Mie-representation of solenoidal vector fields in terms of poloidal and toroidal parts. The wavelet-Mie-representation is shown to provide an adequate tool for geomagnetic modelling in the case of ionospheric magnetic contributions and currents which exhibit spatially localized features. An important example are ionospheric currents flowing radially onto or away from the Earth. To demonstrate the functionality of the approach, such radial currents are calculated from vectorial data of the MAGSAT and CHAMP satellite missions.
* naive examples which show drawbacks of discrete wavelet transform and windowed Fourier transform; * adaptive partition (with a 'best basis' approach) of speech-like signals by means of local trigonometric bases with orthonormal windows. * extraction of formant-like features from the cosine transform; * further proceedingings for classification of vowels or voiced speech are suggested at the end.
With this article we first like to give a brief review on wavelet thresholding methods in non-Gaussian and non-i.i.d. situations, respectively. Many of these applications are based on Gaussian approximations of the empirical coefficients. For regression and density estimation with independent observations, we establish joint asymptotic normality of the empirical coefficients by means of strong approximations. Then we describe how one can prove asymptotic normality under mixing conditions on the observations by cumulant techniques.; In the second part, we apply these non-linear adaptive shrinking schemes to spectral estimation problems for both a stationary and a non-stationary time series setup. For the latter one, in a model of Dahlhaus on the evolutionary spectrum of a locally stationary time series, we present two different approaches. Moreover, we show that in classes of anisotropic function spaces an appropriately chosen wavelet basis automatically adapts to possibly different degrees of regularity for the different directions. The resulting fully-adaptive spectral estimator attains the rate that is optimal in the idealized Gaussian white noise model up to a logarithmic factor.
We derive minimax rates for estimation in anisotropic smoothness classes. This rate is attained by a coordinatewise thresholded wavelet estimator based on a tensor product basis with separate scale parameter for every dimension. It is shown that this basis is superior to its one-scale multiresolution analog, if different degrees of smoothness in different directions are present.; As an important application we introduce a new adaptive wavelet estimator of the time-dependent spectrum of a locally stationary time series. Using this model which was resently developed by Dahlhaus, we show that the resulting estimator attains nearly the rate, which is optimal in Gaussian white noise, simultaneously over a wide range of smoothness classes. Moreover, by our new approach we overcome the difficulty of how to choose the right amount of smoothing, i.e. how to adapt to the appropriate resolution, for reconstructing the local structure of the evolutionary spectrum in the time-frequency plane.
We consider wavelet estimation of the time-dependent (evolutionary) power spectrum of a locally stationary time series. Allowing for departures from stationary proves useful for modelling, e.g., transient phenomena, quasi-oscillating behaviour or spectrum modulation. In our work wavelets are used to provide an adaptive local smoothing of a short-time periodogram in the time-freqeuncy plane. For this, in contrast to classical nonparametric (linear) approaches we use nonlinear thresholding of the empirical wavelet coefficients of the evolutionary spectrum. We show how these techniques allow for both adaptively reconstructing the local structure in the time-frequency plane and for denoising the resulting estimates. To this end a threshold choice is derived which is motivated by minimax properties w.r.t. the integrated mean squared error. Our approach is based on a 2-d orthogonal wavelet transform modified by using a cardinal Lagrange interpolation function on the finest scale. As an example, we apply our procedure to a time-varying spectrum motivated from mobile radio propagation.
The article is concerned with the modelling of ionospheric current systems from induced magnetic fields measured by satellites in a multiscale framework. Scaling functions and wavelets are used to realize a multiscale analysis of the function spaces under consideration and to establish a multiscale regularization procedure for the inversion of the considered vectorial operator equation. Based on the knowledge of the singular system a regularization technique in terms of certain product kernels and corresponding convolutions can be formed. In order to reconstruct ionospheric current systems from satellite magnetic field data, an inversion of the Biot-Savart's law in terms of multiscale regularization is derived. The corresponding operator is formulated and the singular values are calculated. The method is tested on real magnetic field data of the satellite CHAMP and the proposed satellite mission SWARM.
This work is dedicated to the wavelet modelling of regional and temporal variations of the Earth's gravitational potential observed by GRACE. In the first part, all required mathematical tools and methods involving spherical wavelets are introduced. Then we apply our method to monthly GRACE gravity fields. A strong seasonal signal can be identified, which is restricted to areas, where large-scale redistributions of continental water mass are expected. This assumption is analyzed and verified by comparing the time series of regionally obtained wavelet coefficients of the gravitational signal originated from hydrology models and the gravitational potential observed by GRACE. The results are in good agreement to previous studies and illustrate that wavelets are an appropriate tool to investigate regional time-variable effects in the gravitational field.
The thesis is concerned with the modelling of ionospheric current systems and induced magnetic fields in a multiscale framework. Scaling functions and wavelets are used to realize a multiscale analysis of the function spaces under consideration and to establish a multiscale regularization procedure for the inversion of the considered operator equation. First of all a general multiscale concept for vectorial operator equations between two separable Hilbert spaces is developed in terms of vector kernel functions. The equivalence to the canonical tensorial ansatz is proven and the theory is transferred to the case of multiscale regularization of vectorial inverse problems. As a first application, a special multiresolution analysis of the space of square-integrable vector fields on the sphere, e.g. the Earth’s magnetic field measured on a spherical satellite’s orbit, is presented. By this, a multiscale separation of spherical vector-valued functions with respect to their sources can be established. The vector field is split up into a part induced by sources inside the sphere, a part which is due to sources outside the sphere and a part which is generated by sources on the sphere, i.e. currents crossing the sphere. The multiscale technqiue is tested on a magnetic field data set of the satellite CHAMP and it is shown that crustal field determination can be improved by previously applying our method. In order to reconstruct ionspheric current systems from magnetic field data, an inversion of the Biot-Savart’s law in terms of multiscale regularization is defined. The corresponding operator is formulated and the singular values are calculated. Based on the konwledge of the singular system a regularzation technique in terms of certain product kernels and correponding convolutions can be formed. The method is tested on different simulations and on real magnetic field data of the satellite CHAMP and the proposed satellite mission SWARM.
In this paper we introduce a multiscale technique for the analysis of deformation phenomena of the Earth. Classically, the basis functions under use are globally defined and show polynomial character. In consequence, only a global analysis of deformations is possible such that, for example, the water load of an artificial reservoir is hardly to model in that way. Up till now, the alternative to realize a local analysis can only be established by assuming the investigated region to be flat. In what follows we propose a local analysis based on tools (Navier scaling functions and wavelets) taking the (spherical) surface of the Earth into account. Our approach, in particular, enables us to perform a zooming-in procedure. In fact, the concept of Navier wavelets is formulated in such a way that subregions with larger or smaller data density can accordingly be modelled with a higher or lower resolution of the model, respectively.
Wavelets on closed surfaces in Euclidean space R3 are introduced starting from a scale discrete wavelet transform for potentials harmonic down to a spherical boundary. Essential tools for approximation are integration formulas relating an integral over the sphere to suitable linear combinations of functional values (resp. normal derivatives) on the closed surface under consideration. A scale discrete version of multiresolution is described for potential functions harmonic outside the closed surface and regular at infinity. Furthermore, an exact fully discrete wavelet approximation is developed in case of band-limited wavelets. Finally, the role of wavelets is discussed in three problems, namely (i) the representation of a function on a closed surface from discretely given data, (ii) the (discrete) solution of the exterior Dirichlet problem, and (iii) the (discrete) solution of the exterior Neumann problem.
A multiscale method is introduced using spherical (vector) wavelets for the computation of the earth's magnetic field within source regions of ionospheric and magnetospheric currents. The considerations are essentially based on two geomathematical keystones, namely (i) the Mie representation of solenoidal vector fields in terms of toroidal and poloidal parts and (ii) the Helmholtz decomposition of spherical (tangential) vector fields. Vector wavelets are shown to provide adequate tools for multiscale geomagnetic modelling in form of a multiresolution analysis, thereby completely circumventing the numerical obstacles caused by vector spherical harmonics. The applicability and efficiency of the multiresolution technique is tested with real satellite data.
In this paper, the reflection and refraction of a plane wave at an interface between .two half-spaces composed of triclinic crystalline material is considered. It is shown that due to incidence of a plane wave three types of waves namely quasi-P (qP), quasi-SV (qSV) and quasi-SH (qSH) will be generated governed by the propagation condition involving the acoustic tensor. A simple procedure has been presented for the calculation of all the three phase velocities of the quasi waves. It has been considered that the direction of particle motion is neither parallel nor perpendicular to the direction of propagation. Relations are established between directions of motion and propagation, respectively. The expressions for reflection and refraction coefficients of qP, qSV and qSH waves are obtained. Numerical results of reflection and refraction coefficients are presented for different types of anisotropic media and for different types of incident waves. Graphical representation have been made for incident qP waves and for incident qSV and qSH waves numerical data are presented in two tables.
Vorlesung Logik
(2000)
Diese Doktorarbeit befasst sich mit Volatilitätsarbitrage bei europäischen Kaufoptionen und mit der Modellierung von Collateralized Debt Obligations (CDOs). Zuerst wird anhand einer Idee von Carr gezeigt, dass es stochastische Arbitrage in einem Black-Scholes-ähnlichen Modell geben kann. Danach optimieren wir den Arbitrage- Gewinn mithilfe des Erwartungswert-Varianz-Ansatzes von Markowitz und der Martingaltheorie. Stochastische Arbitrage im stochastischen Volatilitätsmodell von Heston wird auch untersucht. Ferner stellen wir ein Markoff-Modell für CDOs vor. Wir zeigen dann, dass man relativ schnell an die Grenzen dieses Modells stößt: Nach dem Ausfall einer Firma steigen die Ausfallintensitäten der überlebenden Firmen an, und kehren nie wieder zu ihrem Ausgangsniveau zurück. Dieses Verhalten stimmt aber nicht mit Beobachtungen am Markt überein: Nach Turbulenzen auf dem Markt stabilisiert sich der Markt wieder und daher würde man erwarten, dass die Ausfallintensitäten der überlebenden Firmen ebenfalls wieder abflachen. Wir ersetzen daher das Markoff-Modell durch ein Semi-Markoff-Modell, das den Markt viel besser nachbildet.
Vigenere-Verschlüsselung
(1999)
The present work deals with the (global and local) modeling of the windfield on the real topography of Rheinland-Pfalz. Thereby the focus is on the construction of a vectorial windfield from low, irregularly distributed data given on a topographical surface. The developed spline procedure works by means of vectorial (homogeneous, harmonic) polynomials (outer harmonics) which control the oscillation behaviour of the spline interpoland. In the process the characteristic of the spline curvature which defines the energy norm is assumed to be on a sphere inside the Earth interior and not on the Earth’s surface. The numerical advantage of this method arises from the maximum-minimum principle for harmonic functions.
In this thesis we classify simple coherent sheaves on Kodaira fibers of types II, III and IV (cuspidal and tacnode cubic curves and a plane configuration of three concurrent lines). Indecomposable vector bundles on smooth elliptic curves were classified in 1957 by Atiyah. In works of Burban, Drozd and Greuel it was shown that the categories of vector bundles and coherent sheaves on cycles of projective lines are tame. It turns out, that all other degenerations of elliptic curves are vector-bundle-wild. Nevertheless, we prove that the category of coherent sheaves of an arbitrary reduced plane cubic curve, (including the mentioned Kodaira fibers) is brick-tame. The main technical tool of our approach is the representation theory of bocses. Although, this technique was mainly used for purely theoretical purposes, we illustrate its computational potential for investigating tame behavior in wild categories. In particular, it allows to prove that a simple vector bundle on a reduced cubic curve is determined by its rank, multidegree and determinant, generalizing Atiyah's classification. Our approach leads to an interesting class of bocses, which can be wild but are brick-tame.
The mathematical modelling of problems in science and engineering leads often to partial differential equations in time and space with boundary and initial conditions.The boundary value problems can be written as extremal problems(principle of minimal potential energy), as variational equations (principle of virtual power) or as classical boundary value problems.There are connections concerning existence and uniqueness results between these formulations, which will be investigated using the powerful tools of functional analysis.The first part of the lecture is devoted to the analysis of linear elliptic boundary value problems given in a variational form.The second part deals with the numerical approximation of the solutions of the variational problems.Galerkin methods as FEM and BEM are the main tools. The h-version will be discussed, and an error analysis will be done.Examples, especially from the elasticity theory, demonstrate the methods.
The shortest path problem in which the \((s,t)\)-paths \(P\) of a given digraph \(G =(V,E)\) are compared with respect to the sum of their edge costs is one of the best known problems in combinatorial optimization. The paper is concerned with a number of variations of this problem having different objective functions like bottleneck, balanced, minimum deviation, algebraic sum, \(k\)-sum and \(k\)-max objectives, \((k_1, k_2)-max, (k_1, k_2)\)-balanced and several types of trimmed-mean objectives. We give a survey on existing algorithms and propose a general model for those problems not yet treated in literature. The latter is based on the solution of resource constrained shortest path problems with equality constraints which can be solved in pseudo-polynomial time if the given graph is acyclic and the number of resources is fixed. In our setting, however, these problems can be solved in strongly polynomial time. Combining this with known results on \(k\)-sum and \(k\)-max optimization for general combinatorial problems, we obtain strongly polynomial algorithms for a variety of path problems on acyclic and general digraphs.
Monte Carlo simulation is one of the commonly used methods for risk estimation on financial markets, especially for option portfolios, where any analytical approximation is usually too inaccurate. However, the usually high computational effort for complex portfolios with a large number of underlying assets motivates the application of variance reduction procedures. Variance reduction for estimating the probability of high portfolio losses has been extensively studied by Glasserman et al. A great variance reduction is achieved by applying an exponential twisting importance sampling algorithm together with stratification. The popular and much faster Delta-Gamma approximation replaces the portfolio loss function in order to guide the choice of the importance sampling density and it plays the role of the stratification variable. The main disadvantage of the proposed algorithm is that it is derived only in the case of Gaussian and some heavy-tailed changes in risk factors.
Hence, our main goal is to keep the main advantage of the Monte Carlo simulation, namely its ability to perform a simulation under alternative assumptions on the distribution of the changes in risk factors, also in the variance reduction algorithms. Step by step, we construct new variance reduction techniques for estimating the probability of high portfolio losses. They are based on the idea of the Cross-Entropy importance sampling procedure. More precisely, the importance sampling density is chosen as the closest one to the optimal importance sampling density (zero variance estimator) out of some parametric family of densities with respect to Kullback - Leibler cross-entropy. Our algorithms are based on the special choices of the parametric family and can now use any approximation of the portfolio loss function. A special stratification is developed, so that any approximation of the portfolio loss function under any assumption of the distribution of the risk factors can be used. The constructed algorithms can easily be applied for any distribution of risk factors, no matter if light- or heavy-tailed. The numerical study exhibits a greater variance reduction than of the algorithm from Glasserman et al. The use of a better approximation may improve the performance of our algorithms significantly, as it is shown in the numerical study.
The literature on the estimation of the popular market risk measures, namely VaR and CVaR, often refers to the algorithms for estimating the probability of high portfolio losses, describing the corresponding transition process only briefly. Hence, we give a consecutive discussion of this problem. Results necessary to construct confidence intervals for both measures under the mentioned variance reduction procedures are also given.
Value Preserving Strategies and a General Framework for Local Approaches to Optimal Portfolios
(1999)
We present some new general results on the existence and form of value preserving portfolio strategies in a general semimartingale setting. The concept of value preservation will be derived via a mean-variance argument. It will also be embedded into a framework for local approaches to the problem of portfolio optimisation.
In this work two main approaches for the evaluation of credit derivatives are analyzed: the copula based approach and the Markov Chain based approach. This work gives the opportunity to use the advantages and avoid disadvantages of both approaches. For example, modeling of contagion effects, i.e. modeling dependencies between counterparty defaults, is complicated under the copula approach. One remedy is to use Markov Chain, where it can be done directly. The work consists of five chapters. The first chapter of this work extends the model for the pricing of CDS contracts presented in the paper by Kraft and Steffensen (2007). In the widely used models for CDS pricing it is assumed that only borrower can default. In our model we assume that each of the counterparties involved in the contract may default. Calculated contract prices are compared with those calculated under usual assumptions. All results are summarized in the form of numerical examples and plots. In the second chapter the copula and its main properties are described. The methods of constructing copulas as well as most common copulas families and its properties are introduced. In the third chapter the method of constructing a copula for the existing Markov Chain is introduced. The cases with two and three counterparties are considered. Necessary relations between the transition intensities are derived to directly find some copula functions. The formulae for default dependencies like Spearman's rho and Kendall's tau for defined copulas are derived. Several numerical examples are presented in which the copulas are built for given Markov Chains. The fourth chapter deals with the approximation of copulas if for a given Markov Chain a copula cannot be provided explicitly. The fifth chapter concludes this thesis.
This thesis deals with risk measures based on utility functions and time consistency of dynamic risk measures. It is therefore aimed at readers interested in both, the theory of static and dynamic financial risk measures in the sense of Artzner, Delbaen, Eber and Heath [7], [8] and the theory of preferences in the tradition of von Neumann and Morgenstern [134].
A main contribution of this thesis is the introduction of optimal expected utility (OEU) risk measures as a new class of utility-based risk measures. We introduce OEU, investigate its main properties, and its applicability to risk measurement and put it in perspective to alternative risk measures and notions of certainty equivalents. To the best of our knowledge, OEU is the only existing utility-based risk measure that is (non-trivial and) coherent if the utility function u has constant relative risk aversion. We present several different risk measures that can be derived with special choices of u and illustrate that OEU reacts in a more sensitive way to slight changes of the probability of a financial loss than value at risk (V@R) and average value at risk.
Further, we propose implied risk aversion as a coherent rating methodology for retail structured products (RSPs). Implied risk aversion is based on optimal expected utility risk measures and, in contrast to standard V@R-based ratings, takes into account both the upside potential and the downside risks of such products. In addition, implied risk aversion is easily interpreted in terms of an individual investor's risk aversion: A product is attractive (unattractive) for an investor if its implied risk aversion is higher (lower) than his individual risk aversion. We illustrate this approach in a case study with more than 15,000 warrants on DAX ® and find that implied risk aversion is able to identify favorable products; in particular, implied risk aversion is not necessarily increasing with respect to the strikes of call warrants.
Another main focus of this thesis is on consistency of dynamic risk measures. To this end, we study risk measures on the space of distributions, discuss concavity on the level of distributions and slightly generalize Weber's [137] findings on the relation of time consistent dynamic risk measures to static risk measures to the case of dynamic risk measures with time-dependent parameters. Finally, this thesis investigates how recursively composed dynamic risk measures in discrete time, which are time consistent by construction, can be related to corresponding dynamic risk measures in continuous time. We present different approaches to establish this link and outline the theoretical basis and the practical benefits of this relation. The thesis concludes with a numerical implementation of this theory.
This thesis deals with the relationship between no-arbitrage and (strictly) consistent price processes for a financial market with proportional transaction costs
in a discrete time model. The exact mathematical statement behind this relationship is formulated in the so-called Fundamental Theorem of Asset Pricing (FTAP). Among the many proofs of the FTAP without transaction costs there
is also an economic intuitive utility-based approach. It relies on the economic
intuitive fact that the investor can maximize his expected utility from terminal
wealth. This approach is rather constructive since the equivalent martingale measure is then given by the marginal utility evaluated at the optimal terminal payoff.
However, in the presence of proportional transaction costs such a utility-based approach for the existence of consistent price processes is missing in the literature. So far, rather deep methods from functional analysis or from the theory of random sets have been used to show the FTAP under proportional transaction costs.
For the sake of existence of a utility-maximizing payoff we first concentrate on a generic single-period model with only one risky asset. The marignal utility evaluated at the optimal terminal payoff yields the first component of a
consistent price process. The second component is given by the bid-ask prices
depending on the investors optimal action. Even more is true: nearby this consistent price process there are many strictly consistent price processes. Their exact structure allows us to apply this utility-maximizing argument in a multi-period model. In a backwards induction we adapt the given bid-ask prices in such a way so that the strictly consistent price processes found from maximizing utility can be extended to terminal time. In addition possible arbitrage opportunities of the 2nd kind vanish which can present for the original bid-ask process. The notion of arbitrage opportunities of the 2nd kind has been so
far investigated only in models with strict costs in every state. In our model
transaction costs need not be present in every state.
For a model with finitely many risky assets a similar idea is applicable. However, in the single-period case we need to develop new methods compared
to the single-period case with only one risky asset. There are mainly two reasons
for that. Firstly, it is not at all obvious how to get a consistent price process
from the utility-maximizing payoff, since the consistent price process has to be
found for all assets simultaneously. Secondly, we need to show directly that the
so-called vector space property for null payoffs implies the robust no-arbitrage condition. Once this step is accomplished we can à priori use prices with a
smaller spread than the original ones so that the consistent price process found
from the utility-maximizing payoff is strictly consistent for the original prices.
To make the results applicable for the multi-period case we assume that the prices are given by compact and convex random sets. Then the multi-period case is similar to the case with only one risky asset but more demanding with regard to technical questions.
We present new results on standard basis computations of a 0-dimensional ideal I in a power series ring or in the localization of a polynomial ring over a computable field K. We prove the semicontinuity of the “highest corner” in a family of ideals, parametrized by the spectrum of a Noetherian domain A. This semicontinuity is used to design a new modular algorithm for computing a standard basis of I if K is the quotient field of A. It uses the computation over the residue field of a “good” prime ideal of A to truncate high order terms in the subsequent computation over K. We prove that almost all prime ideals are good, so a random choice is very likely to be good, and whether it is good is detected a posteriori by the algorithm. The algorithm yields a significant speed advantage over the non-modular version and works for arbitrary Noetherian domains. The most important special cases are perhaps A = ℤ and A = k[t], k any field and t a set of parameters. Besides its generality, the method differs substantially from previously known modular algorithms for A = ℤ, since it does not manipulate the coefficients. It is also usually faster and can be combined with other modular methods for computations in local rings. The algorithm is implemented in the computer algebra system SINGULAR and we present several examples illustrating its power.
Lithium-ion batteries are broadly used nowadays in all kinds of portable electronics, such as laptops, cell phones, tablets, e-book readers, digital cameras, etc. They are preferred to other types of rechargeable batteries due to their superior characteristics, such as light weight and high energy density, no memory effect, and a big number of charge/discharge cycles. The high demand and applicability of Li-ion batteries naturally give rise to the unceasing necessity of developing better batteries in terms of performance and lifetime. The aim of the mathematical modelling of Li-ion batteries is to help engineers test different battery configurations and electrode materials faster and cheaper. Lithium-ion batteries are multiscale systems. A typical Li-ion battery consists of multiple connected electrochemical battery cells. Each cell has two electrodes - anode and cathode, as well as a separator between them that prevents a short circuit.
Both electrodes have porous structure composed of two phases - solid and electrolyte. We call macroscale the lengthscale of the whole electrode and microscale - the lengthscale at which we can distinguish the complex porous structure of the electrodes. We start from a Li-ion battery model derived on the microscale. The model is based on nonlinear diffusion type of equations for the transport of Lithium ions and charges in the electrolyte and in the active material. Electrochemical reactions on the solid-electrolyte interface couple the two phases. The interface kinetics is modelled by the highly nonlinear Butler-Volmer interface conditions. Direct numerical simulations with standard methods, such as the Finite Element Method or Finite Volume Method, lead to ill-conditioned problems with a huge number of degrees of freedom which are difficult to solve. Therefore, the aim of this work is to derive upscaled models on the lengthscale of the whole electrode so that we do not have to resolve all the small-scale features of the porous microstructure thus reducing the computational time and cost. We do this by applying two different upscaling techniques - the Asymptotic Homogenization Method and the Multiscale Finite Element Method (MsFEM). We consider the electrolyte and the solid as two self-complementary perforated domains and we exploit this idea with both upscaling methods. The first method is restricted only to periodic media and periodically oscillating solutions while the second method can be applied to randomly oscillating solutions and is based on the Finite Element Method framework. We apply the Asymptotic Homogenization Method to derive a coupled macro-micro upscaled model under the assumption of periodic electrode microstructure. A crucial step in the homogenization procedure is the upscaling of the Butler-Volmer interface conditions. We rigorously determine the asymptotic order of the interface exchange current densities and we perform a comprehensive numerical study in order to validate the derived homogenized Li-ion battery model. In order to upscale the microscale battery problem in the case of random electrode microstructure we apply the MsFEM, extended to problems in perforated domains with Neumann boundary conditions on the holes. We conduct a detailed numerical investigation of the proposed algorithm and we show numerical convergence of the method that we design. We also apply the developed technique to a simplified two-dimensional Li-ion battery problem and we show numerical convergence of the solution obtained with the MsFEM to the reference microscale one.
Universal Shortest Paths
(2010)
We introduce the universal shortest path problem (Univ-SPP) which generalizes both - classical and new - shortest path problems. Starting with the definition of the even more general universal combinatorial optimization problem (Univ-COP), we show that a variety of objective functions for general combinatorial problems can be modeled if all feasible solutions have the same cardinality. Since this assumption is, in general, not satisfied when considering shortest paths, we give two alternative definitions for Univ-SPP, one based on a sequence of cardinality contrained subproblems, the other using an auxiliary construction to establish uniform length for all paths between source and sink. Both alternatives are shown to be (strongly) NP-hard and they can be formulated as quadratic integer or mixed integer linear programs. On graphs with specific assumptions on edge costs and path lengths, the second version of Univ-SPP can be solved as classical sum shortest path problem.
Universal Algebra
(2004)
An asymptotic preserving numerical scheme (with respect to diffusion scalings) for a linear transport equation is investigated. The scheme is adopted from a class of recently developped schemes. Stability is proven uniformly in the mean free path under a CFL type condition turning into a parabolic CFL condition in the diffusion limit.
Dealing with uncertain structures or data has lately been getting much attention in discrete optimization. This thesis addresses two different areas in discrete optimization: Connectivity and covering.
When discussing uncertain structures in networks it is often of interest to determine how many vertices or edges may fail in order for the network to stay connected.
Connectivity is a broad, well studied topic in graph theory. One of the most important results in this area is Menger's Theorem which states that the minimum number of vertices needed to separate two non-adjacent vertices equals the maximum number of internally vertex-disjoint paths between these vertices. Here, we discuss mixed forms of connectivity in which both vertices and edges are removed from a graph at the same time. The Beineke Harary Conjecture states that for any two distinct vertices that can be separated with k vertices and l edges but not with k-1 vertices and l edges or k vertices and l-1 edges there exist k+l edge-disjoint paths between them of which k+1 are internally vertex-disjoint. In contrast to Menger's Theorem, the existence of the paths is not sufficient for the connectivity statement to hold. Our main contribution is the proof of the Beineke Harary Conjecture for the case that l equals 2.
We also consider different problems from the area of facility location and covering. We regard problems in which we are given sets of locations and regions, where each region has an assigned number of clients. We are now looking for an allocation of suppliers into the locations, such that each client is served by some supplier. The notable difference to other covering problems is that we assume that each supplier may only serve a fixed number of clients which is not part of the input. We discuss the complexity and solution approaches of three such problems which vary in the way the clients are assigned to the suppliers.
In diesem Projekt soll die Bildung von Wirbeln bei der Strömung eines Gases um eine Ecke numerisch untersucht werden. Dabei sollen verschiedene numerische Verfahren getestet und die Ergebnisse mit Versuchsdaten verglichen werden. Ferner soll untersucht werden, wie gut sich diese Verfahren vektorisieren lassen, da komplizierte zweidimensionale und selbst einfache dreidimensionale Probleme der Strömungsdynamik auf den heute üblichen Universalrechnern nicht mit vertretbarem Zeitaufwand zu lösen sind. Die numerischen Berechnungen werden auf der CYBER 205 in Karlsruhe durchgeführt.
In diesem Projekt soll die Bildung von Wirbeln bei der Strömung eines Gases um eine Ecke numerisch untersucht werden. Dabei sollen verschiedene numerische Verfahren getestet und die Ergebnisse mit Versuchsdaten verglichen werden. Ferner soll untersucht werden, wie gut sich diese Verfahren vektorisieren lassen, da kompliziertere zweidimensionale und selbst einfache dreidimensionale Probleme der Strömungsdynamik auf den heute üblichen Universalrechnern nicht mit vertretbarem Zeitaufwand zu lösen sind, besonders, wenn, wie an der Universität Kaiserslautern, nur eine relativ langsame Anlage (Siemens 7551/7561) zur Verfügung steht. Die numerischen Rechnungen werden auf der CYBER 205 in Karlsruhe durchgeführt.
In this thesis we address two instances of duality in commutative algebra.
In the first part, we consider value semigroups of non irreducible singular algebraic curves
and their fractional ideals. These are submonoids of Z^n closed under minima, with a conductor and which fulfill special compatibility properties on their elements. Subsets of Z^n
fulfilling these three conditions are known in the literature as good semigroups and their ideals, and their class strictly contains the class of value semigroup ideals. We examine
good semigroups both independently and in relation with their algebraic counterpart. In the combinatoric setting, we define the concept of good system of generators, and we
show that minimal good systems of generators are unique. In relation with the algebra side, we give an intrinsic definition of canonical semigroup ideals, which yields a duality
on good semigroup ideals. We prove that this semigroup duality is compatible with the Cohen-Macaulay duality under taking values. Finally, using the duality on good semigroup ideals, we show a symmetry of the Poincaré series of good semigroups with special properties.
In the second part, we treat Macaulay’s inverse system, a one-to-one correspondence
which is a particular case of Matlis duality and an effective method to construct Artinian k-algebras with chosen socle type. Recently, Elias and Rossi gave the structure of the inverse system of positive dimensional Gorenstein k-algebras. We extend their result by establishing a one-to-one correspondence between positive dimensional level k-algebras and certain submodules of the divided power ring. We give several examples to illustrate
our result.
The following two norms for holomorphic functions \(F\), defined on the right complex half-plane \(\{z \in C:\Re(z)\gt 0\}\) with values in a Banach space \(X\), are equivalent:
\[\begin{eqnarray*} \lVert F \rVert _{H_p(C_+)} &=& \sup_{a\gt0}\left( \int_{-\infty}^\infty \lVert F(a+ib) \rVert ^p \ db \right)^{1/p}
\mbox{, and} \\ \lVert F \rVert_{H_p(\Sigma_{\pi/2})} &=& \sup_{\lvert \theta \lvert \lt \pi/2}\left( \int_0^\infty \left \lVert F(re^{i \theta}) \right \rVert ^p\ dr \right)^{1/p}.\end{eqnarray*}\] As a consequence, we derive a description of boundary values ofsectorial holomorphic functions, and a theorem of Paley-Wiener typefor sectorial holomorphic functions.
We prove a general monotonicity result about Nash flows in directed networks and use it for the design of truthful mechanisms in the setting where each edge of the network is controlled by a different selfish agent, who incurs costs when her edge is used. The costs for each edge are assumed to be linear in the load on the edge. To compensate for these costs, the agents impose tolls for the usage of edges. When nonatomic selfish network users choose their paths through the network independently and each user tries to minimize a weighted sum of her latency and the toll she has to pay to the edges, a Nash flow is obtained. Our monotonicity result implies that the load on an edge in this setting can not increase when the toll on the edge is increased, so the assignment of load to the edges by a Nash flow yields a monotone algorithm. By a well-known result, the monotonicity of the algorithm then allows us to design truthful mechanisms based on the load assignment by Nash flows. Moreover, we consider a mechanism design setting with two-parameter agents, which is a generalization of the case of one-parameter agents considered in a seminal paper of Archer and Tardos. While the private data of an agent in the one-parameter case consists of a single nonnegative real number specifying the agent's cost per unit of load assigned to her, the private data of a two-parameter agent consists of a pair of nonnegative real numbers, where the first one specifies the cost of the agent per unit load as in the one-parameter case, and the second one specifies a fixed cost, which the agent incurs independently of the load assignment. We give a complete characterization of the set of output functions that can be turned into truthful mechanisms for two-parameter agents. Namely, we prove that an output function for the two-parameter setting can be turned into a truthful mechanism if and only if the load assigned to every agent is nonincreasing in the agent's bid for her per unit cost and, for almost all fixed bids for the agent's per unit cost, the load assigned to her is independent of the agent's bid for her fixed cost. When the load assigned to an agent is continuous in the agent's bid for her per unit cost, it must be completely independent of the agent's bid for her fixed cost. These results motivate our choice of linear cost functions without fixed costs for the edges in the selfish routing setting, but the results also seem to be interesting in the context of algorithmic mechanism design themselves.
A main result of this thesis is a conceptual proof of the fact that the weighted number of tropical curves of given degree and genus, which pass through the right number of general points in the plane (resp., which pass through general points in R^r and represent a given point in the moduli space of genus g curves) is independent of the choices of points. Another main result is a new correspondence theorem between plane tropical cycles and plane elliptic algebraic curves.
This thesis is devoted to two main topics (accordingly, there are two chapters): In the first chapter, we establish a tropical intersection theory with analogue notions and tools as its algebro-geometric counterpart. This includes tropical cycles, rational functions, intersection products of Cartier divisors and cycles, morphisms, their functors and the projection formula, rational equivalence. The most important features of this theory are the following: - It unifies and simplifies many of the existing results of tropical enumerative geometry, which often contained involved ad-hoc computations. - It is indispensable to formulate and solve further tropical enumerative problems. - It shows deep relations to the intersection theory of toric varieties and connected fields. - The relationship between tropical and classical Gromov-Witten invariants found by Mikhalkin is made plausible from inside tropical geometry. - It is interesting on its own as a subfield of convex geometry. In the second chapter, we study tropical gravitational descendants (i.e. Gromov-Witten invariants with incidence and "Psi-class" factors) and show that many concepts of the classical Gromov-Witten theory such as the famous WDVV equations can be carried over to the tropical world. We use this to extend Mikhalkin's results to a certain class of gravitational descendants, i.e. we show that many of the classical gravitational descendants of P^2 and P^1 x P^1 can be computed by counting tropical curves satisfying certain incidence conditions and with prescribed valences of their vertices. Moreover, the presented theory is not restricted to plane curves and therefore provides an important tool to derive similar results in higher dimensions. A more detailed chapter synopsis can be found at the beginning of each individual chapter.
Tropical intersection theory
(2010)
This thesis consists of five chapters: Chapter 1 contains the basics of the theory and is essential for the rest of the thesis. Chapters 2-5 are to a large extent independent of each other and can be read separately. - Chapter 1: Foundations of tropical intersection theory In this first chapter we set up the foundations of a tropical intersection theory covering many concepts and tools of its counterpart in algebraic geometry such as affine tropical cycles, Cartier divisors, morphisms of tropical cycles, pull-backs of Cartier divisors, push-forwards of cycles and an intersection product of Cartier divisors and cycles. Afterwards, we generalize these concepts to abstract tropical cycles and introduce a concept of rational equivalence. Finally, we set up an intersection product of cycles and prove that every cycle is rationally equivalent to some affine cycle in the special case that our ambient cycle is R^n. We use this result to show that rational and numerical equivalence agree in this case and prove a tropical Bézout's theorem. - Chapter 2: Tropical cycles with real slopes and numerical equivalence In this chapter we generalize our definitions of tropical cycles to polyhedral complexes with non-rational slopes. We use this new definition to show that if our ambient cycle is a fan then every subcycle is numerically equivalent to some affine cycle. Finally, we restrict ourselves to cycles in R^n that are "generic" in some sense and study the concept of numerical equivalence in more detail. - Chapter 3: Tropical intersection products on smooth varieties We define an intersection product of tropical cycles on tropical linear spaces L^n_k and on other, related fans. Then, we use this result to obtain an intersection product of cycles on any "smooth" tropical variety. Finally, we use the intersection product to introduce a concept of pull-backs of cycles along morphisms of smooth tropical varieties and prove that this pull-back has all expected properties. - Chapter 4: Weil and Cartier divisors under tropical modifications First, we introduce "modifications" and "contractions" and study their basic properties. After that, we prove that under some further assumptions a one-to-one correspondence of Weil and Cartier divisors is preserved by modifications. In particular we can prove that on any smooth tropical variety we have a one-to-one correspondence of Weil and Cartier divisors. - Chapter 5: Chern classes of tropical vector bundles We give definitions of tropical vector bundles and rational sections of tropical vector bundles. We use these rational sections to define the Chern classes of such a tropical vector bundle. Moreover, we prove that these Chern classes have all expected properties. Finally, we classify all tropical vector bundles on an elliptic curve up to isomorphisms.
This thesis is devoted to furthering the tropical intersection theory as well as to applying the
developed theory to gain new insights about tropical moduli spaces.
We use piecewise polynomials to define tropical cocycles that generalise the notion of tropical Cartier divisors to higher codimensions, introduce an intersection product of cocycles with tropical cycles and use the connection to toric geometry to prove a Poincaré duality for certain cases. Our
main application of this Poincaré duality is the construction of intersection-theoretic fibres under a
large class of tropical morphisms.
We construct an intersection product of cycles on matroid varieties which are a natural
generalisation of tropicalisations of classical linear spaces and the local blocks of smooth tropical
varieties. The key ingredient is the ability to express a matroid variety contained in another matroid variety by a piecewise polynomial that is given in terms of the rank functions of the corresponding
matroids. In particular, this enables us to intersect cycles on the moduli spaces of n-marked abstract
rational curves. We also construct a pull-back of cycles along morphisms of smooth varieties, relate
pull-backs to tropical modifications and show that every cycle on a matroid variety is rationally
equivalent to its recession cycle and can be cut out by a cocycle.
Finally, we define families of smooth rational tropical curves over smooth varieties and construct a tropical fibre product in order to show that every morphism of a smooth variety to the moduli space of abstract rational tropical curves induces a family of curves over the domain of the morphism.
This leads to an alternative, inductive way of constructing moduli spaces of rational curves.
Das Ziel dieser Dissertation ist die Entwicklung und Implementation eines Algorithmus zur Berechnung von tropischen Varietäten über allgemeine bewertete Körper. Die Berechnung von tropischen Varietäten über Körper mit trivialer Bewertung ist ein hinreichend gelöstes Problem. Hierfür kombinieren die Autoren Bogart, Jensen, Speyer, Sturmfels und Thomas eindrucksvoll klassische Techniken der Computeralgebra mit konstruktiven Methoden der konvexer Geometrie.
Haben wir allerdings einen Grundkörper mit nicht-trivialer Bewertung, wie zum Beispiel den Körper der \(p\)-adischen Zahlen \(\mathbb{Q}_p\), dann stößt die konventionelle Gröbnerbasentheorie scheinbar an ihre Grenzen. Die zugrundeliegenden Monomordnungen sind nicht geeignet um Problemstellungen zu untersuchen, die von einer nicht-trivialen Bewertung auf den Koeffizienten abhängig sind. Dies führte zu einer Reihe von Arbeiten, welche die gängige Gröbnerbasentheorie modifizieren um die Bewertung des Grundkörpers einzubeziehen.\[\phantom{newline}\]
In dieser Arbeit präsentieren wir einen alternativen Ansatz und zeigen, wie sich die Bewertung mittels einer speziell eingeführten Variable emulieren lässt, so dass eine Modifikation der klassischen Werkzeuge nicht notwendig ist.
Im Rahmen dessen wird Theorie der Standardbasen auf Potenzreihen über einen Koeffizientenring verallgemeinert. Hierbei wird besonders Wert darauf gelegt, dass alle Algorithmen bei polynomialen Eingabedaten mit ihren klassischen Pendants übereinstimmen, sodass für praktische Zwecke auf bereits etablierte Softwaresysteme zurückgegriffen werden kann. Darüber hinaus wird die Konstruktion des Gröbnerfächers sowie die Technik des Gröbnerwalks für leicht inhomogene Ideale eingeführt. Dies ist notwendig, da bei der Einführung der neuen Variable die Homogenität des Ausgangsideal gebrochen wird.\[\phantom{newline}\]
Alle Algorithmen wurden in Singular implementiert und sind als Teil der offiziellen Distribution erhältlich. Es ist die erste Implementation, welches in der Lage ist tropische Varietäten mit \(p\)-adischer Bewertung auszurechnen. Im Rahmen der Arbeit entstand ebenfalls ein Singular Paket für konvexe Geometrie, sowie eine Schnittstelle zu Polymake.
In the paper we discuss the transition from kinetic theory to macroscopic fluid equations, where the macroscopic equations are defined as aymptotic limits of a kinetic equation. This relation can be used to derive computationally efficient domain decomposition schemes for the simulaion of rarefied gas flows close to the continuum limit. Moreover, we present some basic ideas for the derivation of kinetic induced numerical schemes for macroscopic equations, namely kinetic schemes for general conservation laws as well as Lattice-Boltzmann methods for the incompressible Navier-Stokes equations.
Due to the increasing number of natural or man-made disasters, the application of operations research methods in evacuation planning has seen a rising interest in the research community. From the beginning, evacuation planning has been highly focused on car-based evacuation. Recently, also the evacuation of transit depended evacuees with the help of buses has been considered.
In this case study, we apply two such models and solution algorithms to evacuate a core part of the metropolitan capital city Kathmandu of Nepal as a hypothetical endangered region, where a large part of population is transit dependent. We discuss the computational results for evacuation time under a broad range of possible scenarios, and derive planning suggestions for practitioners.
The Train Marshalling Problem consists of rearranging an incoming train in a marshalling yard in such a way that cars with the same destinations appear consecutively in the final train and the number of needed sorting tracks is minimized. Besides an initial roll-in operation, just one pull-out operation is allowed. This problem was introduced by Dahlhaus et al. who also showed that the problem is NP-complete. In this paper, we provide a new lower bound on the optimal objective value by partitioning an appropriate interval graph. Furthermore, we consider the corresponding online problem, for which we provide upper and lower bounds on the competitiveness and a corresponding optimal deterministic online algorithm. We provide an experimental evaluation of our lower bound and algorithm which shows the practical tightness of the results.
The use of trading stops is a common practice in financial markets for a variety of reasons: it provides a simple way to control losses on a given trade, while also ensuring that profit-taking is not deferred indefinitely; and it allows opportunities to consider reallocating resources to other investments. In this thesis, it is explained why the use of stops may be desirable in certain cases.
This is done by proposing a simple objective to be optimized. Some simple and commonly-used rules for the placing and use of stops are investigated; consisting of fixed or moving barriers, with fixed transaction costs. It is shown how to identify optimal levels at which to set stops, and the performances of different rules and strategies are compared. Thereby, uncertainty and altering of the drift parameter of the investment are incorporated.
Toying with Jordan matrices
(1996)
Over the past 2 decades, there has been much progress on the classification of symplectic linear quotient singularities V/G admitting a symplectic (equivalently, crepant) resolution of singularities. The classification is almost complete but there is an infinite series of groups in dimension 4—the symplectically primitive but complex imprimitive groups—and 10 exceptional groups up to dimension 10, for which it is still open. In this paper, we treat the remaining infinite series and prove that for all but possibly 39 cases there is no symplectic resolution. We thereby reduce the classification problem to finitely many open cases. We furthermore prove non-existence of a symplectic resolution for one exceptional group, leaving 39+9=48 open cases in total. We do not expect any of the remaining cases to admit a symplectic resolution.
Introducing parallelism and exploring its use is still a fundamental challenge for the computer algebra community. In high-performance numerical simulation, on the other hand, transparent environments for distributed computing which follow the principle of separating coordination and computation have been a success story for many years. In this paper, we explore the potential of using this principle in the context of computer algebra. More precisely, we combine two well-established systems: The mathematics we are interested in is implemented in the computer algebra system SINGULAR, whose focus is on polynomial computations, while the coordination is left to the workflow management system GPI-Space, which relies on Petri nets as its mathematical modeling language and has been successfully used for coordinating the parallel execution (autoparallelization) of academic codes as well as for commercial software in application areas such as seismic data processing. The result of our efforts is a major step towards a framework for massively parallel computations in the application areas of SINGULAR, specifically in commutative algebra and algebraic geometry. As a first test case for this framework, we have modeled and implemented a hybrid smoothness test for algebraic varieties which combines ideas from Hironaka’s celebrated desingularization proof with the classical Jacobian criterion. Applying our implementation to two examples originating from current research in algebraic geometry, one of which cannot be handled by other means, we illustrate the behavior of the smoothness test within our framework and investigate how the computations scale up to 256 cores.
Topologie II
(1995)
Several topological necessary conditions of smooth stabilization in the large have been obtained. In particular, if a smooth single-input nonlinear system is smoothly stabilizable in the large at some point of a connected component of equilibria set, then the connected component is to be an unknoted, unbounded curve.
The purpose of Exploration in Oil Industry is to "discover" an oil-containing geological formation from exploration data. In the context of this PhD project this oil-containing geological formation plays the role of a geometrical object, which may have any shape. The exploration data may be viewed as a "cloud of points", that is a finite set of points, related to the geological formation surveyed in the exploration experiment. Extensions of topological methodologies, such as homology, to point clouds are helpful in studying them qualitatively and capable of resolving the underlying structure of a data set. Estimation of topological invariants of the data space is a good basis for asserting the global features of the simplicial model of the data. For instance the basic statistical idea, clustering, are correspond to dimension of the zero homology group of the data. A statistics of Betti numbers can provide us with another connectivity information. In this work represented a method for topological feature analysis of exploration data on the base of so called persistent homology. Loosely, this is the homology of a growing space that captures the lifetimes of topological attributes in a multiset of intervals called a barcode. Constructions from algebraic topology empowers to transform the data, to distillate it into some persistent features, and to understand then how it is organized on a large scale or at least to obtain a low-dimensional information which can point to areas of interest. The algorithm for computing of the persistent Betti numbers via barcode is realized in the computer algebra system "Singular" in the scope of the work.
This paper presents a wavelet analysis of temporal and spatial variations of the Earth's gravitational potential based on tensor product wavelets. The time--space wavelet concept is realized by combining Legendre wavelets for the time domain and spherical wavelets for the space domain. In consequence, a multiresolution analysis for both, temporal and spatial resolution, is formulated within a unified concept. The method is then numerically realized by using first synthetically generated data and, finally, several real data sets.
In this paper a known orthonormal system of time- and space-dependent functions, that were derived out of the Cauchy-Navier equation for elastodynamic phenomena, is used to construct reproducing kernel Hilbert spaces. After choosing one of the spaces the corresponding kernel is used to define a function system that serves as a basis for a spline space. We show that under certain conditions there exists a unique interpolating or approximating, respectively, spline in this space with respect to given samples of an unknown function. The name "spline" here refers to its property of minimising a norm among all interpolating functions. Moreover, a convergence theorem and an error estimate relative to the point grid density are derived. As numerical example we investigate the propagation of seismic waves.
Constructing accurate earth models from seismic data is a challenging task. Traditional methods rely on ray based approximations of the wave equation and reach their limit in geologically complex areas. Full waveform inversion (FWI) on the other side seeks to minimize the misfit between modeled and observed data without such approximation.
While superior in accuracy, FWI uses a gradient based iterative scheme that makes it also very computationally expensive. In this thesis we analyse and test an Alternating Direction Implicit (ADI) scheme in order to reduce the costs of the two dimensional time domain algorithm for solving the acoustic wave equation. The ADI scheme can be seen as an intermediate between explicit and implicit finite difference modeling schemes. Compared to full implicit schemes the ADI scheme only requires the solution of much smaller matrices and is thus less computationally demanding. Using ADI we can handle coarser discretization compared to an explicit method. Although order of convergence and CFL conditions for the examined explicit method and ADI scheme are comparable, we observe that the ADI scheme is less prone to dispersion. Furhter, our algorithm is efficiently parallelized with vectorization and threading techniques. In a numerical comparison, we can demonstrate a runtime advantage of the ADI scheme over an explicit method of the same accuracy.
With the modeling in place, we test and compare several inverse schemes in the second part of the thesis. With the goal of avoiding local minima and improving speed of convergence, we use different minimization functions and hierarchical approaches. In several tests, we demonstrate superior results of the L1 norm compared to the L2 norm – especially in the presence of noise. Furthermore we show positive effects for applying three different multiscale approaches to the inverse problem. These methods focus on low frequency, early recording, or far offset during early iterations of the minimization and then proceed iteratively towards the full problem. We achieve best results with the frequency based multiscale scheme, for which we also provide a heuristical method of choosing iteratively increasing frequency bands.
Finally, we demonstrate the effectiveness of the different methods first on the Marmousi model and then on an extract of the 2004 BP model, where we are able to recover both high contrast top salt structures and lower contrast inclusions accurately.
In this thesis, we deal with the worst-case portfolio optimization problem occuring in discrete-time markets.
First, we consider the discrete-time market model in the presence of crash threats. We construct the discrete worst-case optimal portfolio strategy by the indifference principle in the case of the logarithmic utility. After that we extend this problem to general utility functions and derive the discrete worst-case optimal portfolio processes, which are characterized by a dynamic programming equation. Furthermore, the convergence of the discrete worst-case optimal portfolio processes are investigated when we deal with the explicit utility functions.
In order to further study the relation of the worst-case optimal value function in discrete-time models to continuous-time models we establish the finite-difference approach. By deriving the discrete HJB equation we verify the worst-case optimal value function in discrete-time models, which satisfies a system of dynamic programming inequalities. With increasing degree of fineness of the time discretization, the convergence of the worst-case value function in discrete-time models to that in continuous-time models are proved by using a viscosity solution method.
The Trippstadt Problem
(1984)
Close to Kaiserslautern is the town of Trippstadt, which, together with five other small towns forms a local administration unit (Verbandsgemeinde) called Kaiserslautern-Süd. Trippstadt has its own beautiful public swimming pool, which causes problems though; the cost for the upkeep of the pool is higher than the income and thus has to be divided among the towns belonging to the Verbandsgemeinde. Because of this problem the administration wanted to find out which fraction of the total number of pool visitors came from the different towns. They planned to ask each pool guest where he came from. They did this for only three days though because the waiting lines at the cashiers became unbearably long and they could see that because of this the total number of guests would decrease. Then they wondered how to find a better method to get the same data and that was when I was asked to help with the solution of the problem.
In this paper we consider the stochastic primitive equation for geophysical flows subject to transport noise and turbulent pressure. Admitting very rough noise terms, the global existence and uniqueness of solutions to this stochastic partial differential equation are proven using stochastic maximal
-regularity, the theory of critical spaces for stochastic evolution equations, and global a priori bounds. Compared to other results in this direction, we do not need any smallness assumption on the transport noise which acts directly on the velocity field and we also allow rougher noise terms. The adaptation to Stratonovich type noise and, more generally, to variable viscosity and/or conductivity are discussed as well.
This paper is concerned with numerical algorithms for the bipolar quantum drift diffusion model. For the thermal equilibrium case a quasi-gradient method minimizing the energy functional is introduced and strong convergence is proven. The computation of current - voltage characteristics is performed by means of an extended emph{Gummel - iteration}. It is shown that the involved fixed point mapping is a contraction for small applied voltages. In this case the model equations are uniquely solvable and convergence of the proposed iteration scheme follows. Numerical simulations of a one dimensional resonant tunneling diode are presented. The computed current - voltage characteristics are in good qualitative agreement with experimental measurements. The appearance of negative differential resistances is verified for the first time in a Quantum Drift Diffusion model.
In this dissertation convergence of binomial trees for option pricing is investigated. The focus is on American and European put and call options. For that purpose variations of the binomial tree model are reviewed.
In the first part of the thesis we investigated the convergence behavior of the already known trees from the literature (CRR, RB, Tian and CP) for the European options. The CRR and the RB tree suffer from irregular convergence, so our first aim is to find a way to get the smooth convergence. We first show what causes these oscillations. That will also help us to improve the rate of convergence. As a result we introduce the Tian and the CP tree and we proved that the order of convergence for these trees is \(O \left(\frac{1}{n} \right)\).
Afterwards we introduce the Split tree and explain its properties. We prove the convergence of it and we found an explicit first order error formula. In our setting, the splitting time \(t_{k} = k\Delta t\) is not fixed, i.e. it can be any time between 0 and the maturity time \(T\). This is the main difference compared to the model from the literature. Namely, we show that the good properties of the CRR tree when \(S_{0} = K\) can be preserved even without this condition (which is mainly the case). We achieved the convergence of \(O \left(n^{-\frac{3}{2}} \right)\) and we typically get better results if we split our tree later.
In this work we introduce a new bandlimited spherical wavelet: The Bernstein wavelet. It possesses a couple of interesting properties. To be specific, we are able to construct bandlimited wavelets free of oscillations. The scaling function of this wavelet is investigated with regard to the spherical uncertainty principle, i.e., its localization in the space domain as well as in the momentum domain is calculated and compared to the well-known Shannon scaling function. Surprisingly, they possess the same localization in space although one is highly oscillating whereas the other one shows no oscillatory behavior. Moreover, the Bernstein scaling function turns out to be the first bandlimited scaling function known to the literature whose uncertainty product tends to the minimal value 1.
In this paper we consider a certain class of geodetic linear inverse problems LambdaF=G in a reproducing kernel Hilbert space setting to obtain a bounded generalized inverse operator Lambda. For a numerical realization we assume G to be given at a finite number of discrete points to which we employ a spherical spline interpolation method adapted to the Hilbertspaces. By applying Lambda to the obtained spline interpolant we get an approximation of the solution F. Finally our main task is to show some properties of the approximated solution and to prove convergence results if the data set increases.
The performance of a combustion engine is essentially determined by the charge cycle, i.e. by the inflow of fresh air through the inlet pipe into the cylinder after a combustion cycle. The amount of air, exchanged during this process, depends on many factors, e.g. the number of revolutions per minute, the temperature, the engine and valve geometry. In order to have a tool in designing the engine one is interested in calculating this amount. The proper calculation would involve the solution of three-dimensional hydrodynamical equations governing the gas flow including chemical reactions in a complicated geometry, consisting of the cylinder, valves, inlet and outlet pipe. Since this is clearly too ambitious, we consider a simplified model.
We consider optimal design problems for semiconductor devices which are simulated using the energy transport model. We develop a descent algorithm based on the adjoint calculus and present numerical results for a ballistic diode. Further, we compare the optimal doping profile with results computed on basis of the drift diffusion model. Finally, we exploit the model hierarchy and test the space mapping approach, especially the aggressive space mapping algorithm, for the design problem. This yields a significant reduction of numerical costs and programming effort.
By natural or man-made disasters, the evacuation of a whole region or city may become necessary. Apart from private traffic, the evacuation from collection points to secure shelters outside the endangered region will be realized by a bus fleet made available by emergency relief. The arising Bus Evacuation Problem (BEP) is a vehicle scheduling problem, in which a given number of evacuees needs to be transported from a set of collection points to a set of capacitated shelters, minimizing the total evacuation time, i.e., the time needed until the last person is brought to safety.
In this paper we consider an extended version of the BEP, the Robust Bus Evacuation Problem (RBEP), in which the exact numbers of evacuees are not known, but may stem from a set of probable scenarios. However, after a given reckoning time, this uncertainty is eliminated and planners are given exact figures. The problem is to decide for each bus, if it is better to send it right away -- using uncertain numbers of evacuees -- or to wait until the numbers become known.
We present a mixed-integer linear programming formulation for the RBEP and discuss solution approaches; in particular, we present a tabu search framework for finding heuristic solutions of acceptable quality within short computation time. In computational experiments using both randomly generated instances and the real-world scenario of evacuating the city of Kaiserslautern, we compare our solution approaches.
In the present paper we investigate the Rayleigh-Benard convection in rarefied gases and demonstrate by numerical experiments the transition from purely thermal conduction to a natural convective flow for a large range of Knudsen numbers from 0.02 downto 0.001. We address to the problem how the critical value for the Rayleigh number defined for incompressible vsicous flows may be translated to rarefied gas flows. Moreover, the simulations obtained for a Knudsen number Kn=0.001 and Froude number Fr=1 show a further transition from regular Rayleigh-Benard cells to a pure unsteady behavious with moving vortices.
The thermal equilibrium state of a bipolar, isothermal quantum fluid confined to a bounded domain \(\Omega\subset I\!\!R^d,d=1,2\) or \( d=3\) is the minimizer of the total energy \({\mathcal E}_{\epsilon\lambda}\); \({\mathcal E}_{\epsilon\lambda}\) involves the squares of the scaled Planck's constant \(\epsilon\) and the scaled minimal Debye length \(\lambda\). In applications one frequently has \(\lambda^2\ll 1\). In these cases the zero-space-charge approximation is rigorously justified. As \(\lambda \to 0 \), the particle densities converge to the minimizer of a limiting quantum zero-space-charge functional exactly in those cases where the doping profile satisfies some compatibility conditions. Under natural additional assumptions on the internal energies one gets an differential-algebraic system for the limiting \((\lambda=0)\) particle densities, namely the quantum zero-space-charge model. The analysis of the subsequent limit \(\epsilon \to 0\) exhibits the importance of quantum gaps. The semiclassical zero-space-charge model is, for small \(\epsilon\), a reasonable approximation of the quantum model if and only if the quantum gap vanishes. The simultaneous limit \(\epsilon =\lambda \to 0\) is analyzed.
Consider the primitive equations on ◂+▸R2×(◂,▸z0,z1) with initial data a of the form a=◂+▸a1+a2, where ◂+▸a1∈◂◽.▸BUCσ(◂,▸R2;L1(◂,▸z0,z1)) and ◂+▸a2∈L
∞
σ
(◂,▸R2;L1(◂,▸z0,z1)). These spaces are scaling-invariant and represent the anisotropic character of these equations. It is shown that for a1 arbitrary large and a2 sufficiently small, this set of equations admits a unique strong solution which extends to a global one and is thus strongly globally well posed for these data provided a is periodic in the horizontal variables. The approach presented depends crucially on mapping properties of the hydrostatic Stokes semigroup in the L∞(L1)-setting. It can be seen as the counterpart of the classical iteration schemes for the Navier–Stokes equations, now for the primitive equations in the L∞(L1)-setting.
In this article we prove existence and uniqueness results for solutions to the outer oblique boundary problem for the Poisson equation under very weak assumptions on boundary, coefficients and inhomogeneities. Main tools are the Kelvin transformation and the solution operator for the regular inner problem, provided in [1]. Moreover we prove regularisation results for the weak solutions of both, the inner and the outer problem. We investigate the non-admissible direction for the oblique vector field, state results with stochastic inhomogeneities and provide a Ritz-Galerkinm approximation. The results are applicable to problems from Geomathematics, see e.g. [2] and [3].
In the thesis the author presents a mathematical model which describes the behaviour of the acoustical pressure (sound), produced by a bass loudspeaker. The underlying physical propagation of sound is described by the non--linear isentropic Euler system in a Lagrangian description. This system is expanded via asymptotical analysis up to third order in the displacement of the membrane of the loudspeaker. The differential equations which describe the behaviour of the key note and the first order harmonic are compared to classical results. The boundary conditions, which are derived up to third order, are based on the principle that the small control volume sticks to the boundary and is allowed to move only along it. Using classical results of the theory of elliptic partial differential equations, the author shows that under appropriate conditions on the input data the appropriate mathematical problems admit, by the Fredholm alternative, unique solutions. Moreover, certain regularity results are shown. Further, a novel Wave Based Method is applied to solve appropriate mathematical problems. However, the known theory of the Wave Based Method, which can be found in the literature, so far, allowed to apply WBM only in the cases of convex domains. The author finds the criterion which allows to apply the WBM in the cases of non--convex domains. In the case of 2D problems we represent this criterion as a small proposition. With the aid of this proposition one is able to subdivide arbitrary 2D domains such that the number of subdomains is minimal, WBM may be applied in each subdomain and the geometry is not altered, e.g. via polygonal approximation. Further, the same principles are used in the case of 3D problem. However, the formulation of a similar proposition in cases of 3D problems has still to be done. Next, we show a simple procedure to solve an inhomogeneous Helmholtz equation using WBM. This procedure, however, is rather computationally expensive and can probably be improved. Several examples are also presented. We present the possibility to apply the Wave Based Technique to solve steady--state acoustic problems in the case of an unbounded 3D domain. The main principle of the classical WBM is extended to the case of an external domain. Two numerical examples are also presented. In order to apply the WBM to our problems we subdivide the computational domain into three subdomains. Therefore, on the interfaces certain coupling conditions are defined. The description of the optimization procedure, based on the principles of the shape gradient method and level set method, and the results of the optimization finalize the thesis.
Primary decomposition of an ideal in a polynomial ring over a field belongs to the indispensable theoretical tools in commutative algebra and algebraic geometry. Geometrically it corresponds to the decomposition of an affine variety into irreducible components and is, therefore, also an important geometric concept.The decomposition of a variety into irreducible components is, however, slightly weaker than the full primary decomposition, since the irreducible components correspond only to the minimal primes of the ideal of the variety, which is a radical ideal. The embedded components, although invisible in the decomposition of the variety itself, are, however, responsible for many geometric properties, in particular, if we deform the variety slightly. Therefore, they cannot be neglected and the knowledge of the full primary decomposition is important also in a geometric context.In contrast to the theoretical importance, one can find in mathematical papers only very few concrete examples of non-trivial primary decompositions because carrying out such a decomposition by hand is almost impossible. This experience corresponds to the fact that providing efficient algorithms for primary decomposition of an ideal I ae K[x1; : : : ; xn], K a field, is also a difficult task and still one of the big challenges for computational algebra and computational algebraic geometry.All known algorithms require Gr"obner bases respectively characteristic sets and multivariate polynomial factorization over some (algebraic or transcendental) extension of the given field K. The first practical algorithm for computing the minimal associated primes is based on characteristic sets and the Ritt-Wu process ([R1], [R2], [Wu], [W]), the first practical and general primary decomposition algorithm was given by Gianni, Trager and Zacharias [GTZ]. New ideas from homological algebra were introduced by Eisenbud, Huneke and Vasconcelos in [EHV]. Recently, Shimoyama and Yokoyama [SY] provided a new algorithm, using Gr"obner bases, to obtain the primary decompositon from the given minimal associated primes.In the present paper we present all four approaches together with some improvements and with detailed comparisons, based upon an analysis of 34 examples using the computer algebra system SINGULAR [GPS]. Since primary decomposition is a fairly complicated task, it is, therefore, best explained by dividing it into several subtasks, in particular, while sometimes only one of these subtasks is needed in practice. The paper is organized in such a way that we consider the subtasks separately and present the different approaches of the above-mentioned authors, with several tricks and improvements incorporated. Some of these improvements and the combination of certain steps from the different algorithms are essential for improving the practical performance.