### Refine

#### Year of publication

#### Document Type

- Report (399) (remove)

#### Language

- English (399) (remove)

#### Keywords

- numerical upscaling (7)
- hub location (5)
- Elastoplastizität (4)
- Integer programming (4)
- modelling (4)
- poroelasticity (4)
- Darcy’s law (3)
- Dienstgüte (3)
- Elastic BVP (3)
- Elastoplasticity (3)
- Formalisierung (3)
- Heston model (3)
- Hysterese (3)
- Lagrangian mechanics (3)
- Mathematikunterricht (3)
- Modellierung (3)
- effective heat conductivity (3)
- facility location (3)
- non-Newtonian flow in porous media (3)
- polynomial algorithms (3)
- praxisorientiert (3)
- variational inequalities (3)
- virtual material design (3)
- American options (2)
- Bartlett spectrum (2)
- Chisel (2)
- Elastisches RWP (2)
- Elastoplastisches RWP (2)
- Field-programmable gate array (FPGA) (2)
- HJB equation (2)
- Heuristics (2)
- IMRT planning (2)
- Inverses Problem (2)
- Jiang's model (2)
- Jiang-Modell (2)
- Lineare Algebra (2)
- Logistics (2)
- MAC type grid (2)
- Noether’s theorem (2)
- Nonlinear multigrid (2)
- Portfolio optimisation (2)
- Ratenunabhängigkeit (2)
- Regularisierung (2)
- Rotational spinning process (2)
- Special Cosserat rods (2)
- Supply Chain Management (2)
- Variationsungleichungen (2)
- Wavelet (2)
- adaptive refinement (2)
- algorithmic game theory (2)
- asymptotic homogenization (2)
- branch and cut (2)
- discontinuous coefficients (2)
- discrete mechanics (2)
- domain decomposition (2)
- elastoplastic BVP (2)
- energy minimization (2)
- facets (2)
- fast Fourier transform (2)
- fiber orientation (2)
- fiber-fluid interaction (2)
- filling processes (2)
- finite volume method (2)
- free-surface phenomena (2)
- heuristic (2)
- hysteresis (2)
- image analysis (2)
- injection molding (2)
- integer programming (2)
- interface boundary conditions (2)
- linear algebra (2)
- linear elasticity (2)
- mathematical education (2)
- model reduction (2)
- multibody dynamics (2)
- multigrid (2)
- multilayered material (2)
- non-overlapping constraints (2)
- online optimization (2)
- option pricing (2)
- portfolio choice (2)
- power spectrum (2)
- praxis orientated (2)
- rectangular packing (2)
- simulation (2)
- single phase flow (2)
- software development (2)
- stochastic control (2)
- supply chain management (2)
- valid inequalities (2)
- work effort (2)
- (dynamic) network flows (1)
- 3D (1)
- 3d imaging (1)
- : Navier-Stokes equations (1)
- : multiple criteria optimization (1)
- : multiple objective programming (1)
- AG-RESY (1)
- Abstract linear systems theory (1)
- Ad-hoc-Netz (1)
- AmICA (1)
- Assigment (1)
- Asymptotic Expansion (1)
- Asymptotic homogenization (1)
- Automatic Differentiation (1)
- Automatische Differentiation (1)
- Bayesian Model Averaging (1)
- Bell Number (1)
- Berechnungskomplexität (1)
- Betriebsfestigkeit (1)
- Bingham viscoplastic model (1)
- Biot poroelasticity system (1)
- Biot-Savart Operator (1)
- Biot-Savart operator (1)
- Black–Scholes approach (1)
- Blocked Neural Networks (1)
- Boolean polynomials (1)
- Boundary Value Problem (1)
- Brinkman (1)
- Brinkman equations (1)
- CAD (1)
- CFD (1)
- CHAMP <Satellitenmission> (1)
- CIR model (1)
- Capacitated Hub Location (1)
- Capacity decisions (1)
- Code Inspection (1)
- Competitive Analysis (1)
- Compiler (1)
- Complexity theory (1)
- Constant Maturity Credit Default Swap (1)
- Constrained mechanical systems (1)
- Constraint Programming (1)
- Continuum mechanics (1)
- Convex sets (1)
- Coq (1)
- Core (1)
- Cosserat rod (1)
- Credit Default Swaption (1)
- Customer distribution (1)
- Decision support systems (1)
- Delaunay Triangulation (1)
- Delaunay mesh generation (1)
- Design (1)
- Differentialinklusionen (1)
- Discrete linear systems (1)
- Distortion measure (1)
- Domain Decomposition (1)
- Drahtloses Sensorsystem (1)
- Dynamic Network Flows (1)
- Dynamical Coupling (1)
- Education (1)
- Elastoplastic BVP (1)
- Electrophysiology (1)
- Equicofactor matrix polynomials (1)
- Euler number (1)
- Eulerian-Lagrangian formulation (1)
- Existence of Solutions (1)
- Extraction (1)
- FETI (1)
- FPM (1)
- FPTAS (1)
- Facility location (1)
- Fault Prediction (1)
- Filippov theory (1)
- Filippov-Theorie (1)
- Filtering (1)
- Financial Mathematics (1)
- Finite rotations (1)
- Flexible multibody dynamics (1)
- Flooding (1)
- FlowLoc (1)
- Fluid Structure Interaction (1)
- Fokker-Planck Equation (1)
- Fokker-Planck equations (1)
- Folgar-Tucker equation (1)
- Folgar-Tucker model (1)
- Formal Semantics (1)
- Front Propagation (1)
- Fräsen (1)
- Funknetz (1)
- G2++ model (1)
- Generalized LBE (1)
- Geographical Information Systems (1)
- Geomagnetic Field Modelling (1)
- Geomagnetismus (1)
- Geomathematik (1)
- Geometric (1)
- Geostrophic flow (1)
- Gradual Covering (1)
- Gravimetrie (1)
- Greedy Heuristic (1)
- Green’s function (1)
- Grid Generation (1)
- Gröber basis (1)
- HJM (1)
- Hals-Nasen-Ohren-Chirurgie (1)
- Hals-Nasen-Ohren-Heilkunde (1)
- Hankel matrix (1)
- Hardware Description Langauge (HDL) (1)
- Hardware Description Language (HDL) (1)
- Hedge funds (1)
- Helmholtz-Decomposition (1)
- Helmholtz-Zerlegung (1)
- Heston Model (1)
- Heuristic (1)
- Home Health Care (1)
- Homotopie (1)
- Homotopiehochhebungen (1)
- Homotopy (1)
- Homotopy lifting (1)
- Hub Location (1)
- Hub-and-Spoke-System (1)
- Hull White model (1)
- Human resource modeling (1)
- Hysteresis (1)
- Hörgerät (1)
- IMRT planning on adaptive volume structures – a significant advance of computational complexity (1)
- Implantation (1)
- Incompressible Navier-Stokes equations (1)
- Infiltration (1)
- Injectivity of mappings (1)
- Injektivität von Abbildungen (1)
- Inkorrekt gestelltes Problem (1)
- Investigation (1)
- Isabelle/HOL (1)
- Iterative learning control (1)
- Jiang's constitutive model (1)
- Jiangsches konstitutives Gesetz (1)
- Jiang’s Model of Elastoplasticity (1)
- Kaktusgraph (1)
- Kalman Filter (1)
- Kirchhoff and Cosserat rods (1)
- Kirchhoff\\\'s geometrically theory (1)
- Knowledge Extraction (1)
- Kommunikationsprotokoll (1)
- Komplexitätsklasse NP (1)
- Kontinuumsmechanik (1)
- Konvexe Mengen (1)
- LIBOR market model (1)
- Lagrange formalism (1)
- Large deformations (1)
- Lattice Boltzmann (1)
- Lattice Boltzmann method (1)
- Lattice Boltzmann methods (1)
- Lattice Boltzmann models (1)
- Lattice-Boltzmann method (1)
- Least squares approximation (1)
- Level Set method (1)
- Li Ion Batteries (1)
- Linear Programming (1)
- Linear kinematic hardening (1)
- Linear kinematische Verfestigung (1)
- Lineare Optimierung (1)
- Liquid Polymer Moulding (1)
- Load Balancing (1)
- Locational Planning (1)
- MBS simulation (1)
- MILP formulations (1)
- MIP formulations (1)
- Mapping (1)
- Mastoid (1)
- Mastoidektomie (1)
- Mathematical modeling (1)
- Matrix perturbation theory (1)
- Melt spinning (1)
- Mesh-less methods (1)
- Meshfree Method (1)
- Meshfree method (1)
- Metaheuristics (1)
- Methode der Fundamentallösungen (1)
- Mie-Darstellung (1)
- Mie-Representation (1)
- Model Checking (1)
- Model reduction (1)
- Modeling (1)
- Modelling (1)
- Monte Carlo methods (1)
- Monte-Carlo methods (1)
- Multi-dimensional systems (1)
- Multibody simulation (1)
- Multicriteria decision making (1)
- Multipoint flux approximation (1)
- Multiscale problem (1)
- Multiscale problems (1)
- Multiscale structures (1)
- NP-hard (1)
- Nash equilibria (1)
- Navier-Stokes (1)
- Navier-Stokes equation (1)
- Navier-Stokes-Brinkmann system of equations (1)
- Network Location (1)
- Network design (1)
- Networks (1)
- Neumann problem (1)
- Nichtlineare/große Verformungen (1)
- Node Platform Design (1)
- Non-Newtonian flow (1)
- Non-homogeneous Poisson Process (1)
- Nonequilibrium Thermodynamics (1)
- Nonlinear Regression (1)
- Nonlinear energy (1)
- Nonlinear/large deformations (1)
- Numerical modeling (1)
- OCL 2.0 (1)
- Ohrenchirurgie (1)
- One-dimensional systems (1)
- Online Algorithms (1)
- Optimal parameter estimation (1)
- Optimization (1)
- Option pricing (1)
- Ordered Median Function (1)
- Ornstein-Uhlenbeck Process (1)
- Parallel Programming (1)
- Parameter Identification (1)
- Parameter identification (1)
- Parameteridentifikation (1)
- Parametrisation of rotations (1)
- Parsimonious Heston Model (1)
- Parteto surface (1)
- Particle scheme (1)
- Peer-to-Peer-Netz (1)
- Performance of iterative solvers (1)
- Pleated Filter (1)
- Poisson line process (1)
- Poroelastizität (1)
- Preconditioners (1)
- Profiles (1)
- Projection method (1)
- Pseudopolynomial-Time Algorithm (1)
- Quanto option (1)
- RONAF (1)
- Random set (1)
- Rate-independency (1)
- Realization theory (1)
- Recycling (1)
- Regelung (1)
- Reliability Prediction (1)
- Reservierungsprotokoll (1)
- Restricted Shortest Path (1)
- Ripley’s K function (1)
- Roboter (1)
- Rosenbrock methods (1)
- Rotational Fiber Spinning (1)
- Rounding (1)
- Route Planning (1)
- Routing (1)
- SAW filters (1)
- SDL (1)
- SDL-2000 (1)
- SGG (1)
- SIMPLE (1)
- SST (1)
- Satellitengradiometrie (1)
- Schädelchirurgie (1)
- Sensitivitäten (1)
- Shapley Value (1)
- Shapley value (1)
- Shapleywert (1)
- Sheet ofPaper (1)
- Simplex (1)
- Simulation (1)
- Slender body theory (1)
- Solid-Gas Separation (1)
- Solid-Liquid Separation (1)
- Spezifikation (1)
- Spieltheorie (1)
- Sprachprofile (1)
- Standortplanung (1)
- Stationary heat equation (1)
- Stein equation (1)
- Stochastic Differential Equations (1)
- Stokes-Brinkman equations (1)
- Stop- and Play-Operators (1)
- Stop- und Play-Operator (1)
- Stop-und Play-Operator (1)
- Stress-strain correction (1)
- Stücklisten (1)
- Supply Chain Design (1)
- Switching regression model (1)
- System Abstractions (1)
- Theorie schwacher Lösungen (1)
- Thermal Transport (1)
- Train Rearrangement (1)
- Translation Validation (1)
- UML 2 (1)
- UML Profile (1)
- Unstructured Grid (1)
- VCG payment scheme (1)
- VHDL (1)
- Variational inequalities (1)
- Variationsungleichugen (1)
- Vasicek model (1)
- Vectorial Wavelets (1)
- Vektor-Wavelets (1)
- Vektorkugelfunktionen (1)
- Vektorwavelets (1)
- Viscous Fibers (1)
- Winner Determination Problem (WDP) (1)
- Wireless Communication (1)
- Wireless Sensor Network (1)
- Wireless sensor network (1)
- a posteriori error estimates (1)
- a-priori domain decomposition (1)
- acoustic absorption (1)
- adaptive local refinement (1)
- adaptive triangulation (1)
- additive outlier (1)
- aerodynamic drag (1)
- air drag (1)
- algebraic constraints (1)
- algebraic cryptoanalysis (1)
- algorithm by Bortfeld and Boyer (1)
- aliasing (1)
- anisotropic cicosity (1)
- anisotropy (1)
- applied mathematics (1)
- artial differential algebraic equations (1)
- asymptotic (1)
- asymptotic Cosserat models (1)
- asymptotic limits (1)
- automated analog circuit design (1)
- automatic differentiation (1)
- autoregressive process (1)
- basic systems theoretic properties (1)
- batch presorting problem (1)
- battery modeling (1)
- bedingte Aktionen (1)
- behavioral modeling (1)
- ber dynamics (1)
- big triangle small triangle method (1)
- bills of materials (1)
- bin coloring (1)
- binarization (1)
- boudary condistions (1)
- bounce-back rule (1)
- boundary value problems (1)
- bounds (1)
- cactus graph (1)
- calibration (1)
- calls (1)
- cell volume (1)
- change analysis (1)
- circuit sizing (1)
- cliquet options (1)
- clustering (1)
- clustering and disaggregation techniques (1)
- combinatorial procurement (1)
- competetive analysis (1)
- competitive analysis (1)
- compiler (1)
- complexity (1)
- composite materials (1)
- computational fluid dynamics (1)
- computer algebra (1)
- concentrated electrolyte (1)
- constrained mechanical systems (1)
- constraint propagation (1)
- consumption (1)
- contact problems (1)
- continuous optimization (1)
- controlling (1)
- convergence of approximate solution (1)
- convex (1)
- convex optimization (1)
- cooperative game (1)
- core (1)
- corre- lation (1)
- correlation (1)
- coupled flow in plain and porous media (1)
- credit risk (1)
- credit spread (1)
- cuboidal lattice (1)
- curved viscous fibers (1)
- curved viscous fibers with surface tension (1)
- decision support systems (1)
- decomposition (1)
- defect detection (1)
- deformable bodies (1)
- deformable porous media (1)
- delay management (1)
- design centering (1)
- design optimization (1)
- deterministic technical systems (1)
- dial-a-ride (1)
- dif (1)
- differential algebraic equations (1)
- differential inclusions (1)
- differentialalgebraic equations (1)
- discrete facility location (1)
- discrete location (1)
- discrete optimization (1)
- discrete time setting (1)
- discretisation of control problems (1)
- discriminant analysis (1)
- diusion limits (1)
- dividend discount model (1)
- dividends (1)
- domains (1)
- drag models (1)
- drift due to noise (1)
- durability (1)
- dynamic capillary pressure (1)
- dynamic mode (1)
- dynamic network flows (1)
- earliest arrival flows (1)
- effective elastic moduli (1)
- effective thermal conductivity (1)
- efficient set (1)
- eigenvalue problems (1)
- elastoplasticity (1)
- electrochemical diusive processes (1)
- electrochemical simulation (1)
- electronic circuit design (1)
- elliptic equation (1)
- encapsulation (1)
- energy conservation (1)
- error estimates (1)
- estimation of compression (1)
- evolutionary algorithms (1)
- executive compensation (1)
- executive stockholder (1)
- expert system (1)
- explicit jump (1)
- explicit jump immersed interface method (1)
- exponential utility (1)
- extreme equilibria (1)
- extreme solutions (1)
- fatigue (1)
- fiber dynamics (1)
- fiber model (1)
- fiber-fluid interactions (1)
- fiber-turbulence interaction scales (1)
- fibrous insulation materials (1)
- fibrous materials (1)
- film casting process (1)
- filtration (1)
- financial decisions (1)
- finite difference discretization (1)
- finite differences (1)
- finite element method (1)
- finite elements (1)
- finite sample breakdown point (1)
- finite volume discretization (1)
- finite volume discretization discretization (1)
- finite volume discretizations (1)
- finite volume methods (1)
- finite-volume method (1)
- flexible fibers (1)
- flow in heterogeneous porous media (1)
- flow in porous media (1)
- flow resistivity (1)
- fluid-fiber interactions (1)
- fluid-structure interaction (1)
- force-based simulation (1)
- formal verification (1)
- forward starting options (1)
- fptas (1)
- frameindifference (1)
- free boundary value problem (1)
- free surface (1)
- free surface Stokes flow (1)
- full vehicle model (1)
- functional Hilbert space (1)
- fuzzy logic (1)
- general semi-infinite optimization (1)
- generalized Pareto distribution (1)
- genetic algorithms (1)
- geographical information systems (1)
- geomathematics (1)
- geometrically exact rod models (1)
- geometrically exact rods (1)
- glass processing (1)
- global optimization (1)
- global robustness (1)
- graph laplacian (1)
- guarded actions (1)
- harmonic density (1)
- harmonische Dichte (1)
- heterogeneous porous media (1)
- heuristics (1)
- hierarchical shape functions (1)
- human factors (1)
- human visual system (1)
- hydraulics (1)
- hyperealstic (1)
- image processing (1)
- image segmentation (1)
- impinging jets (1)
- improving and feasible directions (1)
- in-house hospital transportation (1)
- incompressible flow (1)
- inertial and viscous-inertial fiber regimes (1)
- innovation outlier (1)
- integral constitutive equation (1)
- intensity maps (1)
- intensity modulated (1)
- intensity modulated radiotherapy planning (1)
- interactive multi-objective optimization (1)
- interactive navigation (1)
- interfa (1)
- interface problem (1)
- interface problems (1)
- invariants (1)
- ion transport (1)
- isotropy test (1)
- kernel estimate (1)
- kernel function (1)
- kinetic derivation (1)
- knowledge management (1)
- knowledge representation (1)
- kooperative Spieltheorie (1)
- large scale optimization (1)
- lattice Boltzmann equation (1)
- learning curve (1)
- level-set (1)
- lid-driven flow in a (1)
- linear elasticity equations (1)
- linear kinematic hardening (1)
- linear optimization (1)
- liquid composite moulding (1)
- liquid film (1)
- lithium-ion battery (1)
- local approximation of sea surface topography (1)
- local robustness (1)
- locally supported (Green’s) vector wavelets (1)
- locally supported wavelets (1)
- location theory (1)
- log utility (1)
- logistic regression (1)
- logistics (1)
- long slender fibers (1)
- macro modeling (1)
- macroscopic equations (1)
- magnetic field (1)
- mass & spring (1)
- mastoid (1)
- mastoidectomy (1)
- mechanism design (1)
- metal foams (1)
- method of fundamental solutions (1)
- microstructure simulatio (1)
- microstructure simulation (1)
- minimaler Schnittbaum (1)
- minimum cut tree (1)
- models (1)
- modified gradient projection method (1)
- moment matching (1)
- multi-asset (1)
- multi-period planning (1)
- multi-stage stochastic programming (1)
- multibody system simulation (1)
- multicriteria optimization (1)
- multigrid methods (1)
- multiobjective evolutionary algorithms (1)
- multiphase flow (1)
- multiple objective optimization (1)
- multiscale problem (1)
- multiscale problems (1)
- multiscale structures (1)
- multivalued fundamental diagram (1)
- nearest neighbour distance (1)
- neighborhod relationships (1)
- network congestion game (1)
- neural network (1)
- non-Newtonian fluids (1)
- non-linear optimization (1)
- non-linear wealth dynamics (1)
- non-local conditions (1)
- non-woven (1)
- nonlinear programming (1)
- nonlinear stochastic systems (1)
- nonlinearity (1)
- nonparametric regression (1)
- numerical simulation (1)
- numerical solution (1)
- object-orientation (1)
- occupational choice (1)
- oil filters (1)
- on-board simulation (1)
- open cell foam (1)
- operator-dependent prolongation (1)
- optimal control (1)
- optimal control theory (1)
- optimal portfolio choice (1)
- optimization (1)
- optimization algorithms (1)
- optimization strategies (1)
- options (1)
- ordered median (1)
- orientation analysis (1)
- orthogonal orientations (1)
- oscillating coefficients (1)
- otorhinolaryngological surgery (1)
- ownership (1)
- pH-sensitive microelectrodes (1)
- paper machine (1)
- parallel computing (1)
- parallel implementation (1)
- particle methods (1)
- path-connected sublevelsets (1)
- permeability of fractured porous media (1)
- phase space (1)
- phase transitions (1)
- piezoelectric periodic surface acoustic wave filters (1)
- planar location (1)
- polar ice (1)
- political districting (1)
- porous media (1)
- porous microstructure (1)
- power utility (1)
- preconditioner (1)
- pressing section of a paper machine (1)
- price of anarchy (1)
- price of stability (1)
- productivity (1)
- project management and scheduling (1)
- projection-type splitting (1)
- pseudo-plastic fluids (1)
- public transit (1)
- public transport (1)
- public transportation (1)
- puts (1)
- quadratic assignment problem (1)
- quantile estimation (1)
- quasistatic deformations (1)
- quickest path (1)
- radiation therapy planning (1)
- radiotherapy planning (1)
- random -Gaussian aerodynamic force (1)
- random set (1)
- random system of fibers (1)
- rate-independency (1)
- rate-indepenhysteresis (1)
- real-life applications. (1)
- real-time (1)
- real-time simulation (1)
- real-world accident data (1)
- regularization (1)
- regularized models (1)
- representative systems of Pareto solutions (1)
- reproducing kernel (1)
- risk (1)
- robust network flows (1)
- robustness (1)
- rotational spinning processes (1)
- safety critical components (1)
- safety function (1)
- sales territory alignment (1)
- satisfiability (1)
- selfish routing (1)
- semi-infinite programming (1)
- sensitivities (1)
- sequences (1)
- sequential test (1)
- series-parallel graphs (1)
- shape (1)
- shape optimization (1)
- simplex (1)
- single layer kernel (1)
- singularity (1)
- slender- body theory (1)
- slender-body theory (1)
- slenderbody theory (1)
- smoothness (1)
- software process (1)
- spherical decomposition (1)
- spinning processes (1)
- stability (1)
- statistical modeling (1)
- steady Richards’ equation (1)
- steady modified Richards’ equation (1)
- stochastic Hamiltonian system (1)
- stochastic averaging. (1)
- stochastic dif (1)
- stochastic volatility (1)
- stokes (1)
- stop and go waves (1)
- stop- and play-operator (1)
- stop- and play-operators (1)
- strategic (1)
- strength (1)
- strong equilibria (1)
- strut thickness (1)
- subgrid approach (1)
- subgrid approximation (1)
- suspension (1)
- swap (1)
- symbolic analysis (1)
- synchrone Sprachen (1)
- synchronous languages (1)
- system simulation (1)
- tabu search (1)
- technology (1)
- territory desgin (1)
- textile quality control (1)
- texture classification (1)
- theorem prover (1)
- thin films (1)
- topological sensitivity (1)
- topology optimization (1)
- total latency (1)
- tr (1)
- trace stability (1)
- traffic flow (1)
- transfer quality (1)
- translation validation (1)
- transportation (1)
- tree method (1)
- turbulence modeling (1)
- turbulence modelling (1)
- two-grid algorithm (1)
- two-way coupling (1)
- types (1)
- unstructured grid (1)
- upscaling (1)
- urban elevation (1)
- variable aggregation method (1)
- variable neighborhood search (1)
- variational formulation (1)
- vector spherical harmonics (1)
- vectorial wavelets (1)
- viscous thermal jets (1)
- visual (1)
- visual interfaces (1)
- visualization (1)
- volatility (1)
- volume of fluid method (1)
- wave propagation (1)
- weak solution theory (1)
- weakly/ strictly pareto optima (1)
- white noise (1)
- wild bootstrap test (1)

#### Faculty / Organisational entity

Partitioned chain grammars
(1979)

This paper introduces a new class of grammars, the partitioned chain grammars, for which efficient parsers can be automatically generated. Besides being efficiently parsable these grammars possess a number of other properties, which make them very attractive for the use in parser-generators. They for instance form a large grammarclass and describe all deterministic context-free languages. Main advantage of the partitioned chain grammars however is, that given a language it is usually easier to describe it by a partitioned chain grammar than to construct a grammar of some other type commonly used in parser-generators for it.

In this paper we continue the study of p - groups G of square order \(p^{2n}\) and investigate the existence of partial congruence partitions (sets of mutually disjoint subgroups of order \(p^n\)) in G. Partial congruence partitions are used to construct translation nets and partial difference sets, two objects studied extensively in finite geometries and combinatorics. We prove that the maximal number of mutually disjoint subgroups of order \(p^n\) in a group G of order \(p^{2n}\) cannot be more than \((p^{n-1}-1)(p-1)^{-1}\) provided that \(n\ge4\)and that G is not elementary abelian. This improves a result in [6] and as we do not distinguish the cases p=2 and p odd in the present paper, we also have a generalization of D. FROHARDT' s theorem on 2 - groups in [4]. Furthermore we study groups of order \(p^6\). We can show that for each odd prime number, there exist exactly four nonisomorphic groups which contain at least p+2 mutually disjoint subgroups of order \(p^3\). Again, as we do not distinguish between the even and the odd case in advance, we in particular obtain
D. GLUCK' s and A. P. SPRAGUE' s classification of groups of order 64 which contain at least 4 mutually disjoint subgroups of order 8 in [5] and [13] respectively.

In this paper the existence of translation transversal designs which is equivalent to the existence of certain particular partitions in finite groups is studied. All considerations are based on the fact that the particular component of such a partition (the component representing the point classes of the corresponding design) is a normal subgroup of the translation group. With regard to groups admitting an (s,k,\(\lambda\))-partiton, on one hand the already known families of such groups are determined without using R. BAER's, 0.H.KEGEL's and M. SUZUKI' s classification of finite groups with partition and on the other hand some new results on the special structure of p - groups are proved. Furthermore, the existence of a series of nonabelian p - groups of odd order which can be represented as translation groups of certain (s,k,1) - translation transversal designs is shown; moreover, the translation groups are normal subgroups of collineation groups acting regularly on the set of flags of the same designs.

The notion of Q-Gorenstein smoothings has been introduced by Kollar. ([KoJ], 6.2.3). This notion is essential for formulating Kollar's conjectures on smoothing components for rational surface singularities. He conjectures, loosely speaking, that every smoothing of a rational surface singularity can be obtained by blowing down a deformation of a partial resolution, this partial resolution having the property (among others) that the singularities occuring on it all have qG-smoothings. (For more details and precise statements see [Ko], ch. 6.). It is therefore of interest to construct singularities having qG-smoothings.

Limits of instantons
(1991)

Moduli for singularities
(1991)

The aim of this article is to give a survey on recent results about moduli spaces for curve singularities and for modules over the local ring of a fixed curve singularity. We emphasize especially the general concept which lies behind these constructions.
Therefore, the article might be useful to the reader who wishes to have the leading ideas and the main steps of the proofs explained without going into all the details. We also calculate explicit examples (for singularities and for modules) which illustrate
the general theorems.

Trimming of surfaces and volumes, curve and surface modeling via Bézier's idea of destortion, segmentation, reparametrization, geometric continuity are examples of applications of functional composition. This paper shows how to
compose polynomial and rational tensor product Bézier representations. The problem of composing Bezier splines and B-spline representations will also be addressed in this paper.

The use of non-volatile semiconductor memory within an extended storage hierarchy promises significant performance improvements for transaction processing. Although page-addressable semiconductor memories like extended memory, solid-state disks and disk caches are commercially available since several years, no detailed investigation of their use for transaction processing has been performed so far. We present a comprehensive simulation study that compares the performance of these storage types and of different usage forms. The following usage forms are considered: allocation of entire log and database files in non-volatile semiconductor memory, using a so-called write buffer to perform disk writes asynchronously, and caching of database pages at intermediate storage levels (in addition to main memory caching). Our simulations are conducted with both synthetically generated workloads and traces from real-life database applications. In particular, simulation results will be presented for the debit-credit workload frequently used in transaction processing benchmarks. As expected, the greatest performance improvements (but at the highest cost) can be achieved by storing log and database files completely in non-volatile semiconductor memory. For update-intensive
workloads, a limited amount of non-volatile memory used as a write buffer also proved to be very effective. To reduce the number of disk reads; caching of database pages in addition to main memory is best supported by an extended memory buffer. In this respect, disk caches are found to be less effective as they are designed for one-level caching. Different storage costs suggest that it may be cost-effective to use two or even three of the intermediate storage types together. The performance improvements obtainable by the use of non-volatile semiconductor memory is also found to reduce the need for sophisticated DBMS buffer management in order to achieve high transaction processing performance.

We show that the different module structures of GF(\(q^m\)) arising from the intermediate fields of GF(\(q^m\))and GF(q) can be studied simultaneously with the help of some basic properties of cyclotomic polynomials. We use this ideas to give a detailed and constructive proof of the most difficult part of a Theorem of D. Blessenohl and K. Johnsen (1986), i.e., the existence of elements v in GF(\(q^m\)) over GF(q) which generate normal bases over any intermediate field of GF(\(q^m\)) and GF(q), provided that m is a prime power. Such elements are called completely free in GF(\(q^m\)) over GF(q). We develop a recursive formula for the number of completely free elements in GF(\(q^m\)) over GF(q) in the case where m is a prime power. Some of the results can be generalized to finite cyclic Galois extensions
over arbitrary fields.

Let \(a_1, i:=1,\dots,m\), be an i.i.d. sequence taking values in \(\mathbb{R}^n\), whose convex hull is interpreted as a stochastic polyhedron \(P\). For a special class of random variables, which decompose additively relative to their boundary simplices, eg. the volume of \(P\), simple integral representations of its first two moments are given in case of rotationally symmetric distributions in order to facilitate estimations of variances or to quantify large deviations from the mean.

We are concerned with a parameter choice strategy for the Tikhonov regularization \((\tilde{A}+\alpha I)\tilde{x}\) = T* \(\tilde{y}\)+ w where \(\tilde{A}\) is a (not necessarily selfadjoint) approximation of T*T and T*\(\tilde y\)+ w is a perturbed form of the (not exactly computed) term T*y. We give conditions for convergence and optimal convergence rates.

Let \(a_i i:= 1,\dots,m.\) be an i.i.d. sequence taking values in \(\mathbb{R}^n\). Whose convex hull is interpreted as a stochastic polyhedron \(P\). For a special class of random variables which decompose additively relative to their boundary simplices, eg. the volume of \(P\), integral representations of their first two moments are given which lead to asymptotic estimations of variances for special "additive variables" known from stochastic approximation theory in case of rotationally symmetric distributions.

User interfaces for large distributed applications have to handle specific problems: the complexity of the application itself and the integration of online-data into the user interface. A main task of the user interface architecture is to provide powerful tools to design and augment the end-user system easily, hence giving the designer more time to focus on user requirements. Our experiences developing a user interface system for a process control room showed that a lot of time during the development process is wasted for the integration of online-data residing anywhere but not in the user interface itself. Furtheron external data may be kept by different kinds of programs, e.g. C-programs running
a numerical process model or PROLOG-programs running a diagnosis system, both in parallel to the process and in parallel to the user interface. Facing these specific requirements, we developed a user interface architecture following two main goals: 1. integration of external information into high-level graphical objects and 2. the system should be open for any program running as a separate process using its own problem-oriented language. The architecture is based on two approaches: an asynchronous, distributed and language independent communication model and an object model describing the problem domain and the interface using object-oriented techniques. Other areas like rule-based programming are involved, too. With this paper, we will present the XAVIA user interface architecture, the (as far as we know) first user inteface architecture, which is consequently based on a distributed object model.

Gauss Frame Offsets
(1992)

Weighted k-cardinality trees
(1992)

We consider the k -CARD TREE problem, i.e., the problem of finding in a given undirected graph G a subtree with k edges, having minimum weight. Applications of this problem arise in oil-field leasing and facility layout. While the general problem is shown to be strongly NP hard, it can be solved in polynomial time if G is itself a tree. We give an integer programming formulation of k-CARD TREE, and an efficient exact separation routine for a set of generalized subtour elimination constraints. The polyhedral structure of the convex huLl of the integer solutions is studied.

Facility location problems in the plane are among the most widely used tools of Mathematical Programming in modeling real-world problems. In many of these problems restrictions have to be considered which correspond to regions in which a placement of new locations is forbidden. We consider center and median problems where the forbidden set is
a union of pairwise disjoint convex sets. As applications we discuss the assembly of printed circuit boards, obnoxious facility location and the location of emergency facilities.

We present a generalization of Proth's theorem for testing certain large integers for primality. The use of Gauß sums leads to a much simpler approach to these primality criteria as compared to the earlier tests. The running time of the algorithms is bounded by a polynomial in the length of the input string. The applicability of our algorithms is linked to certain diophantine approximations of \(l\)-adic roots of unity.

Hyperidentities
(1992)

The concept of a free algebra plays an essential role in universal algebra and in computer science. Manipulation of terms, calculations and the derivation of identities are performed in free algebras. Word problems, normal forms, system of reductions, unification and finite bases of identities are topics in algebra and logic as well as in computer science. A very fruitful point of view is to consider structural properties of free algebras. A.I. Malcev initiated a thorough research of the congruences of free algebras. Henceforth congruence permutable, congruence distributive and congruence modular varieties are
intensively studied. A lot of Malcev type theorems are connected to the congruence lattice of free algebras. Here we consider free algebras as semigroups of compositions of terms and more specific as clones of terms. The properties of these semigroups and clones are adequately described by hyperidentities. Naturally a lot of theorems of "semigroup" or "clone" type can be derived. This topic of research is still in its beginning and therefore a lot öf concepts and results cannot be presented in a final and polished form. Furthermore a lot of problems and questions are open which are of importance for the further development of the theory of hyperidentities.

Virtual Reality (VR) is to be seen as the superset of simulation and animation. Visualization is done by rendering. The fundamental model of VR accounts for all phenomenons to be modelled with help of a computer. Examples range from simple dragging actions with a mouse device to the complex simulation of physically based animation.

Max ordering (MO) optimization is introduced as tool for modelling production
planning with unknown lot sizes and in scenario modelling. In MO optimization a feasible solution set \(X\) and, for each \(x\in X, Q\) individual objective functions \(f_1(x),\dots,f_Q(x)\) are given. The max ordering objective
\(g(x):=max\) {\(f_1(x),\dots,f_Q(x)\)} is then minimized over all \(x\in X\).
The paper discusses complexity results and describes exact and approximative
algorithms for the case where \(X\) is the solution set of combinatorial
optimization problems and network flow problems, respectively.

Let \(A\):= {\(a_i\mid i= 1,\dots,m\)} be an i.i.d. random sample in (\mathbb{R}^n\), which we consider a random polyhedron, either as the convex hull of the \(a_i\) or as the intersection of halfspaces {\(x \mid a ^T_i x\leq 1\)}. We introduce a class of polyhedral functionals we will call "additive-type functionals", which covers a number of polyhedral functionals discussed in different mathematical fields, where the emphasis in our contribution will be on those, which arise in linear optimization theory. The class of additive-type functionals is a suitable setting in order to unify and to simplify the asymptotic probabilistic analysis of first and second moments of polyhedral functionals. We provide examples of asymptotic results on expectations and on variances.

The article provides an asymptotic probabilistic analysis of the variance of the number of pivot steps required by phase II of the "shadow vertex algorithm" - a parametric variant of the simplex algorithm, which has been proposed by Borgwardt [1] . The analysis is done for data which satisfy a rotationally
invariant distribution law in the \(n\)-dimensional unit ball.

Despite their very good empirical performance most of the simplex algorithm's variants require exponentially many pivot steps in terms of the problem dimensions of the given linear programming problem (LPP) in worst-case situtation. The first to explain the large gap between practical experience and the disappointing worst-case was Borgwardt (1982a,b), who could prove polynomiality on tbe average for a certain variant of the algorithm-the " Schatteneckenalgorithmus (shadow vertex algorithm)" - using a stochastic problem simulation.

Efficient algorithms and structural results are presented for median
problems with 2 new facilities including the classical 2-Median problem,
the 2-Median problem with forbidden regions and bicriterial 2-Median
problems. This is the first paper dealing with multi-facility multiobjective location problems. The time complexity of all presented algorithms is O(MlogM), where M is the number of existing facilities.

Given Q different objective functions, three types of single-facility problems
are considered: Lexicographic, pareto and max ordering problems. After discussing the interrelation between the problem types, a complete characterization of lexicographic locations and some instances of pareto and max ordering locations is given. The characterizations result in efficient solution algorithms for finding these locations. The paper relies heavily on the theory of restricted locations developed by the same authors, and can be further extended, for instance, to multifacility problems with several objectives. The proposed approach is more general than previously published results on multicriteria planar location problems and is particulary suited for modelling real-world problems.

We investigate two versions of multiple objective minimum spanning tree
problems defined on a network with vectorial weights. First, we want to minimize the maximum of Q linear objective functions taken over the set of all spanning trees (max linear spanning tree problem ML-ST). Secondly, we look for efficient spanning trees (multi criteria spanning tree problem MC-ST). Problem ML-ST is shown to be NP-complete. An exact algorithm which is based on ranking is presented. The procedure can also be used as an approximation scheme. For solving the bicriterion MC-ST, which in the worst case may have an exponential number of efficient trees, a two-phase procedure is presented. Based on the computation of extremal efficient spanning trees we use neighbourhood search to determine a sequence of solutions with the property that the distance
between two consecutive solutions is less than a given accuracy.

Shadow-Mapping
(1993)

Most radiosity techniques store radiosities in certain sample points, typically the vertices of polyhedral scenes. As diffuse radiosities are view independent they can be used for an interactive 'walk-through'. This paper presents an algorithm for storing radiosities independent of the representation of the object. A distributed rendering system, which uses this shadow-mapping technique is described. The basic thermophysical definitions, needed to derive a sum formula for a form factor calculation of polygons, are explained.

This paper describes some new algorithms for the accurate calculation of surface properties. In the first part an arithmetic on Bézier surfaces is introduced. Formulas are given, which determine the Bézier points and weights of the resulting surface from the points and weights of the operand surfaces. An application of the arithmetic operations to the surface interrogation methods are described in the second part. It turns out, that the quality analysis can be reduced to a few numerical stable operations. Finally the advantages and disadvantages of this method are discussed.

The composition of Bézier curves and tensor product Bézier surfaces, polynomial as well as rational, is applied to exactly and explicitely represent trim curves of tensor product Bézier surfaces. Trimming curves are assumed to be defined as Bézier curves in surface parameter domain. A Bézier spline approximation of lower polynomial degree is built up as weil which is based on the exact trim curve representation in coordinate space.

Order-semi-primal lattices
(1994)

A nonequilibrium situation governed by kinetic equations with strongly contrasted Knudsen numbers in different subdomains is discussed. We consider a domain decomposition problem for Boltzmann- and Euler equations, establish the correct coupling conditions and prove the validity of the obtained coupled solution . Moreover numerical examples comparing different types of coupling conditions are presented.

Let (\(a_i)_{i\in \bf{N}}\) be a sequence of identically and independently distributed random vectors drawn from the \(d\)-dimensional unit ball \(B^d\)and let \(X_n\):= convhull \((a_1,\dots,a_n\)) be the random polytope generated by \((a_1,\dots\,a_n)\). Furthermore, let \(\Delta (X_n)\) : = (Vol \(B^d\) \ \(X_n\)) be the deviation of the polytope's volume from the volume of the ball. For uniformly distributed \(a_i\) and \(d\ge2\), we prove that tbe limiting distribution of \(\frac{\Delta (X_n)} {E(\Delta (X_n))}\) for \(n\to\infty\) satisfies a 0-1-law. Especially, we provide precise information about the asymptotic behaviour of the variance of \(\Delta (X_n\)). We deliver analogous results for spherically symmetric distributions in \(B^d\) with regularly varying tail.

We introduce the concept of streamballs for fluid flow visualization. Streamballs are based upon implicit surface generation techniques adopted from the well-known metaballs. Their property to split or merge automatically in areas of significant divergence or convergence makes them an ideal tool for the visualization of arbitrary complex flow fields. Using convolution surfaces generated by continuous skeletons for streamball construction offers the possibility to visualize even tensor fields.

The problem to interpolate Hermite-type data (i.e. two points with attached tangent vectors) with elastic curves of prescribed tension is known to have multiple solutions. A method is presented that finds all solutions of length not exceeding one period of its curvature function. The algorithm is based on algebraic relations between discrete curvature information which allow to transform the problem into a univariate one. The method operates with curves that by construction partially interpolate the given data. Hereby the objective function of the problem is drastically simplified. A bound on the maximum curvature value is established that provides an interval containing all solutions.

Best-Fit Pattern Matching
(1994)

This report shows that dispatching of methods in object oriented languages is in principle the same as best fit pattern matching. A general conceptual description of best fit pattern matching is presented. Many object oriented features are modelled by means of the general concept. This shows that simple methods, multi methods, overloading of functions, pattern matching,
dynamic and union types, and extendable records can be combined in a single comprehensive concept.

We study the complexity of local solution of Fredholm integral equations. This means that we want to compute not the full solution, but rather a functional (weighted mean, value in a point) of it. For certain Sobolev classes of multivariate periodic functions we prove matching upper and lower bounds and construct an algorithm of the optimal order, based on Fourier coefficients and a hyperbolic cross approximation.

In this paper the complexity of the local solution of Fredholm integral equations
is studied. For certain Sobolev classes of multivariate periodic functions with dominating mixed derivative we prove matching lower and upper bounds. The lower bound is shown using relations to s-numbers. The upper bound is proved in a constructive way providing an implementable algorithm of optimal order based on Fourier coefficients and a hyperbolic cross approximation.

The Basic Reference Model of ODP introduces a number of basic concepts in order to provide a common basis for the development of a coherent set of standards. To achieve this objective, a clear understanding of the basic concepts is one prerequisite. This paper makes an effort at clarifying some of the basic concepts independently of standardized or non-standardized formal description techniques. Among the basic concepts considered here are: agent, action, interaction, interaction point, architecture, behaviour, system, composition, refinement, and abstraction. In a case study, it is then shown how these basic concepts can be represented in a formal specification written in temporal logic.

Free Form Volumes
(1994)

Visualization of large data sets, especially on small machines, requires advanced techniques in image processing and image generation. Our hybrid raytracer is capable of rendering volumetric and geometric data simultaneously, without loss of accuracy due to data conversion. Compound data sets, consisting of several types of data, are called "hybrid data sets". There is only one rendering pipeline to obtain loss-less and efficient visualization of hybrid data. Algorithms apply to both types of data. Optical material properties are stored in the same data base for both volumetric and geometric objects, and anti-aliasing methods appeal to both data types. Stereoscopic display routines have been added to obtain true three-dimensional visualization on various media, and animation features allow generation of recordable 3-D sequences.

The main problem in computer graphics is to solve the global illumination problem,
which is given by a Fredholm integral equation of the second kind, called the radiance equation (REQ). In order to achieve realistic images, a very complex kernel
of the integral equation, modelling all physical effects of light, must be considered. Due to this complexity Monte Carlo methods seem to be an appropriate approach to solve the REQ approximately. We show that replacing Monte Carlo by quasi-Monte Carlo in some steps of the algorithm results in a faster convergence.

Optimization of Projection Methods for Solving ill-posed Problems. In this paper we propose a modification of the projection scheme for solving ill-posed problems. We show that this modification allows to obtain the best possible order of accuracy of Tikhonov Regularization using an amount of information which is far less than for the standard projection technique.

The rapid development of any field of knowledge brings with it unavoidable fragmentation and proliferation of new disciplines. The development of computer science is no exception. Software engineering (SE) and human-computer interaction (HCI) are both relatively new disciplines of computer science. Furthermore, as both names suggest, they each have strong connections with other subjects. SE is concerned with methods and tools for general software development based on engineering principles. This discipline has its roots not only in computer science but also in a number of traditional engineering disciplines. HCI is concerned with methods and tools for the development of human-computer interfaces, assessing the usability of computer systems and with broader issues about how people interact with computers. It is based on theories about how humans process information and interact with computers, other objects and other people in the organizational and social contexts in
which computers are used. HCI draws on knowledge and skills from psychology, anthropology and sociology in addition to computer science. Both disciplines need ways of measuring how well their products and development processes fulfil their intended requirements. Traditionally SE has been concerned with 'how software is constructed' and HCI with 'how people use software'. Given the
different histories of the disciplines and their different objectives, it is not surprising that they take different approaches to measurement. Thus, each has its own distinct 'measurement culture.' In this paper we analyse the differences and the commonalties of the two cultures by examining the measurement approaches used by each. We then argue the need for a common measurement taxonomy and framework, which is derived from our analyses of the two disciplines. Next we demonstrate the usefulness of the taxonomy and framework via specific example studies drawn from our own work and that of others and show that, in fact, the two disciplines have many important similarities as well as differences and that there is some evidence to suggest that they are growing closer. Finally, we discuss the role of the taxonomy as a framework to support: reuse, planning future studies, guiding practice and facilitating communication between the two disciplines.

Hardware / Software Codesign
(1994)

Monte Carlo integration is often used for antialiasing in rendering processes.
Due to low sampling rates only expected error estimates can be stated, and the variance can be high. In this article quasi-Monte Carlo methods are presented, achieving a guaranteed upper error bound and a convergence rate essentially as fast as usual Monte Carlo.

The radiance equation, which describes the global illumination problem in computer graphics, is a high dimensional integral equation. Estimates of the solution are usually computed on the basis of Monte Carlo methods. In this paper we propose and investigate quasi-Monte Carlo methods, which means that we replace (pseudo-) random samples by low discrepancy sequences, yielding deterministic algorithms. We carry out a comparative numerical study between Monte Carlo and quasi-Monte Carlo methods. Our results show that quasi-Monte Carlo converges considerably faster.

This report presents a generalization of tensor-product B-spline surfaces. The new scheme permits knots whose endpoints lie in the interior of the domain rectangle of a surface. This allows local refinement of the knot structure for approximation purposes as well as modeling surfaces with local tangent or curvature discontinuities. The surfaces are represented in terms of B-spline basis functions, ensuring affine invariance, local control, the convex hull property, and evaluation by de Boor's algorithm. A dimension formula for a class of generalized tensor-product spline spaces is developed.

In this paper, the complexity of full solution of Fredholm integral equations of the second kind with data from the Sobolev class \(W^r_2\) is studied. The exact order of information complexity is derived. The lower bound is proved using a Gelfand number technique. The upper bound is shown by providing a concrete algorithm of optimal order, based on a specific hyperbolic cross approximation of the kernel function. Numerical experiments are included, comparing the optimal algorithm with the standard Galerkin method.

Let \(a_1,\dots,a_m\) be i.i .d. vectors uniform on the unit sphere in \(\mathbb{R}^n\), \(m\ge n\ge3\) and let \(X\):= {\(x \in \mathbb{R}^n \mid a ^T_i x\leq 1\)} be the random polyhedron generated by. Furthermore, for linearly independent vectors \(u\), \(\bar u\) in \(\mathbb{R}^n\), let \(S_{u, \bar u}(X)\) be the number of shadow vertices of \(X\) in \(span (u, \bar u\)). The paper provides an asymptotic expansion of the expectation value \(E (S_{u, \bar u})\) for fixed \(n\) and \(m\to\infty\). The first terms of the expansion are given explicitly. Our investigation of \(E (S_{u, \bar u})\) is closely connected to Borgwardt's probabilistic analysis of the shadow vertex algorithm - a parametric variant of the simplex algorithm. We obtain an improved asymptotic upper bound for the number of pivot steps required by the shadow vertex algorithm for uniformly on the sphere distributed data.

Experience gathered from applying the software process modeling language MVP-L in software development organizations has shown the need for graphical representations of process models. Project members (i.e„ non MVP-L specialists) review models much more easily by using graphical representations. Although several various graphical notations were developed for individual projects in which MVP-L was applied, there was previously no consistent definition of a mapping between textual MVP-L models and graphical representations. This report defines a graphical representation schema for MVP-L
descriptions and combines previous results in a unified form. A basic set of building blocks (i.e., graphical symbols and text fragments) is defined, but because we must first gain experience with the new symbols, only rudimentary guidelines are given for composing basic
symbols into a graphical representation of a model.

Intellectual control over software development projects requires the existence of an integrated set of explicit models of the products to be developed, the processes used to develop them, the resources needed, and the productivity and quality aspects involved. In recent years the development of languages, methods and tools for modeling software processes, analyzing and enacting them has become a major emphasis of software engineering research. The majority of current process research concentrates on prescriptive modeling of small, completely formalizable processes and their execution entirely on computers. This research direction has produced process modeling languages suitable for machine rather than human consumption. The MVP project, launched at the University of Maryland and continued at Universität Kaiserslautern, emphasizes building descriptive models of large, real-world processes and their use by humans and computers for the purpose of understanding, analyzing, guiding and improving software development projects. The language MVP-L has been developed with these purposes in mind. In this paper, we
motivate the need for MVP-L, introduce the prototype language, and demonstrate its uses. We assume that further improvements to our language will be triggered by lessons learned from applications and experiments.

Optimal degree reductions, i.e. best approximations of \(n\)-th degree Bezier curves
by Bezier curves of degree \(n\) - 1, with respect to different norms are studied. It
is shown that for any \(L_p\)-norm the euclidean degree reduction where the norm is applied to the euclidean distance function of two curves is identical to componentwise degree reduction. The Bezier points of the degree reductions are found to lie on parallel lines through the Bezier points of any Taylor expansion of degree \(n\) - 1 of the original curve. This geometric situation is shown to hold also in the case of constrained degree reduction. The Bezier points of the degree reduction are explicitly given in the unconstrained case for \(p\) = 1 and \(p\) = 2 and in the constrained case for \(p\) = 2.

The local solution problem of multivariate Fredholm integral equations is studied. Recent research proved that for several function classes the complexity of this problem is closely related to the Gelfand numbers of some characterizing operators. The generalization of this approach to the situation of arbitrary Banach spaces is the subject of the present paper.
Furthermore, an iterative algorithm is described which - under some additional conditions - realizes the optimal error rate. The way these general theorems work is demonstrated by applying them to integral equations in a Sobolev space of periodic functions with dominating mixed derivative of various order.

In multiple criteria optimization an important research topic is the topological structure of the set \( X_e \) of efficient solutions. Of major interest is the connectedness of \( X_e \), since it would allow the determination of \( X_e \) without considering non-efficient solutions in the
process. We review general results on the subject,including the connectedness result for efficient solutions in multiple criteria linear programming. This result can be used to derive a definition of connectedness for discrete optimization problems. We present a counterexample to a previously stated result in this area, namely that the set of efficient solutions of the shortest path problem is connected. We will also show that connectedness does not hold for another important problem in discrete multiple criteria optimization: the spanning tree problem.

In this paper we will introduce the concept of lexicographic max-ordering solutions for multicriteria combinatorial optimization problems. Section 1 provides the basic notions of
multicriteria combinatorial optimization and the definition of lexicographic max-ordering solutions. In Section 2 we will show that lexicographic max-ordering solutions are pareto optimal as well as max-ordering optimal solutions. Furthermore lexicographic max-ordering solutions can be used to characterize the set of pareto solutions. Further properties of lexicographic max-ordering solutions are given. Section 3 will be devoted to algorithms. We give a polynomial time algorithm for the two criteria case where one criterion is a sum and one is a bottleneck objective function, provided that the one criterion sum problem is solvable in polynomial time. For bottleneck functions an algorithm for the general case of Q criteria is presented.

In this paper we investigate two optimization problems for matroids with multiple objective functions, namely finding the pareto set and the max-ordering problem which conists in finding a basis such that the largest objective value is minimal. We prove that the decision versions of both problems are NP-complete. A solution procedure for the max-ordering problem is presented and a result on the relation of the solution sets of the two problems is given. The main results are a characterization of pareto bases by a basis exchange property and finally a connectivity result for proper pareto solutions.

This paper introduces a new high Level programming language for a novel
class of computational devices namely data-procedural machines. These machines are by up to several orders of magnitude more efficient than the von Neumann paradigm of computers and are as flexible and as universal as computers. Their efficiency and flexibility is achieved by using field-programmable logic as the essential technology platform. The paper briefly summarizes and illustrates the essential new features of this language by means of two example programs.

A new variance reduction technique for the Monte Carlo solution of integral
equations is introduced. It is based on separation of the main part. A neighboring equation with exactly known solution is constructed by the help of a deterministic Galerkin scheme. The variance of the method is analyzed, and an application to the radiosity equation of computer graphics, together with numerical test results is given.

In this paper an analytic hidden surface removal algorithm is presented which uses a combination
of 2D and 3D BSP trees without involving point sampling or scan conversion. Errors like aliasing
which result from sampling do not occur while using this technique. An application of this
algorithm is outlined which computes the energy locally reflected from a surface having an
arbitrary BRDF. A simplification for diffuse reflectors is described, which has been implemented
to compute analytic form factors from diffuse light sources to differential receivers as they are needed for shading and radiosity algorithms.

The CAD/CAM-based design of free-form surfaces is the beginning of a chain of operations, which ends with the numerically controlled (NC-) production of the designed object. During this process the shape control is an important step to amount efficiency. Several surface interrogation methods already exist to analyze curvature and continuity behaviour of the shape. This paper deals with a new aspect of shape control: the stability of surfaces with respect to infnitesimal bendings. Each inEnitesimal bending of a surface determines a so called instability surface, which is used for the stability investigations. The kinematic meaning of this instability surface will be discussed and we present algorithms to calculate it.

Computer processing of free form surfaces forms the basis of a closed construction process starting with surface design and up to NC-production.
Numerical simulation and visualization allow quality analysis before manufacture. A new aspect in surface analysis is described, the stability
of surfaces versus infinitesimal bendings. The stability concept is derived
from the kinetic meaning of a special vector field which is given by the deformation. Algorithms to calculate this vector field together with an appropriate visualization method give a tool able to analyze surface stability.

The \(L_2\)-discrepancy is a quantitative measure of precision for multivariate quadrature rules. It can be computed explicitly. Previously known algorithms needed \(O(m^2\)) operations, where \(m\) is the number of nodes. In this paper we present algorithms which require
\(O(m(log m)^d)\) operations.

Let \(a_1,\dots,a_n\) be independent random points in \(\mathbb{R}^d\) spherically symmetrically but not necessarily identically distributed. Let \(X\) be the random polytope generated as the convex hull of \(a_1,\dots,a_n\) and for any \(k\)-dimensional subspace \(L\subseteq \mathbb{R}^d\) let \(Vol_L(X) :=\lambda_k(L\cap X)\) be the volume of \(X\cap L\) with respect to the \(k\)-dimensional Lebesgue measure \(\lambda_k, k=1,\dots,d\). Furthermore, let \(F^{(i)}\)(t):= \(\bf{Pr}\) \(\)(\(\Vert a_i \|_2\leq t\)),
\(t \in \mathbb{R}^+_0\) , be the radial distribution function of \(a_i\). We prove that the expectation
functional \(\Phi_L\)(\(F^{(1)}, F^{(2)},\dots, F^{(n)})\) := \(E(Vol_L(X)\)) is strictly decreasing in
each argument, i.e. if \(F^{(i)}(t) \le G^{(i)}(t)t\), \(t \in {R}^+_0\), but \(F^{(i)} \not\equiv G^{(i)}\), we show \(\Phi\) \((\dots, F^{(i)}, \dots\)) > \(\Phi(\dots,G^{(i)},\dots\)). The proof is clone in the more general framework
of continuous and \(f\)- additive polytope functionals.

A polynomial function \(f : L \to L\) of a lattice \(\mathcal{L}\) = \((L; \land, \lor)\) is generated by the identity function id \(id(x)=x\) and the constant functions \(c_a (x) = a\) (for every \(x \in L\)), \(a \in L\) by applying the operations \(\land, \lor\) finitely often. Every polynomial function in one or also in several variables is a monotone function of \(\mathcal{L}\).
If every monotone function of \(\mathcal{L}\)is a polynomial function then \(\mathcal{L}\) is called orderpolynomially complete. In this paper we give a new characterization of finite order-polynomially lattices. We consider doubly irreducible monotone functions and point out their relation to tolerances, especially to central relations. We introduce chain-compatible lattices
and show that they have a non-trivial congruence if they contain a finite interval and an infinite chain. The consequences are two new results. A modular lattice \(\mathcal{L}\) with a finite interval is order-polynomially complete if and only if \(\mathcal{L}\) is finite projective geometry. If \(\mathcal{L}\) is simple modular lattice of infinite length then every nontrivial interval is of infinite length and has the same cardinality as any other nontrivial interval of \(\mathcal{L}\). In the last sections we show the descriptive power of polynomial functions of
lattices and present several applications in geometry.

Let \(a_1,\dots,a_m\) be independent random points in \(\mathbb{R}^n\) that are independent and identically distributed spherically symmetrical in \(\mathbb{R}^n\). Moreover, let \(X\) be the random polytope generated as the convex hull of \(a_1,\dots,a_m\) and let \(L_k\) be an arbitrary \(k\)-dimensional
subspace of \(\mathbb{R}^n\) with \(2\le k\le n-1\). Let \(X_k\) be the orthogonal projection image of \(X\) in \(L_k\). We call those vertices of \(X\), whose projection images in \(L_k\) are vertices of \(X_k\)as well shadow vertices of \(X\) with respect to the subspace \(L_k\) . We derive a distribution independent sharp upper bound for the expected number of shadow vertices of \(X\) in \(L_k\).