Filtern
Erscheinungsjahr
Dokumenttyp
- Dissertation (1251) (entfernen)
Sprache
- Deutsch (667)
- Englisch (583)
- Mehrsprachig (1)
Schlagworte
- Visualisierung (18)
- Simulation (16)
- Apoptosis (12)
- Katalyse (12)
- Finite-Elemente-Methode (11)
- Phasengleichgewicht (11)
- Stadtplanung (11)
- Mobilfunk (10)
- Modellierung (10)
- Eisen (9)
Fachbereich / Organisatorische Einheit
- Fachbereich Chemie (304)
- Fachbereich Maschinenbau und Verfahrenstechnik (228)
- Fachbereich Mathematik (225)
- Fachbereich Informatik (139)
- Fachbereich Biologie (99)
- Fachbereich ARUBI (71)
- Fachbereich Elektrotechnik und Informationstechnik (65)
- Fachbereich Bauingenieurwesen (38)
- Fachbereich Sozialwissenschaften (29)
- Fachbereich Raum- und Umweltplanung (21)
Der direkte Einstieg in die Phosphorchemie ist durch Reaktion einer Rohlösung von [Cp=Ru(CO)2H] (1) (Cp= = C5H3(SiMe3)2) mit P4 möglich. Dabei erhält man den Komplex [Cp=Ru(CO)2PH2] (6a) mit freier Phosphanidogruppe, der selbst nicht isoliert, dessen Existenz jedoch spektroskopisch und durch Folgereaktionen eindeutig belegt werden kann. Das freie Elektronenpaar des Phosphors kann sowohl durch [M(CO)5(thf)] (M = Cr, Mo, W) (7,8,9) als auch durch [Cp*Re(CO)2(thf)] (14) nach Verlust von deren thf-Liganden komplexiert werden. Das bei der Reaktion mit 14 als Nebenprodukt entstandene [{Cp*(OC)2Re}2(µ-CO)] (17) konnte erstmals röntgenstrukturanalytisch untersucht werden. Die Cothermolyse von [{Cp=Ru(CO)2}2] (4) mit P4 liefert [Cp=Ru(Eta5-P5)] (18) und [{Cp=Ru}2(µ-Eta;2:2-P2)2] (19), die beide röntgenstrukturanalytisch charakterisiert werden konnten. 18 erweist sich als erstes Pentaphosphametallocen, das eine annähernd ideal ekliptische Konformation aufweist. Im Zweikernkomplex 19 konnte ein Komplex der 8. Gruppe mit zwei separaten P2-Brückenliganden eindeutig belegt werden. Die Cothermolyse von 4 mit [Cp*Fe(Eta5-P5)] liefert neben 18 und 19 die Homo- und Heterotrimetallcluster [{Cp=Ru}n{Cp*Fe}3-nP5] (n = 1, 2, 3) mit verzerrt dreiecksdodekaedrischen Gerüststrukturen. Die Röntgenstrukturanalysen von [{Cp=Ru}3P5] (24), [{Cp=Ru}2{Cp*Fe}P5] (25) und [{Cp*Fe}2{Cp=Ru}P5] (26) zeigen, daß die Metallfragmente {Cp*Fe} und {Cp=Ru} nahezu ohne Auswirkungen auf die Gerüststruktur ausgetauscht werden können. Der Vergleich zu den Cp-Derivaten zeigt die annähernde Identität der Verbindungen. Bei 26 bleibt die Stellung des Cp=-Liganden am Ruthenium nach Einbau eines Eisenfragmentes erhalten, während in [{Cp*Fe}2{CpRu}P5] (26a) ein Wechsel in der Stellung des Cp-Liganden stattfindet. Die Photolyse von [Cp=Ru(CO)2H] (1) in THF ergibt die spektroskopisch identifizierten Verbindungen [{Cp=Ru(CO)}2(µ-H)2] (28), [{Cp=Ru(CO)}2{Cp=Ru(CO)H}] (29) und [{Cp=Ru}4(µ3-CO)4] (30). Das in geringen Mengen anfallende 28 konnte als dimere Struktur mit zwei verbrückenden Wasserstoffatomen identifiziert werden. Für Komplex 29 wurde eine Struktur als triangularer Metalldreiring postuliert, der den Verbindungen 33 und 34 sehr ähnlich ist. Bei 29 handelt es sich um einen 46 VE-Komplex mit einer Ru-Ru-Doppelbindung im Dreiring. Verbindung 30 mit tetraedrischem Ru4-Gerüst kann in hoher Ausbeute dargestellt werden. 30 ist luft- und wasserstabil und überaus reaktionsträge. Führt man die Photolyse von 1 in Hexan als nichtkoordinierendem Lösungsmittel durch, so erhält man neben 28 und 29 die Komplexe [{Cp=Ru(µ-CO)}2{Cp=Ru(CO)H}] (33), [{{{Cp=Ru}2(µ-CO)}{Cp=RuH}}(µ3-CO)2] (34) sowie die unbekannte Verbindung 35. Die Röntgenstrukturanalysen von 33 und 34 zeigen zwei mit 46 VE elektronendefiziente, triangulare Ru3-Cluster mit einer M-M-Doppelbindung im Ring, welche beide der Vorgabe der Magischen Zahlen entsprechen. 33 reagiert bereitwillig mit P4 und ergibt [{Cp=Ru}(µ-Eta4:1:1-P4){Ru(CO)Cp=}] (40). Die RSA von 40 zeigt, daß Ru1 um 25,3° aus der Ebene der vier Phosphoratome P1-P4 abgewinkelt und kein planares Tetraphospharuthenol verwirklicht ist. Die Aufsicht auf die P4Ru-Ebene läßt jedoch deutlich das Bestreben erkennen, die Konstitution einer pentagonalen Pyramide auszubilden.
Destructive diseases of the lung like lung cancer or fibrosis are still often lethal. Also in case of fibrosis in the liver, the only possible cure is transplantation.
In this thesis, we investigate 3D micro computed synchrotron radiation (SR\( \mu \)CT) images of capillary blood vessels in mouse lungs and livers. The specimen show so-called compensatory lung growth as well as different states of pulmonary and hepatic fibrosis.
During compensatory lung growth, after resecting part of the lung, the remaining part compensates for this loss by extending into the empty space. This process is accompanied by an active vessel growing.
In general, the human lung can not compensate for such a loss. Thus, understanding this process in mice is important to improve treatment options in case of diseases like lung cancer.
In case of fibrosis, the formation of scars within the organ's tissue forces the capillary vessels to grow to ensure blood supply.
Thus, the process of fibrosis as well as compensatory lung growth can be accessed by considering the capillary architecture.
As preparation of 2D microscopic images is faster, easier, and cheaper compared to SR\( \mu \)CT images, they currently form the basis of medical investigation. Yet, characteristics like direction and shape of objects can only properly be analyzed using 3D imaging techniques. Hence, analyzing SR\( \mu \)CT data provides valuable additional information.
For the fibrotic specimen, we apply image analysis methods well-known from material science. We measure the vessel diameter using the granulometry distribution function and describe the inter-vessel distance by the spherical contact distribution. Moreover, we estimate the directional distribution of the capillary structure. All features turn out to be useful to characterize fibrosis based on the deformation of capillary vessels.
It is already known that the most efficient mechanism of vessel growing forms small torus-shaped holes within the capillary structure, so-called intussusceptive pillars. Analyzing their location and number strongly contributes to the characterization of vessel growing. Hence, for all three applications, this is of great interest. This thesis provides the first algorithm to detect intussusceptive pillars in SR\( \mu \)CT images. After segmentation of raw image data, our algorithm works automatically and allows for a quantitative evaluation of a large amount of data.
The analysis of SR\( \mu \)CT data using our pillar algorithm as well as the granulometry, spherical contact distribution, and directional analysis extends the current state-of-the-art in medical studies. Although it is not possible to replace certain 3D features by 2D features without losing information, our results could be used to examine 2D features approximating the 3D findings reasonably well.
The various uses of fiber-reinforced composites, for example in the enclosures of planes, boats and cars, generates the demand for a detailed analysis of these materials. The final goal is to optimize fibrous materials by the means of “virtual material design”. New fibrous materials are virtually created as realizations of a stochastic model and evaluated with physical simulations. In that way, materials can be optimized for specific use cases, without constructing expensive prototypes or performing mechanical experiments. In order to design a practically fabricable material, the stochastic model is first adapted to an existing material and then slightly modified. The virtual reconstruction of the existing material requires a precise knowledge of the geometry of its microstructure. The first part of this thesis describes a fiber quantification method by the means of local measurements of the fiber radius and orientation. The combination of a sparse chord length transform and inertia moments leads to an efficient and precise new algorithm. It outperforms existing approaches with the possibility to treat different fiber radii within one sample, with high precision in continuous space and comparably fast computing time. This local quantification method can be directly applied on gray value images by adapting the directional distance transforms on gray values. In this work, several approaches of this kind are developed and evaluated. Further characterization of the fiber system requires a segmentation of each single fiber. Using basic morphological operators with specific structuring elements, it is possible to derive a probability for each pixel describing if the pixel belongs to a fiber core in a region without overlapping fibers. Tracking high probabilities leads to a partly reconstruction of the fiber cores in non crossing regions. These core parts are then reconnected over critical regions, if they fulfill certain conditions ensuring the affiliation to the same fiber. In the second part of this work, we develop a new stochastic model for dense systems of non overlapping fibers with a controllable level of bending. Existing approaches in the literature have at least one weakness in either achieving high volume fractions, producing non overlapping fibers, or controlling the bending or the orientation distribution. This gap can be bridged by our stochastic model, which operates in two steps. Firstly, a random walk with the multivariate von Mises-Fisher orientation distribution defines bent fibers. Secondly, a force-biased packing approach arranges them in a non overlapping configuration. Furthermore, we provide the estimation of all parameters needed for the fitting of this model to a real microstructure. Finally, we simulate the macroscopic behavior of different microstructures to derive their mechanical and thermal properties. This part is mostly supported by existing software and serves as a summary of physical simulation applied to random fiber systems. The application on a glass fiber reinforced polymer proves the quality of the reconstruction by our stochastic model, as the effective properties match for both the real microstructure and the realizations of the fitted model. This thesis includes all steps to successfully perform virtual material design on various data sets. With novel and efficient algorithms it contributes to the science of analysis and modeling of fiber reinforced materials.
Photolysiert man [{Cp R Ru(CO)2}2] (Cp R = Cp", Cp*) (1a,b), in Gegenwart von weißem Phosphor, so können keine phosphorhaltigen Produkte isoliert werden. Setzt man [{Cp"Ru(CO)2}2] (1a) mit weißem Phosphor bei 190 °C in Dekalin um, dann lassen sich als einzige Verbindungen das Pentaphospharuthenocen-Derivat [Cp"Ru(h 5 -P5)] (3a, 6 % Ausbeute) und [Cp"2Ru2P4] (4a, 17 % Ausbeute) säulenchromatographisch abtrennen. Für 4a wird auf Grund seiner spektroskopischen Eigenschaften eine den röntgenstruktur-analytisch charakterisierten pseudo-Tripeldecker-Komplexen [{Cp R Fe}2(micro-h 4:4 -P4)] (Cp R = Cp" [35] , Cp"' [11] ) analoge Struktur mit s-cis-Tetraphosphabutadiendiyl-"Mitteldeck" vorgeschlagen; es sind jedoch auch zwei micro-h 2:2 -P2-Liganden denkbar. Die Cothermolyse von [{Cp"Ru(CO)2}2] (1a) und [Cp*Fe(h 5 -P5)] (2b) ergibt ein breites Produktbild. Während [Cp"Ru(h 5 -P5)] (3a), das chromatographisch nicht von 2b abgetrennt werden konnte, und [Cp"2Ru2P4] (4a, 6 % Ausbeute) in vergleichsweise geringen Mengen entstehen, können [{Cp"Ru}3P5] (6) in 17 % Ausbeute, [{Cp"Ru}2{Cp*Fe}P5] (7) in 7 % Ausbeute, [{Cp*Fe}2{Cp"Ru}P5] (8) in 22 % Ausbeute und [{Cp"Ru}3{Cp*Fe}(P2)2] (9) in 14 % Ausbeute isoliert werden. Für 9 wird, basierend auf den NMR-spektroskopischen Befunden, eine zu [{CpFe}4(P2)2] (10) [24] analoge Struktur mit einem hier verzerrten Dreiecksdodekaedergerüst vorge-schlagen, dessen vier Ecken der Konnektivität fünf von drei Rutheniumatomen und einem Eisenatom besetzt sind und dessen vier Ecken der Konnektivität vier zwei micro-h 2:2:1:1 -P2-Einheiten einnehmen. Es handelt sich bei 9 wie bei 10 um Cluster vom hypercloso-Typ (n+1 = 8 GEP). [30,31] Röntgenstrukturanalysen zeigen, daß 6, 7 und 8 ebenfalls verzerrt dreiecksdodekaedrische Gerüststrukturen besitzen. Damit ist auch die Struktur der bereits früher synthetisierten und spektroskopisch charakterisierten Komplexe [{Cp R Fe}3P5] (Cp R = Cp*, Cp*') (11b,c) [8] geklärt, deren NMR-Daten auf eine enge Verwandtschaft insbesondere mit 6 hinweisen. Gegenüber 9 bzw. 10 [24] mit einem M4P4-Gerüst (M = allgem. Übergangsmetallatom) ist in den Clustern 6, 7, 8 und 11b,c [8] mit M3P5-Gerüst formal ein 13 VE-Metallkomplex-fragment (Konnektivität fünf; 1 GE) durch ein Phosphoratom (3 GE) ersetzt, wodurch der Übergang zum closo-Strukturtyp des Dreiecksdodekaeders (n+1 = 9 GEP) [30,31] vollzogen wird. Die Cluster 6, 7, 8 und 11b,c [8] enthalten eine bisher unbekannte Koordinationsform der P5-Einheit. Bei der thermischen Umsetzung von [{Cp*Ru(CO)2}2] (1b) mit [Cp*Ru(h 5 -P5)] (3b) erhält man als Hauptprodukt den Dreikernkomplex [{Cp*Ru}3(P4)(P)] (17b, 62 % Ausbeute). Als einziges Nebenprodukt kann [Cp*2Ru2P4] (4b, 5 % Ausbeute) säulen-chromatographisch isoliert werden. Die NMR-Daten von 17b lassen in Analogie zum röntgenstrukturanalytisch charakterisierten [{Cp*'Fe}3(P4)(P){Mo(CO)5}] (18c) [8] auf eine cubanartige Struktur schließen, in der die fünf Phosphoratome in Form einer Isotetraphosphid-Einheit und eines einzelnen Phosphoratoms vorliegen. Die Gesamtzahl von 64 Valenzelektronen ist im Einklang mit drei Metall-Metall-Bindungen [31] . Für 4b ist wie für das voranstehend besprochene Cp"-Derivat 4a eine Pseudo- Tripeldecker-Struktur mit s-cis-Tetraphosphabutadiendiyl-"Mitteldeck" oder mit zwei micro-h 2:2 -P2-Liganden zu diskutieren. Thermolysiert man [{Cp* Ru(CO)2}2] (1c) mit 3b, so erhält man die Komplexe [Cp*'nCp*2-nRu2P4] (n = 0,1,2) (4b,d,c) und [{Cp*'Ru}n{Cp*Ru}3-n(P4)(P)] (n = 1,2,3) (17e,d,c) jeweils als nicht auftrennbare Gemische von analog aufgebauten Verbindungen, die sich nur im Zahlenverhältnis der verschiedenen Cyclopentadienyl-Liganden Cp*' und Cp* unterscheiden. In geringem Umfang beobachtet man dabei auch eine cyclo-P5-Übertragung unter Bildung des literaturbekannten [Cp*'Ru(h 5 -P5)] [7,10] , das im Gemisch mit nicht abreagiertem 3b anfällt. Durch Umsetzung mit [W(CO)5(thf)] gelingt die Komplexierung des dreiecks-dodekaedrischen Clusters [{Cp*Fe}2{Cp"Ru}P5] (8) zum Monoaddukt [{Cp*Fe}2{Cp"Ru}P5{W(CO)5}] (19), während beim Komplexierungsversuch des ebenfalls dreiecksdodekaedrischen Clusters [{Cp*'Fe}3P5] (11c) mit [Mo(CO)5(thf)] [8] eine Gerüstumlagerung zur cubanartig aufgebauten Verbindung [{Cp*'Fe}3(P4)(P){Mo(CO)5}] (18c) erfolgt. Der Strukturvorschlag für 19 basiert auf dem 31 P-NMR-Spektrum. Versuche zur Oxidation von 8 mit gelbem Schwefel bei Raumtemperatur führen zu unspezifischer Zersetzung bzw. Folgereaktionen von 8. Orientierende Versuche mit grauem Selen als milderem Oxidationsmittel deuten darauf hin, daß unter geeigneten Reaktionsbedingungen eine einfache Selenierung am endständigen Phosphoratom des P5-Liganden von 8 erfolgt. Bei der Umsetzung von [{Cp"Ru}3{Cp*Fe}(P2)2] (9) mit gelbem Schwefel können je nach Reaktionsbedingungen bis zu drei Phosphoratome oxidiert werden. Vollständige Sulfurierung wie im Falle von [{CpFe}4(P2S2)2] [24] wird nicht beobachtet. Die Sulfurierung ist regioselektiv. Für einen bestimmten Sulfurierungsgrad wird jeweils nur ein Produkt erhalten. Die Strukturvorschläge für die Cluster 21 23 werden anhand ihrer 31 P-NMR-spektroskopischen Daten abgeleitet.
A Consistent Large Eddy Approach for Lattice Boltzmann Methods and its Application to Complex Flows
(2015)
Lattice Boltzmann Methods have shown to be promising tools for solving fluid flow problems. This is related to the advantages of these methods, which are among others, the simplicity in handling complex geometries and the high efficiency in calculating transient flows. Lattice Boltzmann Methods are mesoscopic methods, based on discrete particle dynamics. This is in contrast to conventional Computational Fluid Dynamics methods, which are based on the solution of the continuum equations. Calculations of turbulent flows in engineering depend in general on modeling, since resolving of all turbulent scales is and will be in near future far beyond the computational possibilities. One of the most auspicious modeling approaches is the large eddy simulation, in which the large, inhomogeneous turbulence structures are directly computed and the smaller, more homogeneous structures are modeled.
In this thesis, a consistent large eddy approach for the Lattice Boltzmann Method is introduced. This large eddy model includes, besides a subgrid scale model, appropriate boundary conditions for wall resolved and wall modeled calculations. It also provides conditions for turbulent domain inlets. For the case of wall modeled simulations, a two layer wall model is derived in the Lattice Boltzmann context. Turbulent inlet conditions are achieved by means of a synthetic turbulence technique within the Lattice Boltzmann Method.
The proposed approach is implemented in the Lattice Boltzmann based CFD package SAM-Lattice, which has been created in the course of this work. SAM-Lattice is feasible of the calculation of incompressible or weakly compressible, isothermal flows of engineering interest in complex three dimensional domains. Special design targets of SAM-Lattice are high automatization and high performance.
Validation of the suggested large eddy Lattice Boltzmann scheme is performed for pump intake flows, which have not yet been treated by LBM. Even though, this numerical method is very suitable for this kind of vortical flows in complicated domains. In general, applications of LBM to hydrodynamic engineering problems are rare. The results of the pump intake validation cases reveal that the proposed numerical approach is able to represent qualitatively and quantitatively the very complex flows in the intakes. The findings provided in this thesis can serve as the basis for a broader application of LBM in hydrodynamic engineering problems.
The growing computational power enables the establishment of the Population Balance Equation (PBE)
to model the steady state and dynamic behavior of multiphase flow unit operations. Accordingly, the twophase
flow
behavior inside liquid-liquid extraction equipment is characterized by different factors. These
factors include: interactions among droplets (breakage and coalescence), different time scales due to the
size distribution of the dispersed phase, and micro time scales of the interphase diffusional mass transfer
process. As a result of this, the general PBE has no well known analytical solution and therefore robust
numerical solution methods with low computational cost are highly admired.
In this work, the Sectional Quadrature Method of Moments (SQMOM) (Attarakih, M. M., Drumm, C.,
Bart, H.-J. (2009). Solution of the population balance equation using the Sectional Quadrature Method of
Moments (SQMOM). Chem. Eng. Sci. 64, 742-752) is extended to take into account the continuous flow
systems in spatial domain. In this regard, the SQMOM is extended to solve the spatially distributed
nonhomogeneous bivariate PBE to model the hydrodynamics and physical/reactive mass transfer
behavior of liquid-liquid extraction equipment. Based on the extended SQMOM, two different steady
state and dynamic simulation algorithms for hydrodynamics and mass transfer behavior of liquid-liquid
extraction equipment are developed and efficiently implemented. At the steady state modeling level, a
Spatially-Mixed SQMOM (SM-SQMOM) algorithm is developed and successfully implemented in a onedimensional
physical spatial domain. The integral spatial numerical flux is closed using the mean mass
droplet diameter based on the One Primary and One Secondary Particle Method (OPOSPM which is the
simplest case of the SQMOM). On the other hand the hydrodynamics integral source terms are closed
using the analytical Two-Equal Weight Quadrature (TEqWQ). To avoid the numerical solution of the
droplet rise velocity, an analytical solution based on the algebraic velocity model is derived for the
particular case of unit velocity exponent appearing in the droplet swarm model. In addition to this, the
source term due to mass transport is closed using OPOSPM. The resulting system of ordinary differential
equations with respect to space is solved using the MATLAB adaptive Runge–Kutta method (ODE45). At
the dynamic modeling level, the SQMOM is extended to a one-dimensional physical spatial domain and
resolved using the finite volume method. To close the mathematical model, the required quadrature nodes
and weights are calculated using the analytical solution based on the Two Unequal Weights Quadrature
(TUEWQ) formula. By applying the finite volume method to the spatial domain, a semi-discreet ordinary
differential equation system is obtained and solved. Both steady state and dynamic algorithms are
extensively validated at analytical, numerical, and experimental levels. At the numerical level, the
predictions of both algorithms are validated using the extended fixed pivot technique as implemented in
PPBLab software (Attarakih, M., Alzyod, S., Abu-Khader, M., Bart, H.-J. (2012). PPBLAB: A new
multivariate population balance environment for particulate system modeling and simulation. Procedia
Eng. 42, pp. 144-562). At the experimental validation level, the extended SQMOM is successfully used
to model the steady state hydrodynamics and physical and reactive mass transfer behavior of agitated
liquid-liquid extraction columns under different operating conditions. In this regard, both models are
found efficient and able to follow liquid extraction column behavior during column scale-up, where three
column diameters were investigated (DN32, DN80, and DN150). To shed more light on the local
interactions among the contacted phases, a reduced coupled PBE and CFD framework is used to model
the hydrodynamic behavior of pulsed sieve plate columns. In this regard, OPOSPM is utilized and
implemented in FLUENT 18.2 commercial software as a special case of the SQMOM. The dropletdroplet
interactions
(breakage
and
coalescence)
are
taken
into
account
using
OPOSPM,
while
the
required
information
about
the
velocity
field
and
energy
dissipation
is
calculated
by
the
CFD
model.
In
addition
to
this,
the proposed coupled OPOSPM-CFD framework is extended to include the mass transfer. The
proposed framework is numerically tested and the results are compared with the published experimental
data. The required breakage and coalescence parameters to perform the 2D-CFD simulation are estimated
using PPBLab software, where a 1D-CFD simulation using a multi-sectional gird is performed. A very
good agreement is obtained at the experimental and the numerical validation levels.
A prime motivation for using XML to directly represent pieces of information is the ability of supporting ad-hoc or 'schema-later' settings. In such scenarios, modeling data under loose data constraints is essential. Of course, the flexibility of XML comes at a price: the absence of a rigid, regular, and homogeneous structure makes many aspects of data management more challenging. Such malleable data formats can also lead to severe information quality problems, because the risk of storing inconsistent and incorrect data is greatly increased. A prominent example of such problems is the appearance of the so-called fuzzy duplicates, i.e., multiple and non-identical representations of a real-world entity. Similarity joins correlating XML document fragments that are similar can be used as core operators to support the identification of fuzzy duplicates. However, similarity assessment is especially difficult on XML datasets because structure, besides textual information, may exhibit variations in document fragments representing the same real-world entity. Moreover, similarity computation is substantially more expensive for tree-structured objects and, thus, is a serious performance concern. This thesis describes the design and implementation of an effective, flexible, and high-performance XML-based similarity join framework. As main contributions, we present novel structure-conscious similarity functions for XML trees - either considering XML structure in isolation or combined with textual information -, mechanisms to support the selection of relevant information from XML trees and organization of this information into a suitable format for similarity calculation, and efficient algorithms for large-scale identification of similar, set-represented objects. Finally, we validate the applicability of our techniques by integrating our framework into a native XML database management system; in this context we address several issues around the integration of similarity operations into traditional database architectures.
This thesis presents a novel, generic framework for information segmentation in document images.
A document image contains different types of information, for instance, text (machine printed/handwritten), graphics, signatures, and stamps.
It is necessary to segment information in documents so that to process such segmented information only when required in automatic document processing workflows.
The main contribution of this thesis is the conceptualization and implementation of an information segmentation framework that is based on part-based features.
The generic nature of the presented framework makes it applicable to a variety of documents (technical drawings, magazines, administrative, scientific, and academic documents) digitized using different methods (scanners, RGB cameras, and hyper-spectral imaging (HSI) devices).
A highlight of the presented framework is that it does not require large training sets, rather a few training samples (for instance, four pages) lead to high performance, i.e., better than previously existing methods.
In addition, the presented framework is simple and can be adapted quickly to new problem domains.
This thesis is divided into three major parts on the basis of document digitization method (scanned, hyper-spectral imaging, and camera captured) used.
In the area of scanned document images, three specific contributions have been realized.
The first of them is in the domain of signature segmentation in administrative documents.
In some workflows, it is very important to check the document authenticity before processing the actual content.
This can be done based on the available seal of authenticity, e.g., signatures.
However, signature verification systems expect pre-segmented signature image, while signatures are usually a part of document.
To use signature verification systems on document images, it is necessary to first segment signatures in documents.
This thesis shows that the presented framework can be used to segment signatures in administrative documents.
The system based on the presented framework is tested on a publicly available dataset where it outperforms the state-of-the-art methods and successfully segmented all signatures, while less than half of the found signatures are false positives.
This shows that it can be applied for practical use.
The second contribution in the area of scanned document images is segmentation of stamps in administrative documents.
A stamp also serves as a seal for documents authenticity.
However, the location of stamp on the document can be more arbitrary than a signature depending on the person sealing the document.
This thesis shows that a system based on our generic framework is able to extract stamps of any arbitrary shape and color.
The evaluation of the presented system on a publicly available dataset shows that it is also able to segment black stamps (that were not addressed in the past) with a recall and precision of 83% and 73%, respectively.
%Furthermore, to segment colored stamps, this thesis presents a novel feature set which is based on intensity gradient, is able to extract unseen, colored, arbitrary shaped, textual as well as graphical stamps, and outperforms the state-of-the-art methods.
The third contribution in the scanned document images is in the domain of information segmentation in technical drawings (architectural floorplans, maps, circuit diagrams, etc.) containing usually a large amount of graphics and comparatively less textual components. Further, as in technical drawings, text is overlapping with graphics.
Thus, automatic analysis of technical drawings uses text/graphics segmentation as a pre-processing step.
This thesis presents a method based on our generic information segmentation framework that is able to detect the text, which is touching graphical components in architectural floorplans and maps.
Evaluation of the method on a publicly available dataset of architectural floorplans shows that it is able to extract almost all touching text components with precision and recall of 71% and 95%, respectively.
This means that almost all of the touching text components are successfully extracted.
In the area of hyper-spectral document images, two contributions have been realized.
Unlike normal three channels RGB images, hyper-spectral images usually have multiple channels that range from ultraviolet to infrared regions including the visible region.
First, this thesis presents a novel automatic method for signature segmentation from hyper-spectral document images (240 spectral bands between 400 - 900 nm).
The presented method is based on a part-based key point detection technique, which does not use any structural information, but relies only on the spectral response of the document regardless of ink color and intensity.
The presented method is capable of segmenting (overlapping and non-overlapping) signatures from varying backgrounds like, printed text, tables, stamps, logos, etc.
Importantly, the presented method can extract signature pixels and not just the bounding boxes.
This is substantial when signatures are overlapping with text and/or other objects in image. Second, this thesis presents a new dataset comprising of 300 documents scanned using a high-resolution hyper-spectral scanner. Evaluation of the presented signature segmentation method on this hyper-spectral dataset shows that it is able to extract signature pixels with the precision and recall of 100% and 79%, respectively.
Further contributions have been made in the area of camera captured document images. A major problem in the development of Optical Character Recognition (OCR) systems for camera captured document images is the lack of labeled camera captured document images datasets. In the first place, this thesis presents a novel, generic, method for automatic ground truth generation/labeling of document images. The presented method builds large-scale (i.e., millions of images) datasets of labeled camera captured / scanned documents without any human intervention. The method is generic and can be used for automatic ground truth generation of (scanned and/or camera captured) documents in any language, e.g., English, Russian, Arabic, Urdu. The evaluation of the presented method, on two different datasets in English and Russian, shows that 99.98% of the images are correctly labeled in every case.
Another important contribution in the area of camera captured document images is the compilation of a large dataset comprising 1 million word images (10 million character images), captured in a real camera-based acquisition environment, along with the word and character level ground truth. The dataset can be used for training as well as testing of character recognition systems for camera-captured documents. Various benchmark tests are performed to analyze the behavior of different open source OCR systems on camera captured document images. Evaluation results show that the existing OCRs, which already get very high accuracies on scanned documents, fail on camera captured document images.
Using the presented camera-captured dataset, a novel character recognition system is developed which is based on a variant of recurrent neural networks, i.e., Long Short Term Memory (LSTM) that outperforms all of the existing OCR engines on camera captured document images with an accuracy of more than 95%.
Finally, this thesis provides details on various tasks that have been performed in the area closely related to information segmentation. This includes automatic analysis and sketch based retrieval of architectural floor plan images, a novel scheme for online signature verification, and a part-based approach for signature verification. With these contributions, it has been shown that part-based methods can be successfully applied to document image analysis.
For many years real-time task models have focused the timing constraints on execution windows defined by earliest start times and deadlines for feasibility.
However, the utility of some application may vary among scenarios which yield correct behavior, and maximizing this utility improves the resource utilization.
For example, target sensitive applications have a target point where execution results in maximized utility, and an execution window for feasibility.
Execution around this point and within the execution window is allowed, albeit at lower utility.
The intensity of the utility decay accounts for the importance of the application.
Examples of such applications include multimedia and control; multimedia application are very popular nowadays and control applications are present in every automated system.
In this thesis, we present a novel real-time task model which provides for easy abstractions to express the timing constraints of target sensitive RT applications: the gravitational task model.
This model uses a simple gravity pendulum (or bob pendulum) system as a visualization model for trade-offs among target sensitive RT applications.
We consider jobs as objects in a pendulum system, and the target points as the central point.
Then, the equilibrium state of the physical problem is equivalent to the best compromise among jobs with conflicting targets.
Analogies with well-known systems are helpful to fill in the gap between application requirements and theoretical abstractions used in task models.
For instance, the so-called nature algorithms use key elements of physical processes to form the basis of an optimization algorithm.
Examples include the knapsack problem, traveling salesman problem, ant colony optimization, and simulated annealing.
We also present a few scheduling algorithms designed for the gravitational task model which fulfill the requirements for on-line adaptivity.
The scheduling of target sensitive RT applications must account for timing constraints, and the trade-off among tasks with conflicting targets.
Our proposed scheduling algorithms use the equilibrium state concept to order the execution sequence of jobs, and compute the deviation of jobs from their target points for increased system utility.
The execution sequence of jobs in the schedule has a significant impact on the equilibrium of jobs, and dominates the complexity of the problem --- the optimum solution is NP-hard.
We show the efficacy of our approach through simulations results and 3 target sensitive RT applications enhanced with the gravitational task model.
Numerical Godeaux surfaces are minimal surfaces of general type with the smallest possible numerical invariants. It is known that the torsion group of a numerical Godeaux surface is cyclic of order \(m\leq 5\). A full classification has been given for the cases \(m=3,4,5\) by the work of Reid and Miyaoka. In each case, the corresponding moduli space is 8-dimensional and irreducible.
There exist explicit examples of numerical Godeaux surfaces for the orders \(m=1,2\), but a complete classification for these surfaces is still missing.
In this thesis we present a construction method for numerical Godeaux surfaces which is based on homological algebra and computer algebra and which arises from an experimental approach by Schreyer. The main idea is to consider the canonical ring \(R(X)\) of a numerical Godeaux surface \(X\) as a module over some graded polynomial ring \(S\). The ring \(S\) is chosen so that \(R(X)\) is finitely generated as an \(S\)-module and a Gorenstein \(S\)-algebra of codimension 3. We prove that the canonical ring of any numerical Godeaux surface, considered as an \(S\)-module, admits a minimal free resolution whose middle map is alternating. Moreover, we show that a partial converse of this statement is true under some additional conditions.
Afterwards we use these results to construct (canonical rings of) numerical Godeaux surfaces. Hereby, we restrict our study to surfaces whose bicanonical system has no fixed component but 4 distinct base points, in the following referred to as marked numerical Godeaux surfaces.
The particular interest of this thesis lies on marked numerical Godeaux surfaces whose torsion group is trivial. For these surfaces we study the fibration of genus 4 over \(\mathbb{P}^1\) induced by the bicanonical system. Catanese and Pignatelli showed that the general fibre is non-hyperelliptic and that the number \(\tilde{h}\) of hyperelliptic fibres is bounded by 3. The two explicit constructions of numerical Godeaux surfaces with a trivial torsion group due to Barlow and Craighero-Gattazzo, respectively, satisfy \(\tilde{h} = 2\).
With the method from this thesis, we construct an 8-dimensional family of numerical Godeaux surfaces with a trivial torsion group and whose general element satisfy \(\tilde{h}=0\).
Furthermore, we establish a criterion for the existence of hyperelliptic fibres in terms of a minimal free resolution of \(R(X)\). Using this criterion, we verify experimentally the
existence of a numerical Godeaux surface with \(\tilde{h}=1\).