Kaiserslautern - Fachbereich Informatik
Refine
Year of publication
- 1999 (267)
- 1996 (50)
- 1994 (49)
- 1995 (48)
- 1998 (38)
- 1997 (35)
- 2021 (27)
- 2016 (25)
- 2019 (24)
- 2022 (24)
- 1993 (22)
- 2015 (22)
- 2023 (22)
- 2001 (21)
- 2020 (21)
- 2007 (19)
- 2013 (18)
- 2018 (18)
- 2002 (17)
- 2003 (16)
- 2014 (15)
- 2012 (14)
- 1992 (13)
- 2000 (13)
- 2004 (12)
- 2006 (11)
- 2009 (11)
- 2008 (9)
- 2005 (8)
- 2017 (8)
- 1991 (7)
- 2010 (7)
- 2024 (7)
- 2011 (5)
- 1979 (2)
- 1980 (1)
- 1983 (1)
- 1990 (1)
Document Type
- Preprint (346)
- Doctoral Thesis (235)
- Report (139)
- Article (134)
- Master's Thesis (45)
- Study Thesis (13)
- Conference Proceeding (8)
- Bachelor Thesis (4)
- Part of a Book (2)
- Habilitation (2)
Has Fulltext
- yes (928)
Keywords
- AG-RESY (64)
- PARO (31)
- Case-Based Reasoning (20)
- Visualisierung (20)
- SKALP (16)
- CoMo-Kit (15)
- Fallbasiertes Schliessen (12)
- RODEO (12)
- Robotik (12)
- HANDFLEX (11)
- META-AKAD (9)
- Robotics (8)
- Visualization (8)
- Deep Learning (7)
- Abstraction (6)
- Case Based Reasoning (6)
- motion planning (6)
- COMOKIT (5)
- Case-Based Planning (5)
- Computergraphik (5)
- Evaluation (5)
- Fallbasiertes Schließen (5)
- Force-Torque (5)
- RONAF (5)
- SDL (5)
- SIMERO (5)
- case-based problem solving (5)
- industrial robots (5)
- parallel processing (5)
- Assembly (4)
- Dienstgüte (4)
- Implementierung (4)
- Java 2 Enterprise Edition (4)
- Knowledge Acquisition (4)
- LOADBAL (4)
- Manipulation skills (4)
- Maschinelles Lernen (4)
- Mensch-Roboter-Kooperation (4)
- Wissensverarbeitung (4)
- case-based reasoning (4)
- deformable objects (4)
- resolution (4)
- safe human robot cooperation (4)
- search algorithms (4)
- Artificial Intelligence (3)
- Bildverarbeitung (3)
- CODET (3)
- Expertensysteme (3)
- Fallbasiertes Planen (3)
- Formalisierung (3)
- Geoinformationssystem (3)
- Internet (3)
- Java (3)
- Knowledge acquisition (3)
- Kooperation (3)
- Machine Learning (3)
- Model Checking (3)
- Mustererkennung (3)
- Navigation (3)
- Optimierung (3)
- Optische Zeichenerkennung (3)
- Ray casting (3)
- Recommender Systems (3)
- Requirements/Specifications (3)
- Roboter (3)
- Scientific Visualization (3)
- Semantic Web (3)
- Simulation (3)
- Software Engineering (3)
- Unsicherheit (3)
- artificial intelligence (3)
- case-based planning (3)
- computer aided planning (3)
- computer graphics (3)
- distributed software development (3)
- distributed software development process (3)
- document analysis (3)
- explanation-based learning (3)
- machine learning (3)
- on-line algorithms (3)
- optical character recognition (3)
- path planning (3)
- problem solving (3)
- reuse (3)
- theorem prover (3)
- theorem proving (3)
- verification (3)
- vibration (3)
- visualization (3)
- AKLEON (2)
- Activity recognition (2)
- Ad-hoc-Netz (2)
- Algorithmus (2)
- Automation (2)
- Automatische Differentiation (2)
- Bahnplanung (2)
- CAP (2)
- CAPlan (2)
- CIM-OSA (2)
- Code Generation (2)
- Computational Fluid Dynamics (2)
- Datenanalyse (2)
- Datenbank (2)
- Decision Trees (2)
- Deduction (2)
- Deformable Objects (2)
- Diagnose technischer Systeme (2)
- Distributed Software Development (2)
- Dreidimensionale Bildverarbeitung (2)
- Effizienter Algorithmus (2)
- Eingebettetes System (2)
- Endlicher Automat (2)
- Entscheidungsbäume (2)
- Experiment (2)
- Fallbasierte Planning (2)
- Fallbasierte Planung (2)
- Fräsen (2)
- GPU (2)
- HOT (2)
- Hals-Nasen-Ohren-Chirurgie (2)
- Hals-Nasen-Ohren-Heilkunde (2)
- Interaction (2)
- Interaktion (2)
- KLUEDO (2)
- Kommunikation (2)
- Kontextbezogenes System (2)
- Layout (2)
- MOLTKE-Projekt (2)
- Machine learning (2)
- Manipulation (2)
- Mapping (2)
- Mensch-Maschine-Kommunikation (2)
- Mensch-Roboter-Koexistenz (2)
- Merkmalsextraktion (2)
- Modellgetriebene Entwicklung (2)
- Modellierung (2)
- Natural Language Processing (2)
- Network Calculus (2)
- Network Protocols (2)
- Netzwerk (2)
- PATDEX (2)
- Partial functions (2)
- Pattern Recognition (2)
- Prozesssteuerung (2)
- Raumakustik (2)
- Regelung (2)
- Room acoustics (2)
- Routing (2)
- SWEEPING (2)
- Scattered-Data-Interpolation (2)
- Server (2)
- Software (2)
- Software Agents (2)
- Software engineering (2)
- Softwareentwicklung (2)
- Sprachprofile (2)
- Suchraum (2)
- Systemarchitektur (2)
- TOVE (2)
- Term rewriting systems (2)
- Topology (2)
- Translation Validation (2)
- UML (2)
- UML 2 (2)
- Uncertain Data (2)
- Uncertainty Visualization (2)
- Visual Analytics (2)
- Voronoi-Diagramm (2)
- Wearable computing (2)
- WiFi (2)
- Wissensakquisition (2)
- World Wide Web (2)
- XML (2)
- analogy (2)
- analysis of algorithms (2)
- application (2)
- artificial neural networks (2)
- automated theorem proving (2)
- automotive (2)
- building automation (2)
- case based reasoning (2)
- combined systems with sha (2)
- computer vision (2)
- computer-supported cooperative work (2)
- confluence (2)
- conservative extension (2)
- consistency (2)
- deep learning (2)
- design patterns (2)
- discretization (2)
- disjoint union (2)
- distributed (2)
- distributed computing (2)
- experience base (2)
- fallbasiertes Schliessen (2)
- formal specification (2)
- frames (2)
- genetic algorithms (2)
- graph search (2)
- human robot cooperation (2)
- image processing (2)
- innermost termination (2)
- layout analysis (2)
- learning system (2)
- load balancing (2)
- many-valued logic (2)
- mobile robots (2)
- modularity (2)
- parallel algorithms (2)
- problem formulation (2)
- regelbasiertes Problemlösen (2)
- requirements engineering (2)
- self-organization (2)
- semantic web (2)
- sensor fusion (2)
- tactics (2)
- temporal logic (2)
- termination (2)
- trajectory planning (2)
- virtual acoustics (2)
- weak termination (2)
- 3D Gene Expression (1)
- 3D Point Data (1)
- ASM (1)
- AUTOSAR (1)
- Ablagestruktur (1)
- Ableitungsschätzung (1)
- Abrechnungsmanagement (1)
- Abstandsregeltempomat (1)
- Abstraction-Based Controller Design (1)
- Abstraktion (1)
- Accelerometer (1)
- Access Points (1)
- Access System (1)
- Accounting (1)
- Accounting Agent (1)
- Ad-hoc networks (1)
- Ad-hoc workflow (1)
- Adaption (1)
- Adaptive Data Structure (1)
- Affine Arithmetic (1)
- Agents (1)
- Akquisition (1)
- Algorithm (1)
- Algorithmic Differentiation (1)
- Ambient Intelligence (1)
- Amharic, Attention, Factored Convolutional Neural Network, OCR (1)
- Anfrageverarbeitung (1)
- Anomaly Detection (1)
- Application Framework (1)
- Aspektorientierte Programmierung (1)
- Association (1)
- Audiodeskription (1)
- Ausdrucksfähig (1)
- Ausdrucksfähigkeit (1)
- Automat <Automatentheorie> (1)
- Automated Calibration (1)
- Automated Reasoning (1)
- Automated theorem proving (1)
- Automatic Differentiation (1)
- Automatic Image Captioning (1)
- Automatic Theorem Provi (1)
- Automatische Indexierung (1)
- Automatische Klassifikation (1)
- Automatische Messung (1)
- Automatisches Beweisverfahren (1)
- Automotive (1)
- Autonomer Agent (1)
- Autonomer Roboter (1)
- Autonomous mobile robots (1)
- Autorensystem (1)
- Bebauungsplanung (1)
- Befahrbarkeitsanalyse (1)
- Benutzer (1)
- Beschränkte Arithmetik (1)
- Bewegungsplanung (1)
- Bibligraphic References (1)
- Bibliothekskatalog (1)
- Bio-inspired (1)
- Bioinformatik (1)
- Bipedal Locomotion (1)
- Bitlisten (1)
- Blackboard architecture (1)
- Bluetooth (1)
- Boosting (1)
- C (1)
- CAN-Bus (1)
- CAR <Roboter> (1)
- CAS (1)
- CBR (1)
- CNC-Maschine (1)
- CUDA (1)
- Caching (1)
- Carbon footprint (1)
- Case Study (1)
- Case Study Erfahrungsdatenbank (1)
- Case-Based Classification Algorithms (1)
- Case-Based Diagnosis (1)
- Case-Based Learning (1)
- Case-Based Reasoning Systems (1)
- Case-Based Representability (1)
- Case-based problem solving (1)
- Castor (1)
- Causal Ordering (1)
- Causality (1)
- Certifying Compilers (1)
- Channel Hopping (1)
- Channel Scheduling (1)
- Channel hopping (1)
- Channel sensing (1)
- Chromium Browser (1)
- Classification (1)
- Classification Tasks (1)
- Classification of biomedical signals (1)
- Cluster-Analyse (1)
- Clustering (1)
- Cochlea Implant (1)
- Cochlea Implantat (1)
- Cognitive Amplification (1)
- Collaboration (1)
- Combinatorial Testing (1)
- Compiler (1)
- Completion (1)
- Computer Graphic (1)
- Computer Science (1)
- Computer Supported Cooperative Work (1)
- Computer assistierte Chirurgie (1)
- Computer assistierte Chirurgie (CAS) (1)
- Computer graphics (1)
- Computer supported cooperative work (1)
- Computer-Aided Diagnosis (1)
- Computerphysik (1)
- Computersimulation (1)
- Computertomographie (1)
- Computerunterstütztes Lernen (1)
- Computervision (1)
- Concept mapping (1)
- Concept maps (1)
- Concurrent data structures (1)
- Congitive Radio Networks (1)
- Constraint Graphs (1)
- Containertypen (1)
- Containertypes (1)
- Content Management (1)
- Context of Use (1)
- Context-sensitive Assistance (1)
- Contract net (1)
- Controller Synthesis (1)
- Cooperative decision making (1)
- Coq (1)
- Correct-by-Design Controller Synthesis (1)
- Correlation (1)
- Cyber-Physical Systems (1)
- Cycle Accuracy (1)
- DCE <Programm> (1)
- DES (1)
- DFG (1)
- DPN (1)
- Data Modeling (1)
- Dataset (1)
- Datenbanken (1)
- Datenfusion (1)
- Datenreduktion (1)
- Decision Support Systems (1)
- Declarative and Procedural Knowledge (1)
- Delaunay-Triangulierung (1)
- Dependency Factors (1)
- Derivative Estimation (1)
- Design Rationales (1)
- Diagnosesystem (1)
- Didaktik (1)
- Dienstschnittstellen (1)
- Difference Reduction (1)
- Direct Numerical Simulation (1)
- Discrete Event Simulation (DES) (1)
- Distributed Computation (1)
- Distributed Deb (1)
- Distributed Multimedia Applications (1)
- Distributed Rendering (1)
- Distributed Software Development Projects (1)
- Distributed System (1)
- Distributed software development support (1)
- Distributed system (1)
- Distributed systems (1)
- Distribution and Combination of Theorem Provers (1)
- DoS (1)
- DoS-Attacke (1)
- Domänenspezifische Sprachen (1)
- Doppler-Radar (1)
- Drahtloses lokales Netz (1)
- Dreidimensionale Rekonstruktion (1)
- Dreidimensionale Strömung (1)
- Dublin Core (1)
- Duplicate Identification (1)
- Duplikaterkennung (1)
- Dynamischer Test (1)
- E-Learning (1)
- EBG (1)
- ECMAScript (1)
- ESTELLE (1)
- Earley-Parser (1)
- Ecommerce (1)
- Effizienz (1)
- Effizienzsteigerung (1)
- Eingebettete Systeme (1)
- Elektrizitätsverbrauch (1)
- Embedded software (1)
- Empfehlungssysteme (1)
- Energie (1)
- Engineering 4.0 (1)
- Ensemble Visualization (1)
- Enterprise modeling (1)
- Enterprise modelling (1)
- Entwurf (1)
- Equality reasoning (1)
- Equational Reasoning (1)
- Ernergie effiziente Bewegung (1)
- Erstellung von Expertensystemen (1)
- Eventual consistency (1)
- Evolutionary Algorithm (1)
- Experience Base (1)
- Experience Database (1)
- Experimentation (1)
- Explainability (1)
- Explainable Artificial Intelligence (1)
- Exposed Datapath Architectures (1)
- Expressiveness (1)
- Extensibility (1)
- Extraction (1)
- Extrapolation (1)
- Eyewear Computing (1)
- FERAL (1)
- Fachdidaktik (1)
- Fahrassistenzsysteme (1)
- Fahrtkostenmodelle (1)
- Fallbasierte Diagnose (1)
- Fallstu (1)
- Farbmodell (1)
- Fast Mode-Signaling (1)
- Fault Tree Analysis (1)
- Feasibility study (1)
- Feature (1)
- Feature Detection (1)
- Feature Extraction (1)
- Fehlerbaumanalyse (1)
- Fließanalyse (1)
- Flooding Attack (1)
- Flow Visualization (1)
- Force Feedback (1)
- Forgetting-enabled Information Systems (1)
- Formal Semantics (1)
- Formale Beschreibungstechnik (1)
- Formale Grammatik (1)
- Formale Sprache (1)
- Formaler Beweis (1)
- Framework (1)
- Framework <Informatik> (1)
- Frequenzsprungverfahren (1)
- Functional Programs (1)
- Functional Safety (1)
- Funknetz (1)
- Funktionale Sicherheit (1)
- Fusion (1)
- Future Internet (1)
- GRAPHICS (1)
- Gateway (1)
- Gauß-Filter (1)
- Gebäude (1)
- Gebäudeautomation (1)
- Gefahren- und Risikoanalyse (1)
- Gene expression programming (1)
- General Knowledge (1)
- Generic Methods (1)
- Generierung (1)
- Genetic Algorithm (1)
- Geovisualization (1)
- Global Predicate Detection (1)
- Global Software Highway (1)
- Google Earth (1)
- HOL (1)
- HPC (1)
- HTE (1)
- HTML (1)
- Harvest (1)
- Hazard Analysis (1)
- Heterogeneous (1)
- Hitting families (1)
- Hub-and-Spoke-System (1)
- Huffman (1)
- Huffman-Code (1)
- Human Pose (1)
- Human-Computer Interaction (1)
- Human-Robot-Coexistence (1)
- Human-Robot-Cooperation (1)
- Hybrid CBR (1)
- Hybridregler (1)
- Hypergraph (1)
- Hörgerät (1)
- IDEA (1)
- IEEE 802.15.4 (1)
- IMU (1)
- INRECA (1)
- IP Address (1)
- IP Traffic Accounting (1)
- IP-XACT (1)
- ISO 26262 (1)
- ISO26262 (1)
- Image Processing (1)
- Imote2 (1)
- Implan (1)
- Implantation (1)
- Implementation (1)
- Incremental recomputation (1)
- Indexierung <Inhaltserschließung> (1)
- Inductive Theorem Proving (1)
- Induktive Logische Programmierung (1)
- Induktive logische Programmierung (1)
- Induktivem Schliessen (1)
- Induktiven Logischen Programmierung (1)
- Information Extraction (1)
- Information Management (1)
- Information Visualization (1)
- Instance-based Learning (1)
- Intel XScale (1)
- Intelligent Agents (1)
- Intelligent Object Fusion (1)
- Intelligent agents (1)
- Intensity estimation (1)
- Intentional Forgetting (1)
- Interleaved Planning (1)
- Intermediate Composition (1)
- Internet Based Software Process Management Environment (1)
- Internet knowledge base (1)
- Internet knowledge reuse (1)
- Interpolation (1)
- Invariante (1)
- Invariante Momente (1)
- Isabelle/HOL (1)
- JSR 170 JCR (1)
- Jacobian (1)
- JavaScript (1)
- Kellerautomat (1)
- Klassifikation (1)
- Klima (1)
- Knowledge Management (1)
- Knowledge Work (1)
- Knowledge-based Techniques (1)
- Knuth-Bendix completion (1)
- Knuth-Bendix completion algorithm (1)
- Knuth-Bendix-Vervollständigung (1)
- Kollisionserkennung (1)
- Kommunikationsprotokoll (1)
- Komplexitätsklasse NP (1)
- Komponente <Software> (1)
- Komponentenmodell (1)
- Kompression (1)
- Komprimierung (1)
- Korrelationsanalyse (1)
- Kraftregelung (1)
- Kraftrückkopplung (1)
- Kryptographie (1)
- Kryptologie (1)
- Kurve (1)
- Künstliche Intelligenz (1)
- LIDAR (1)
- LIR-Tree (1)
- Language Constructs (1)
- Large Data (1)
- Large High-Resolution Displays (1)
- Laser Wakefield Particle Accelerator (1)
- Laufkomprimierung (1)
- Learning Analytics (1)
- Leistungsmessung (1)
- Lernalgorithmen (1)
- Link Metric (1)
- Linked Data (1)
- Linking Data Analysis and Visualization (1)
- Linux (1)
- Location Awareness (1)
- Logical Time (1)
- Logische Programmierung (1)
- Lokalisierung (1)
- Lärmbelastung (1)
- Lärmimmission (1)
- MAC protocols (1)
- MDA <Vorgehensmodell> (1)
- MEGI (1)
- MHEG (1)
- MIDI <Musikelektronik> (1)
- MIP-Emissionsspektroskopie (1)
- MIP-Massenspektrometrie (1)
- MIR (1)
- MOO (1)
- MP3 (1)
- MVP-L (1)
- Magnetfeldbasierter Lokalisierung (1)
- Magnetfelder (1)
- Magnetometer (1)
- Manufacturing (1)
- Map Building (1)
- MapReduce (1)
- Maschinelle Übersetzung (1)
- Mastoid (1)
- Mastoidektomie (1)
- Matrix Completion (1)
- Maturity of Software Engineering (1)
- Maximum Intensity Projection (1)
- Medical Image Analysis (1)
- Mehragentensystem (1)
- Memory Architecture (1)
- Memory Consistency (1)
- Menschenmenge (1)
- Merkmalsraum (1)
- Mesh-Free (1)
- Metadaten (1)
- Meter (1)
- Methods (1)
- Microelectromechanical Systems (1)
- Mikrocontroller AVR (1)
- Mikroskopie (1)
- Minimal Cut Set Visualization (1)
- Minimal training (1)
- MoCAS/2 (1)
- Mobile Computing (1)
- Mobile system (1)
- Mobiler Roboter (1)
- Mode-Based Scheduling with Fast Mode-Signaling (1)
- Model Based User Interface Development (1)
- Modelling (1)
- Modularisierung (1)
- Modusbasierte Signalisierung (1)
- Molekulare Bioinformatik (1)
- Moment Invariants (1)
- Moment-Generating Functions (1)
- Mood-based Music Recommendations (1)
- Multi-Edge Graph (1)
- Multi-Field (1)
- Multi-Variate Data (1)
- Multidisciplinary Optimization (1)
- Multifield Data (1)
- Mund-Kiefer-Gesichts-Chirurgie (1)
- Music Information Retrieval (1)
- Musik / Artes liberales (1)
- NP-hard (1)
- Nachhaltigkeit (1)
- Natural Neighbor (1)
- Natural Neighbor Interpolation (1)
- Natürliche Nachbarn (1)
- Nearest-Neighbor Classification (1)
- Network Architecture (1)
- Netz-Architekturen (1)
- Netzwerkmanagement (1)
- Neural Architecture Search (1)
- Neural Networks (1)
- NoSQL (1)
- Node-Link Diagram (1)
- Numerische Strömungssimulation (1)
- Nvidia (1)
- OCL 2.0 (1)
- OCR (1)
- OWL (1)
- Object-OrientedCase Representation (1)
- Object-Relational DataBase Management Systems (ORDBMS) (1)
- Object-Relational Database Systems (1)
- Object-orientation (1)
- Objektorientierung (1)
- Off-road Robotics (1)
- Off-road Robotik (1)
- Ohrenchirurgie (1)
- Online chain partitioning (1)
- Ontolingua (1)
- Ontology (1)
- Open Estelle (1)
- Open-Source (1)
- Operationsroboter (1)
- Optimierender Compiler (1)
- Optimization (1)
- P2P (1)
- PABS-Methode (1)
- PATDEX 2 (1)
- PC-based robot control (1)
- PDF <Dateiformat> (1)
- PERA (1)
- PI-Regler (1)
- PLAN Abstraction (1)
- PVM (1)
- Parallel Virtual Machines (1)
- Paralleler Hybrid (1)
- Pareto Optimality (1)
- Parser (1)
- Partial Differential Equations (1)
- Partially ordered sets (1)
- Participatory Sensing (1)
- Peer-to-Peer-Netz (1)
- Performance (1)
- Personalisation (1)
- Pervasive health (1)
- Physical activity monitoring (1)
- Planar Pressure (1)
- Planning and Verification (1)
- Position Sensitive Device (1)
- Position- and Orientation Estimation (1)
- Privacy (1)
- Problem Solvers (1)
- Problemlösung (1)
- Process Data (1)
- Process Management (1)
- Process configuration (1)
- Process creation (1)
- Process support (1)
- Processor Architecture (1)
- Processor Architectures (1)
- Produktionsdesign (1)
- Profiles (1)
- Project Management (1)
- Protocol Composition (1)
- Prototyp (1)
- Prototype (1)
- Prozessmodellen (1)
- Prozessvisualisierung (1)
- Quality (1)
- Quality Improvement Paradigm (QIP) (1)
- Qualität (1)
- Quantitative Bildanalyse (1)
- Quartz (1)
- Quicksort (1)
- RDF (1)
- RGB (1)
- RLE (1)
- RNA interaction (1)
- RSA (1)
- Radar (1)
- Random testing (1)
- Rate Gyro (1)
- Raumordnung (1)
- Ray tracing (1)
- ReasoningSystems (1)
- Rechnernetze (1)
- Rechtecksgitter (1)
- Rectilinear Grid (1)
- Redundanzvermeidung (1)
- Regler (1)
- Reglerentwurf (1)
- Reinforcement Learning (1)
- Rekonstruktion (1)
- Relationales Datenbanksystem (1)
- Repositories (1)
- Repository <Informatik> (1)
- Representation (1)
- Requirements engineering (1)
- Reservierungsprotokoll (1)
- Resource Description Framework (1)
- Reuse (1)
- Risikobewertung (1)
- Risk Assessment (1)
- Robot Calibration (1)
- Roboterarm (1)
- Rogue AP (1)
- Räumliche Statistik (1)
- SAHARA (1)
- SAX2 (1)
- SCAD (1)
- SCM (1)
- SDL extensions (1)
- SDL-2000 (1)
- SDL-oriented Object Modeling Technique (1)
- SDL-pattern a (1)
- SFB-EB (1)
- SIMD architectures (1)
- SOMT (1)
- SPARQL (1)
- SPARQL query learning (1)
- SQL (1)
- Safety Analysis (1)
- Scalar (1)
- Scheduling (1)
- Schema <Informatik> (1)
- Schematisation (1)
- Schematisierung (1)
- Schädelbasis (1)
- Schädelchirurgie (1)
- Scientific Community Analysis (1)
- Scientific Computing (1)
- Self-Referencing (1)
- Self-organization (1)
- Self-supervised Learning (1)
- Semantic Desktop (1)
- Semantic Index (1)
- Semantic Wikis (1)
- Semantics of Programming Languages (1)
- Semantische Modellierung (1)
- Semantisches Datenmodell (1)
- Sensing (1)
- Sensor (1)
- Sensor Fusion (1)
- Sensors (1)
- Service Access Points (1)
- Shared Resource Modeling (1)
- Sicherheit (1)
- Sicherheitsanalyse (1)
- Similarity Assessment (1)
- Similarity Join (1)
- Similarity Joins (1)
- Simulationen (1)
- Simultaneous quantifier elimination (1)
- Skalar (1)
- Skalierbarkeit (1)
- Smalltalk (1)
- Smart City (1)
- Smart Device (1)
- Smart Mobile Device (1)
- Smart Textile (1)
- SmartFactory (1)
- Smartphone (1)
- Smartwatch (1)
- Socio-Semantic Web (1)
- Software Comprehension (1)
- Software Configuration Management (1)
- Software Dependencies (1)
- Software Evolution (1)
- Software Maintenance (1)
- Software Measurement (1)
- Software Process Support (1)
- Software Testing (1)
- Software Visualization (1)
- Software development (1)
- Software development environment (1)
- Software transactional memory (1)
- Software-Architektur (1)
- Software-Entwicklung (1)
- Softwaremetrie (1)
- Softwarespezifikation (1)
- Softwarewartung (1)
- Spezifikation (1)
- Sprachdefinition (1)
- Sprache (1)
- Sprachen (1)
- Statistical Independence (1)
- Stereovision (1)
- Stimmungsbasierte Musikempfehlungen (1)
- Stochastic Dependence (1)
- Stochastic Network Calculus (1)
- Stokes Equations (1)
- Streaming (1)
- Structural Adaptation (1)
- Structure (1)
- Structuring Approach (1)
- Strukturiertes Gitter (1)
- Strömung (1)
- Suchve (1)
- Support-Vektor-Maschine (1)
- Surface Reconstruction (1)
- Symbolic Methods (1)
- Synchronous Control Asynchronous Dataflow (1)
- System Abstractions (1)
- SystemC (1)
- Systemdesign (1)
- TCP/IP (1)
- Tactics (1)
- Task-based (1)
- Tcl (1)
- Technology combination (1)
- Technology decision (1)
- Technology selection (1)
- Temporal Decoupling (1)
- Temporal Logic (1)
- Temporal data processing (1)
- Tensor (1)
- Tensorfeld (1)
- Termination (1)
- Tesselation (1)
- Tetraeder (1)
- Tetraedergi (1)
- Tetrahedral Grid (1)
- Tetrahedral Mesh (1)
- Textual CBR (1)
- Themenbasierte Empfehlungen von Ressourcen (1)
- Time Series (1)
- Time series classification (1)
- Time-motion-Ultraschallkardiographie (1)
- Time-slotted (1)
- Tonsignal (1)
- Topic-based Resource Recommendations (1)
- Topologie (1)
- Topology Preserving Networks (1)
- Topology visualization (1)
- Transaction Level Modeling (TLM) (1)
- Transport Protocol (1)
- Traversability Analysis (1)
- UML Activity (1)
- UML Profile (1)
- Ubiquitous system (1)
- Ultraschallkardiographie (1)
- Umweltinformatik (1)
- Uncertainty Estimation (1)
- Universal Control Device (1)
- Unobtrusive instrumentations (1)
- Unorganized Data (1)
- Unstrukturiertes Gitter (1)
- Unterricht (1)
- Urban sprawl (1)
- UrbanSim (1)
- Usage Control (1)
- User Model (1)
- VIACOBI (1)
- VIROP (1)
- Validierung (1)
- Vector (1)
- Vector Field (1)
- Vector Time (1)
- Vektor (1)
- Vektorfelder (1)
- Verifikation (1)
- Versionierungssysteme (1)
- Verteiltes System (1)
- Verzerrungstensor (1)
- Virtual Corporation (1)
- Virtual Prototyping (1)
- Virtual Software Projects (1)
- Vision (1)
- Visual Queries (1)
- Visualization Theory (1)
- Volume rendering (1)
- Volumen-Rendering (1)
- Voronoi diagram (1)
- WETICE 98 (1)
- Weak Memory Model (1)
- Wearable Computing (1)
- Wi-Fi (1)
- Wide Area Multimedia Group Interaction (1)
- Wide-column stores (1)
- Wireless Networks (1)
- Wissenschaftliches Rechnen (1)
- Wissenserwerb (1)
- Workflow Replication (1)
- Workflowmanagement (1)
- Workstation-Cluster (1)
- World-Wide Web (1)
- XDBMS (1)
- XEC (1)
- XML query estimation (1)
- XML summary (1)
- Yaroslavskiy-Bentley-Bloch Quicksort (1)
- Zeitplanung (1)
- Zielverfolgung (1)
- Zugriffstruktur (1)
- Zugriffsystem (1)
- abstract description (1)
- access rights (1)
- acoustic modeling (1)
- active damping (1)
- adaption (1)
- affective user interface (1)
- affine arithmetic (1)
- aliasing (1)
- analogical reasoning (1)
- architecture (1)
- artificial neural network (1)
- aspect-oriented programming (1)
- assembly sequence design (1)
- associations (1)
- authentication (1)
- autoencoder (1)
- automated code generation (1)
- automated computer learning (1)
- automated proof planner (1)
- automated synchronization (1)
- autonomes Lernen (1)
- autonomous learning (1)
- autonomous systems (1)
- average-case analysis (1)
- bedingte Aktionen (1)
- behaviour-based system (1)
- behaviourbased (1)
- bi-directional search (1)
- bidirectional search (1)
- binary countdown protocol (1)
- biological motivated (1)
- biologisch motiviert (1)
- black bursts (1)
- body-IMU calibration (1)
- branch-and-bound (1)
- buildroot (1)
- business process modelling (1)
- business process reengineering (1)
- byte code (1)
- case-based planner (1)
- classification (1)
- client/server-architecture (1)
- climate (1)
- clustering (1)
- co-learning (1)
- collaborative information visualization (1)
- collaborative mobile sensing (1)
- collective intelligence (1)
- collision detection (1)
- combinatorial algorithms (1)
- communication architectures (1)
- communication protocols (1)
- communication subsystem (1)
- compilation (1)
- compiler (1)
- completeness (1)
- complex System Development (1)
- component-based (1)
- comprehensive reuse (1)
- computational biology (1)
- computational fluid dynamics (1)
- concept representation (1)
- conceptual design (1)
- conceptual representation (1)
- concurrent (1)
- concurrent software (1)
- constraint exploration (1)
- constraint satisfaction problem (CSP) (1)
- constraint-based reasoning (1)
- content-and-structure summary (1)
- continuous master theorem (1)
- continuous media (1)
- cooperative problem solving (1)
- crowd condition estimation (1)
- crowd density estimation (1)
- crowd scanning (1)
- crowd sensing (1)
- crowdsourcing (1)
- curves and surfaces (1)
- customization of communication protocols (1)
- data race (1)
- data sets (1)
- data-flow (1)
- dataset (1)
- decidability (1)
- decision support (1)
- deformable object (1)
- dependability (1)
- dependable systems (1)
- dependency management (1)
- description of reactive systems (1)
- design (1)
- design processes (1)
- deterministic arbitration (1)
- diagnostic problems (1)
- directed graphs (1)
- discosat (1)
- display algorithms (1)
- distributed and parallel processing (1)
- distributed c (1)
- distributed compliant control (1)
- distributed control system (1)
- distributed deduction (1)
- distributed document management (1)
- distributed enterprise (1)
- distributed groupware environment (1)
- distributed multi-platform software development (1)
- distributed multi-platform software development projects (1)
- distributed processing (1)
- distributed real-time systems (1)
- distributed software configuration management (1)
- distributed softwaredevelopment tools (1)
- distributed tasks (1)
- distributedknowledge-based systems (1)
- domain-specific language (1)
- domains (1)
- efficiency (1)
- embedded (1)
- embedding (1)
- emotion visualization (1)
- encapsulation (1)
- end-to-end learning (1)
- energy consumption (1)
- ensemble (1)
- environmental noise (1)
- epidemic algorithms (1)
- epidemische Algorithmen (1)
- ernergy effcient motion (1)
- evolutionary algorithm (1)
- experience management (1)
- experimental software engineering (1)
- explainability (1)
- fallbasiertes planen (1)
- firewall (1)
- flexible workflows (1)
- flexible-link (1)
- flexible-link robot (1)
- flow visualization (1)
- force control (1)
- force following (1)
- formal description techniques (1)
- formal reasoning (1)
- foundational translation validation (1)
- frame buffer operations (1)
- framework (1)
- gaussian filter (1)
- generative Programmierung (1)
- generative programming (1)
- generic design of a customized communication subsystem (1)
- geographic information systems (1)
- geology (1)
- goal oriented completion (1)
- graph drawing algorithm (1)
- graph embedding (1)
- graph layout (1)
- graphic processors (1)
- guarded actions (1)
- hand pose, hand shape, depth image, convolutional neural networks (1)
- heterogeneous large-scale distributed DBMS (1)
- high-level caching of potentially shared networked documents (1)
- higher order logic (1)
- higher order tableau (1)
- higher-order calculi (1)
- higher-order tableaux calculus (1)
- higher-order theorem prover (1)
- historical documents (1)
- hub location (1)
- human body motion tracking (1)
- humanoid robot (1)
- humanoide Roboter (1)
- hybrid control (1)
- hybrid knowledge representation (1)
- hypercubes (1)
- hypergraph (1)
- iB2C (1)
- idle times (1)
- image analysis (1)
- implementation (1)
- industrial supervision (1)
- inertial sensors (1)
- information systems (1)
- information systems engineering (1)
- intelligent agents (1)
- intentional programming (1)
- interference (1)
- internet event synchronizer (1)
- interpolation (1)
- interpreter (1)
- interval arithmetic (1)
- invariant (1)
- inverses Pendel (1)
- isochronous streams (1)
- knowledge processing (1)
- knowledge space (1)
- knowledge-based planning (1)
- konzeptuelle Modelierung (1)
- kraftbasiertes Führen (1)
- language definition (1)
- language modeling (1)
- language profiles (1)
- learning (1)
- learning algorithms (1)
- linked abstraction workflows (1)
- linked data (1)
- load sharing (1)
- local communication (1)
- long short-term memory (1)
- long tail (1)
- machine-checkable proof (1)
- magnetic field based localization (1)
- magnetometer calibration (1)
- manipulation (1)
- mastoid (1)
- mastoidectomy (1)
- mathematical concept (1)
- matrix visualization (1)
- measurement (1)
- message-passing (1)
- metadata (1)
- microelectronics ontology (1)
- middleware (1)
- migration (1)
- mixed-signal (1)
- mobile agents (1)
- mobile agents approach (1)
- modelling time (1)
- modularisation (1)
- moment (1)
- monitoring and managing distributed development processes (1)
- morphism (1)
- multi-agent architecture (1)
- multi-language (1)
- multicore (1)
- multidimensional datasets (1)
- multimedia (1)
- multinomial regression (1)
- multiple context free grammar (1)
- multiple-view product modeling (1)
- multithreading (1)
- multitype code coupling (1)
- multiway partitioning (1)
- narrowing (1)
- natural language semantics (1)
- navigation (1)
- negotiation (1)
- nestable tangibles (1)
- object frameworks (1)
- object-orientation (1)
- object-oriented software modeling (1)
- off-line programming (1)
- operations research (1)
- optimization (1)
- optimization correctness (1)
- order-sorted logic (1)
- oscillating magnetic fields (1)
- oscillation (1)
- otorhinolaryngological surgery (1)
- out-of-order (1)
- ownership (1)
- parallel (1)
- parallelism and concurrency (1)
- paramodulation (1)
- participatory sensing (1)
- path cost models (1)
- pattern recognition (1)
- peer-to-peer (1)
- persistence (1)
- pivot sampling (1)
- plan enactment (1)
- planning (1)
- point cloud (1)
- point-to-point (1)
- problem solvers (1)
- process model (1)
- process modelling (1)
- process support system (PROSYT) (1)
- process-centred environments (1)
- profiles (1)
- programmable client-server systems (1)
- programming by demonstration (1)
- project coordination (1)
- proof generating optimizer (1)
- proof plans (1)
- proof presentation (1)
- protocol (1)
- quality assurance (1)
- rapid authoring (1)
- raster graphics (1)
- rate control (1)
- ray casting (1)
- ray tracing (1)
- reactive systems (1)
- real time (1)
- real-time (1)
- real-time tasks (1)
- real-time temporal logic (1)
- reasoning (1)
- receptive safety properties (1)
- redundancy (1)
- redundant robots (1)
- rekursiv aufzählbare Sprachfamilien (1)
- relaxed memory models (1)
- reliability (1)
- requirements (1)
- reuse repositories (1)
- review (1)
- rings (1)
- risk management (1)
- robot (1)
- robot calibration (1)
- robot control (1)
- robot control architectures (1)
- robot kinematics (1)
- robot motion planning (1)
- robotergestützt (1)
- robotics (1)
- robustness (1)
- roles (1)
- rule-based reasoning (1)
- safe human robot coexistence (1)
- scalar field (1)
- scene flow (1)
- sdr (1)
- search algorithm (1)
- search alogorithms (1)
- search-space-problem (1)
- second order logic (1)
- secondary structure prediction (1)
- security domain (1)
- seed filling (1)
- self-localization (1)
- semiring parsing (1)
- sequent calculus (1)
- shortest sequence (1)
- similarity measure (1)
- skolemization (1)
- small-multiples node-link visualization (1)
- sniffing (1)
- social media (1)
- software agents (1)
- software comprehension (1)
- software engineering (1)
- software engineering task (1)
- software project (1)
- software project management (1)
- software reuse (1)
- sorted logic (1)
- soundness (1)
- sparse-to-dense (1)
- spatial statistics (1)
- state-based formalism (1)
- static load balancing (1)
- static software structure (1)
- stationary sensing (1)
- statistics (1)
- stochastic context free grammar (1)
- structural summary (1)
- subjective evaluation (1)
- symbolic simulation (1)
- synchrone Sprachen (1)
- synchronous (1)
- synchronous languages (1)
- system architecture (1)
- system behaviour (1)
- tableau (1)
- tabletop (1)
- task sequence (1)
- tensor (1)
- tensorfield (1)
- terrain rendering (1)
- time-dependent (1)
- time-varying flow fields (1)
- topology preserving maps (1)
- tori (1)
- touch surfaces (1)
- traceability (1)
- trajectory optimization (1)
- translation (1)
- translation contract (1)
- translation validation (1)
- traveling salesman problem (1)
- types (1)
- typical examples (1)
- typical instance (1)
- ultrasound signals (1)
- urban planning (1)
- user-centered design (1)
- vector field visualization (1)
- vectorfield (1)
- verhaltensbasiert (1)
- verlustfrei (1)
- verlässlichkeit (1)
- verteilte Berechnung (1)
- vetreilte nachgiebige Regelung (1)
- virtual market place (1)
- virtual programming (1)
- virtual reality (1)
- visual analytics (1)
- visual analytics, machine learning, interaction, user experience, assistive technologies (1)
- visual process modelling environment (1)
- visual structure (1)
- wearable systems (1)
- weighted finite-state transducers (1)
- wireless signal (1)
- wissensbasierte Systeme (1)
- wissensbasierter Systeme der Arbeitsplanerstellung (1)
- work coordination (1)
- world modelling (1)
- world-modelling (1)
- xai (1)
- zeitabhängige Strömungen (1)
- Ähnlichkeit (1)
- Übersetzung (1)
Faculty / Organisational entity
Distributed message-passing systems have become ubiquitous and essential for our daily lives. Hence, designing and implementing them correctly is of utmost importance. This is, however, very challenging at the same time. In fact, it is well-known that verifying such systems is algorithmically undecidable in general due to the interplay of asynchronous communication (messages are buffered) and concurrency. When designing communication in a system, it is natural to start with a global protocol specification of the desired communication behaviour. In such a top-down approach, the implementability problem asks, given such a global protocol, if the specified behaviour can be implemented in a distributed setting without additional synchronisation. This problem has been studied from two perspectives in the literature. On the one hand, there are Multiparty Session Types (MSTs) from process algebra, with global types to specify protocols. Key to the MST approach is a so-called projection operator, which takes a global type and tries to project it onto every participant: if successful, the local specifications are safe to use. This approach is efficient but brittle. On the other hand, High-level Message Sequence Charts (HMSCs) study the implementability problem from an automata-theoretic perspective. They employ very few restrictions on protocol specifications, making the implementability problem for HMSCs undecidable in general. The work in this thesis is the first to formally build a bridge between the world of MSTs and HMSCs. To start, we present a generalised projection operator for sender-driven choice. This allows a sender to send to different receivers when branching, which is crucial to handle common communication patterns from distributed computing. Despite this first step, we also show that the classical MST projection approach is inherently incomplete. We present the first formal encoding from global types to HMSCs. With this, we prove decidability of the implementability problem for global types with sender-driven choice. Furthermore, we develop the first direct and complete projection operator for global types with sender-driven choice, using automata-theoretic techniques, and show its effectiveness with a prototype implementation. We are the first to provide an upper bound for the implementability problem for global types with sender-driven (or directed) choice and show it to be in PSPACE. We also provide a session type system that uses the results from our projection operator. Last, we introduce protocol state machines (PSMs) – an automata-based protocol specification formalism – that subsume both global types from MSTs and HMSCs with regard to expressivity. We use transformations on PSMs to show that many of the syntactic restrictions of global types are not restrictive in terms of protocol expressivity. We prove that the implementability problem for PSMs with mixed choice, which requires no dedicated sender for a branch but solely all labels to be distinct, is undecidable in general. With our results on expressivity, this answers an open question: the implementability problem for mixed-choice global types is undecidable in general.
Machine learning algorithms are widely applied to create powerful prediction models. With increasingly complex models, humans' ability to understand the decision function (that maps from a high-dimensional input space) is quickly exceeded. To explain a model's decisions, black-box methods have been proposed that provide either non-linear maps of the global topology of the decision boundary, or samples that allow approximating it locally. The former loses information about distances in input space, while the latter only provides statements about given samples, but lacks a focus on the underlying model for precise ‘What-If'-reasoning. In this paper, we integrate both approaches and propose an interactive exploration method using local linear maps of the decision space. We create the maps on high-dimensional hyperplanes—2D-slices of the high-dimensional parameter space—based on statistical and personal feature mutability and guided by feature importance. We complement the proposed workflow with established model inspection techniques to provide orientation and guidance. We demonstrate our approach on real-world datasets and illustrate that it allows identification of instance-based decision boundary structures and can answer multi-dimensional ‘What-If'-questions, thereby identifying counterfactual scenarios visually.
Edit distances between merge trees of scalar fields have many applications in scientific visualization, such as ensemble analysis, feature tracking or symmetry detection. In this paper, we propose branch mappings, a novel approach to the construction of edit mappings for merge trees. Classic edit mappings match nodes or edges of two trees onto each other, and therefore have to either rely on branch decompositions of both trees or have to use auxiliary node properties to determine a matching. In contrast, branch mappings employ branch properties instead of node similarity information, and are independent of predetermined branch decompositions. Especially for topological features, which are typically based on branch properties, this allows a more intuitive distance measure which is also less susceptible to instabilities from small-scale perturbations. For trees with 𝒪(n) nodes, we describe an 𝒪(n4) algorithm for computing optimal branch mappings, which is faster than the only other branch decomposition-independent method in the literature by more than a linear factor. Furthermore, we compare the results of our method on synthetic and real-world examples to demonstrate its practicality and utility.
The development of algorithmic differentiation (AD) tools focuses mostly on handling floating point types in the target language. Taping optimizations in these tools mostly focus on specific operations like matrix vector products. Aggregated types like std::complex are usually handled by specifying the AD type as a template argument. This approach provides exact results, but prevents the use of expression templates. If AD tools are extended and specialized such that aggregated types can be added to the expression framework, then this will result in reduced memory utilization and improve the timing for applications where aggregated types such as complex number or matrix vector operations are used. Such an integration requires a reformulation of the stored data per expression and a rework of the tape evaluation process. We will demonstrate the overheads on a synthetic benchmark and show the improvement when aggregated types are handled properly by the expression framework of the AD tool.
To increase situational awareness of the crane operator, the aim of this thesis is to develop a vision-based deep learning object detection from crane load-view using an adaptive perception in the construction area. Conventional worker detection methods are based on simple shape or color features from the worker's appearances. Nonetheless, these methods can fail to recognize the workers who do not wear the protective gears. To find out an image representation of the object from the top view manually or handcrafted feature is crucial. We, therefore, employed deep learning methods to automatically learn those features.
To yield optimal results, deep learning methods require mass amount of data.
Due to the data deficit especially in the construction domain, we developed the photorealistic world to create the data in addition to our samples collected from the real construction area. The simulated platform does not benefit only from diverse data types, but also concurrent research development which speeds up the pipeline at a low cost.
Our research findings indicate that the combination of synthetic and real training samples improved the state-of-the-art detector. In line with previous studies to bridge the gap between synthetic and real data, the results of preprocessed synthetic images are substantially better than using the raw data by approximately 10%.
Finding the right deep learning model for load-view detection is challenging.
By investigating our training data, it becomes evident that the majority of bounding box sizes are very small with a complex background.
In addition, we gave the priority to speed over accuracy based on the construction safety criteria. Finally, RetinaNet is chosen out of the three primary object detection models.
Nevertheless, the data-driven detection algorithm can fail to handle scale invariance, especially for detectors whose input size is changed in an extremely wide range.
The adaptive zoom feature can enhance the quality of the worker detection.
To avoid further data gathering and extensive retraining, the proposed automatic zoom method of the load-view crane camera supports the deep learning algorithm, specifically in the high scale variant problem. The finite state machine is employed for control strategies to adapt the zoom level to cope not only with inconsistent detection but also abrupt camera movement during lifting operation. Consequently, the detector is able to detect a small size object by smooth continuous zoom control without additional training.
The adaptive zoom control not only enhances the performance of the top-view object detection but also reduces the interaction of the crane operator with camera system, reducing the risk of fatality during load lifting operation.
Turbulence models, which are a means to fix the closure problem arising from Reynolds averaging of Navier-Stokes equations, are economical stop-gaps but suffer from accuracy issues. Modifying turbulence models by incorporating corrections in their functional form is one approach to improve their accuracy. We estimate correction functionals for the Spalart - Allmaras turbulence model, based on an inverse problem with PDE constraints emphasizing the issue of regularization.
Algorithmic decision-making (ADM) systems have come to support, pre-empt or substitute for human decisions in manifold areas, with potentially significant impacts on individuals' lives. Achieving transparency and accountability has been formulated as a general goal regarding the use of these systems. However, concrete applications differ widely in the degree of risk and the accountability problems they entail for data subjects. The present paper addresses this variation and presents a framework that differentiates regulatory requirements for a range of ADM system uses. It draws on agency theory to conceptualize accountability challenges from the point of view of data subjects with the purpose to systematize instruments for safeguarding algorithmic accountability. The paper furthermore shows how such instruments can be matched to applications of ADM based on a risk matrix. The resulting comprehensive framework can guide the evaluation of ADM systems and the choice of suitable regulatory provisions.
We describe a novel technique for the simultaneous visualization of multiple scalar fields, e.g. representing the members of an ensemble, based on their contour trees. Using tree alignments, a graph-theoretic concept similar to edit distance mappings, we identify commonalities across multiple contour trees and leverage these to obtain a layout that can represent all trees simultaneously in an easy-to-interpret, minimally-cluttered manner. We describe a heuristic algorithm to compute tree alignments for a given similarity metric, and give an algorithm to compute a joint layout of the resulting aligned contour trees. We apply our approach to the visualization of scalar field ensembles, discuss basic visualization and interaction possibilities, and demonstrate results on several analytic and real-world examples.
Editorial
(2021)
In recent years, there has been a growing need for accurate 3D scene reconstruction. Recent developments in the automotive industry have led to the increased use of ADAS where 3D reconstruction techniques are used, for example, as part of a collision detection system. For such applications, scene geometry reconstruction is usually performed in the form of depth estimation, where distances to scene objects are obtained.
In general, depth estimation systems can be divided into active and passive. Both systems have their advantages and disadvantages, but passive systems are usually cheaper to produce and easier to assemble and integrate than active systems. Passive systems can be stereo- or multiple-view based. Up to a certain limit, increasing the number of views in multi-view systems usually results in improved depth estimation accuracy.
One potential problem for ensuring the reliability of multi-view systems is the need to accurately estimate the orientation of their optical sensors. One way to ensure sensor placement for multi-view systems is to rigidly fix the sensors at the manufacturing stage. Unlike arbitrary sensor placement, using of a simplified and known sensor placement geometry further simplifies the depth estimation.
We meet with the concept of light field, which parameterizes all visible light passing through all viewpoints by their intersection with angular and spatial planes. When applied to computer vision, this gives us a 2D set of 2D images, where the physical distances between each image are fixed and proportional to each other.
Existing light field depth estimation methods provide good accuracy, which is suitable for industrial applications. However, the main problems of these methods are related to their running time and resource requirements. Most of the algorithms presented in the literature are typically sharpened for accuracy, can only be run on high-performance machines and often require a significant amount of time to process and obtain results.
Real-world applications often have running time requirements. Also, often there is a power-consumption limitation. In this dissertation, we investigate the problem of building a depth estimation system with an light field camera that satisfies the operating time and power consumption constraints without significant loss of estimation accuracy.
First, an algorithm for calibrating light field cameras is proposed, together with an algorithm for automatic calibration refinement, that works on arbitrary captured scenes. An algorithm for classical geometric depth estimation using light field cameras is proposed. Ways to optimize the algorithm for real-time use without significant loss of accuracy are presented. Finally, the ways how the presented depth estimation methods can be extended using modern deep learning paradigms under the two previously mentioned constraints are shown.
In order to discuss the kinds of reasoning a visualization supports and the conclusions that can be drawn within the analysiscontext, a theoretical framework is needed that enables a formal treatment of the reasoning process. Such a model needs toencompass three stages of the visualization pipeline: encoding, decoding and interpretation. The encoding details how dataare transformed into a visualization and what can be seen in the visualization. The decoding explains how humans constructgraphical contexts inside the depicted visualization and how they interpret them assigning meaning to displayed structuresaccording to a formal reasoning strategy. In the presented model, we adapt and combine theories for the different steps intoa unified formal framework such that the analysis process is modelled as an assignment of meaning to displayed structuresaccording to a formal reasoning strategy. Additionally, we propose the ConceptGraph, a combined graph-based representationof the finite-state transducers resulting from the three stages, that can be used to formalize and understand the reasoning process.We apply the new model to several visualization types and investigate reasoning strategies for various tasks.
Knowledge workers face an ever increasing flood of information in their daily work. They live in a “multi-tasking craziness”, involving activities like creating, finding, processing, assessing or organizing information while constantly switching from one context to another, each being associated with different tasks, documents, mails, etc. Hence, their personal information sphere consisting of file, mail and bookmark folders as well as their content, calendar entries, etc. is cluttered with information that has become irrelevant. Finding important information thus gets harder and much of previously gained knowledge is practically lost.
This thesis explores new ways of solving this problem by investigating the potential of self-(re)organizing and especially forgetting-enabled personal knowledge assistants in the given scenario. It utilizes so-called Managed Forgetting, which is an escalating set of measures to overcome the binary keep-or-delete paradigm, ranging from temporal hiding, to condensation, to adaptive reorganization, synchronization, archiving and deletion. Managed Forgetting is combined with two other major ideas: First, it uses the Semantic Desktop as an ecosystem, which brings Semantic Web and thus knowledge graph technologies to a user’s desktop, making it possible to capture and represent major parts of a user’s personal mental model in a machine-understandable way and exploit it in many different applications. Second, the system uses explicated context information – so-called Context Spaces: context is seen as an explicit interaction element users can work with (i.e. a “tangible” object similar to a folder) and in (immersion). The thesis is structured according to the basic interaction cycle with such a system, ranging from evidence collection to information extraction and context elicitation, followed by information value assessment and the actual support measures consisting of self-(re)organization decisions (back-end) and user interface updates (front-end). The system’s data foundation are personal or group knowledge graphs as well as native data. This work makes contributions to all of these aspects, whereas several of them have been investigated and developed in interdisciplinary research with cognitive scientists. On a more general level, searching and trust in such highly autonomous assistants have also been investigated.
In summary, a self-(re)organizing and especially forgetting-enabled support system for information management and knowledge work has been realized. Its different features vary in maturity: the most mature ones are already in practical use (also in industry), while the latest are just well elaborated (position papers) or rough ideas. Different evaluation strategies have been applied ranging from mere data-driven experiments to various user studies. Some of them were rather short-term with controlled laboratory conditions, others less controlled but spanning several months. Different benefits of working with such a system could be quantified, e.g. cognitive offloading effects and reduced task switching/resumption time. Other benefits were gathered qualitatively, e.g. tidiness of the information sphere and its better alignment with the user’s mental model. The presented approach has been shown to hold a lot of potential. In some aspects, however, only first steps have been taken towards tapping it, e.g. several support measures can be further refined and automation further increased.
Editorial
(2020)
Editorial
(2020)
Several governmental organizations all over the world aim for algorithmic accountability of artificial intelligence systems. However, there are few specific proposals on how exactly to achieve it. This article provides an extensive overview of possible transparency and inspectability mechanisms that contribute to accountability for the technical components of an algorithmic decision-making system. Following the different phases of a generic software development process, we identify and discuss several such mechanisms. For each of them, we give an estimate of the cost with respect to time and money that might be associated with that measure.
Editorial
(2020)
This thesis focuses on the operation of reliability-constrained routes in wireless ad-hoc networks. A complete communication protocol that is capable of guaranteeing a statistical minimum reliability level would have to support several functionalities: first, routes that are capable of supporting the specified Quality of Service requirement have to be discovered. During operation of discovered routes, the current Quality of Service level has to be monitored continuously. Whenever significant deviations are detected and the required level of Quality of Service is endangered, route maintenance has to ensure continuous operation. All four functionalities, route discovery, route operation, route maintenance and collection and distribution of network status information, will be addressed in this thesis.
In the first part of the thesis, we propose a new approach for Quality-of- Service routing in wireless ad-hoc networks called rmin-routing, with the provision of statistical minimum route reliability as main route selection criterion. To achieve specified minimum route reliabilities, we improve the reliability of individual links by well-directed retransmissions, to be applied during the operation of routes. To select among a set of candidate routes, we define and apply route quality criteria concerning network load.
High-quality information about the network status is essential for the discovery and operation of routes and clusters in wireless ad-hoc networks. This requires permanent observation and assessment of nodes, links, and link metrics, and the exchange of gathered status data. In the second part of the thesis, we present cTEx, a configurable topology explorer for wireless ad-hoc networks that efficiently detects and exchanges high-quality network status information during operation.
In the third part, we propose a decentralized algorithm for the discovery and operation of reliability-constrained routes in wireless ad-hoc networks called dRmin-routing. The algorithm uses locally available network status information about network topology and link properties that is collected proactively in order to discover a preliminary route candidate. This is followed by a distributed, reactive search along this preselected route to remove imprecisions of the locally recorded network status before making a final route selection. During route operation, dRmin-routing monitors routes and performs different kinds of route repair actions to maintain route reliability in order to overcome varying link reliabilities.
Modeling and Simulation of Internet of Things Infrastructures for Cyber-Physical Energy Systems
(2024)
This dissertation presents a novel approach to the model-based development and simulation-based validation of Internet of Things (IoT) infrastructures within the context of Cyber-Physical Energy Systems (CPES). CPES represents an evolution in energy management, seamlessly blending physical and cyber components for efficient, secure, and dependable energy distribution. However, the intricate interplay of these components demands innovative modeling and simulation strategies.
The work begins by establishing a robust foundation, exploring essential background elements such as requirements engineering, model-based systems engineering, digitalization approaches, and the intricacies of IoT platforms. It introduces the novel concept of homomorphic encryption, a critical enabler for securing IoT data within CPES.
In the exploration of the state of the art, the dissertation delves into the multifaceted landscape of IoT simulation, emphasizing the significance of versatility, community support, scalability, and synchronization.
The core contribution emerges in the chapter on simulating IoT networks. It introduces a sophisticated framework that encompasses hardware-in-the-loop, software-in-the-loop, and human-in-the-loop simulation. This innovative framework extends the boundaries of conventional simulation, enabling holistic evaluations of IoT systems.
A practical case study on smart energy usage showcases the application of the framework. Detailed SysML models, including requirements, package diagrams, block definition diagrams, internal block diagrams, state machine diagrams, and activity diagrams, are meticulously examined. The performance evaluation encompasses diverse aspects, from hardware and software validation to human interaction.
In conclusion, this dissertation represents a significant leap forward in the integration of IoT infrastructures within CPES. Its contributions extend from a comprehensive understanding of foundational elements to the practical implementation of a holistic simulation framework. This work not only addresses the current challenges but also outlines a path for future research, shaping the landscape of IoT integration within the dynamic realm of CPES. It offers invaluable insights for researchers, engineers, and stakeholders working towards resilient, secure, and energy-efficient infrastructures.
In many applications, visual analytics (VA) has developed into a standard tool to ease data access and knowledge generation. VA describes a holistic cycle transforming data into hypothesis and visualization to generate insights that enhance the data. Unfortunately, many data sources used in the VA process are affected by uncertainty. In addition, the VA cycle itself can introduce uncertainty to the knowledge generation process but does not provide a mechanism to handle these sources of uncertainty. In this manuscript, we aim to provide an extended VA cycle that is capable of handling uncertainty by quantification, propagation, and visualization, defined as uncertainty-aware visual analytics (UAVA). Here, a recap of uncertainty definition and description is used as a starting point to insert novel components in the visual analytics cycle. These components assist in capturing uncertainty throughout the VA cycle. Further, different data types, hypothesis generation approaches, and uncertainty-aware visualization approaches are discussed that fit in the defined UAVA cycle. In addition, application scenarios that can be handled by such a cycle, examples, and a list of open challenges in the area of UAVA are provided.
Dataflow process networks (DPNs) are intrinsically data-driven, i.e., node actions are not synchronized among each other and may fire whenever sufficient input operands arrived at a node. While the general model of computation (MoC) of DPNs does not impose further restrictions, many different subclasses of DPNs representing different dataflow MoCs have been considered over time. These classes mainly differ in the kinds of behaviors of the processes. A DPN may be heterogeneous in that different processes in the network belong to different classes of DPNs. A heterogeneous DPN can therefore be effectively used to model and to implement different components of a system with different kinds of processes and, therefore, different dataflow MoCs. This paper presents a model-based design based on different dataflow MoCs including their heterogeneous combinations. In particular, it covers the automatic software synthesis of systems from DPN models. The main objective is to validate, evaluate and compare the artifacts exhibited by different dataflow MoCs at the implementation level of systems under the supervision of a common design tool. Moreover, this work also offers an efficient synthesis method that targets and exploits heterogeneity in DPNs by generating implementations based on the kinds of behaviors of the processes. The proposed synthesis method provides a tool chain including different specialized code generators for specific dataflow MoCs, and a runtime system that finally maps models using a combination of different dataflow MoCs on cross-vendor target hardware.
This paper presents an iterative finite element (FE)–based method to calculate the gravity-free shape of nonrigid parts from
an optical measurement performed on a non-over-constrained fixture. Measuring these kinds of parts in a stress-free state
is almost impossible because deflections caused by their weight occur. To solve this problem, a simulation model of the
measurement is created using available methods of reverse engineering. Then, an iterative algorithm calculates the gravityfree
shape. The approach does not require a CAD model of the measured part, implying the whole part can be fully scanned.
The application of this method mainly addresses thin, unstable sheet metal parts, like those commonly used in the automotive
or aerospace industry. To show the performance of the proposed method, validations with simulation and experimental
data are presented. The shown results meet the predefined quality goal to predict shapes within a tolerance of ±0.05 mm
measured in surface normal direction.
We propose a universal method for the evaluation of generalized standard materials that greatly simplifies the material law implementation process. By means of automatic differentiation and a numerical integration scheme, AutoMat reduces the implementation effort to two potential functions. By moving AutoMat to the GPU, we close the performance gap to conventional evaluation routines and demonstrate in detail that the expression level reverse mode of automatic differentiation as well as its extension to second order derivatives can be applied inside CUDA kernels. We underline the effectiveness and the applicability of AutoMat by integrating it into the FFT-based homogenization scheme of Moulinec and Suquet and discuss the benefits of using AutoMat with respect to runtime and solution accuracy for an elasto-viscoplastic example.
When considering complex systems, identifying the most important actors is often of relevance. When the system is modeled
as a network, centrality measures are used which assign each node a value due to its position in the network. It is often
disregarded that they implicitly assume a network process flowing through a network, and also make assumptions of how
the network process flows through the network. A node is then central with respect to this network process (Borgatti in Soc
Netw 27(1):55–71, 2005, https ://doi.org/10.1016/j.socne t.2004.11.008). It has been shown that real-world processes often
do not fulfill these assumptions (Bockholt and Zweig, in Complex networks and their applications VIII, Springer, Cham,
2019, https ://doi.org/10.1007/978-3-030-36683 -4_7). In this work, we systematically investigate the impact of the measures’
assumptions by using four datasets of real-world processes. In order to do so, we introduce several variants of the betweenness
and closeness centrality which, for each assumption, use either the assumed process model or the behavior of the real-world
process. The results are twofold: on the one hand, for all measure variants and almost all datasets, we find that, in general,
the standard centrality measures are quite robust against deviations in their process model. On the other hand, we observe a
large variation of ranking positions of single nodes, even among the nodes ranked high by the standard measures. This has
implications for the interpretability of results of those centrality measures. Since a mismatch of the behaviour of the real
network process and the assumed process model does even affect the highly-ranked nodes, resulting rankings need to be
interpreted with care.
Since the h-index has been invented, it is the most frequently discussed bibliometric value and one of the most commonly used metrics to quantify a researcher’s scientific output. The more it is increasingly gaining popularity to use the metric as an indication of the quality of a job applicant or an employee the more important it is to assure its correctitude. Many platforms offer the h-index of a scientist as a service, sometimes without the explicit knowledge of the respective person. In this article we show that looking up the h-index for a researcher on the five most commonly used platforms, namely AMiner, Google Scholar, ResearchGate, Scopus and Web of Science, results in a variance that is in many cases as large as the average value. This is due to the varying definitions of what a scientific article is, the underlying data basis, and different qualities of the entity recognition problem. To perform our study, we crawled the h-index of the worlds top researchers according to two different rankings, all the Nobel Prize laureates except Literature and Peace, and the teaching staff of the computer science department of the TU Kaiserslautern Germany with whom we additionally computed their h-index manually. Thus we showed that the individual h-indices differ to an alarming extent between the platforms. We observed that researchers with an extraordinary high h-index and researchers with an index appropriate to the scientific career path and the respective scientific field are affected alike by these problems.
Weak memory consistency models capture the outcomes of concurrent
programs that appear in practice and yet cannot be explained by thread
interleavings. Such outcomes pose two major challenges to formal
methods. First, establishing that a memory model satisfies its
intended properties (e.g., supports a certain compilation scheme) is
extremely error-prone: most proposed language models were initially
broken and required multiple iterations to achieve soundness. Second,
weak memory models make verification of concurrent programs much
harder, as a result of which there are no scalable verification
techniques beyond a few that target very simple models.
This thesis presents solutions to both of these problems.
First, it shows that the relevant metatheory of weak memory
models can be effectively decided (sparing years of manual proof
efforts), and presents Kater, a tool that can answer metatheoretic
queries in a matter of seconds. Second, it presents GenMC, the first
(and only) scalable stateless model checker that is parametric in the
choice of the memory model, often improving the prior state of the art
by orders of magnitude.
In one-dimensional (1-D) Ultrasound (US) measurements, signals are
acquired that form the basis of more sophisticated two-dimensional (2-D) or
three-dimensional (3-D) US imaging. These 1-D signals contain a lot of raw
information about the US wave propagation and interaction with the
medium that is only processed in parts during image generation. While
image representations are easy to interpret for humans, the analysis of US
wave signals is hard to perform without applying algorithms to extract
desired features.
This work investigates reliable and fast 1-D US signal classifications to
distinguish between different stages or states in biomedical US scenarios and
shows how the new field of Machine Learning (ML) on raw US wave data
provides advantages and different applications. To achieve good results, the
input signals are treated as time series, which requires the deployment of
comparatively complex Time Series Classification (TSC) algorithms.
The literature shows that a lot of research efforts have previously only
tackled the classification and segmentation of US Brightness mode (B-Mode)
images, while neglecting approaches to classify 1-D signals to a large extent.
This research contributes by developing, deploying and evaluating
classification approaches for three distinct biomedical US classification tasks
and finds that respective signal classifications for different scenarios are
possible with varying degrees of accuracies. It entails the comparison of
several combinations of data types (e.g. temporal, spectral and statistical
features or raw signals), ML models and pre-processing steps to provide a
strong foundation for robust, binary classifications of 1-D US signals for
scenarios based on low-cost wearable, mobile and stationary devices. This
research addresses scientific questions not answered before by informing on
detailed descriptions of beneficial domain specific knowledge (domain specific
knowledge (DSK)), achieved accuracies and times needed for training and
evaluation of the examined ML models.
The resulting ML pipelines includes solutions based on data acquired from
custom experimental setups or clinical trials. Possible real-world applications
might include muscle contraction trackers, muscle fatigue detectors,
epiphyseal radius bone closure detectors or devices providing information
about advanced liver disease stages.
Automated machine-assisted
classifications requiring as little DSK as possible from the end user enable
application scenarios ranging from fitness or rehabilitation trackers as
consumer devices to solutions providing diagnostic support without requiring
extensive knowledge from professional medical practitioners. For example,
decision support systems for bone age assessments in clinical use or liver
health assessment systems for gastroenterologists.
This work shows that reliable, robust and fast classifications based on 1-D
US signals are possible with high degrees of accuracies depending on the
examined scenario with achieved F 1 -scores ranging from ≈ 70% to ≈ 87%.
These results prove that real-life applications for recreational purposes are
already possible and that critical applications for clinical use are highly likely
to be achieved once the presented approaches are further optimized in the future.
The field of 3D reconstruction is one of the most important areas in computer
vision. It is not only of theoretical importance, but it is also increasingly
used in practical applications, be it in reverse engineering, quality control or
robotics. In practical applications, where high precision reconstructions are
required for a large variety of different objects, structured light reconstruction
is often the method of choice. It allows to achieve accurate and dense
point correspondences over the entire scene, regardless of object texture or
features. Techniques that project phase-shifted sinusoidals are widely used
because, based on the harmonic addition theorem, they theoretically allow
surface encoding in full camera resolution invariant to the object’s coloring.
In this thesis, a fully-automatic reconstruction pipeline based on the sinusoidal
structured light technique is presented. From the projection of the
fringe patterns for encoding the object’s surface, the robust matching of the
point correspondences in sub-pixel accuracy, the auto-calibration of the setup
including the active device, up to the fully-automatic alignment of the partial
reconstructions, all steps will be described and examined in detail. During
that, improvements will be achieved in the area of matching, obtaining highly
accurate and topologically consistent correspondences in sub-pixel precision
between all the devices used. Furthermore, the auto-calibration from point
correspondences, based on the epipolar geometry of the structured light system
is improved. Weaknesses of previous methods in the extraction of focal
lengths from the fundamental matrices are discovered and addressed. The partial
point clouds, reconstructed from the auto-calibrated devices, are finally
pre-aligned using a neural network approach, based on light-resistant optical
flow estimation and subsequently refined using a global approach.
The weaknesses of the structured light method itself will also be addressed
and partially fixed during the course of this work. Since it is an active reconstruction method, certain surface properties can affect the quality of the
reconstruction. It will be shown how these problems can be eliminated or at
least be reduced using an iterative approach that combines fringe patterns with
an inverse texture. Another weakness of the method is its time-consuming acquisition procedure. Typically, a large number of horizontal and vertical fringe
patterns are projected onto the scene to achieve high-precision encoding despite
the limited dynamic range and resolution of the projector. Therefore, a
method will be presented which allows to combine the horizontal and vertical
patterns for a simultaneous two dimensional surface encoding.
During our daily lives, we are confronted with vast amounts of data, the processing of which can dramatically influence our lives, both positively and negatively. The enormous amount of data (images, texts, tables, and time series), its variety and possible applications are not always obvious. Due to advancements in the internet of things (IoT), there exist billions of sensors that produce time series which can be found everywhere, whether in medicine, the financial sector or the agricultural economy. This incredible amount of time series data has many hidden features which are useful for industry as well as for daily use, e.g. improving the cancer prediction can save real human lives. Recently, several deep learning methods have been proposed for analyzing this time series data. However, due to their black box nature, their applicability is limited in critical sectors like medicine, finance, and communication. In addition, it is now a compulsion as per artificial intelligence (AI) Act and per General Data Protection Regulation (GDPR) to protect any sensitive data and provide explanations in safety-critical domains. To enable use of DNNs in a broader domain scope, this thesis presents a framework for privacy-preserved and interpretable time series analysis. TimeFrame consists of four main components, namely, post-hoc interpretability, intrinsic interpretability, direct privacy, and indirect privacy. Interpretability is indispensable to avoid damaging people or the infrastructure. In the past years, the development mostly focused on image data, which prevented the full potential of DNNs in time series processing from being exploited. To overcome this limitation, TimeFrame introduces five (Time to Focus, TSViz, TimeREISE, TSInsight, Data Lens) novel post-hoc and two (PatchX, P2ExNet) novel intrinsic interpretability components. TimeFrame addresses multiple perspectives such as attribution, compression, visualization, influence, prototyping, and hierarchical splitting. Compared to existing methods, the components show better explanations, robustness, and scalability. Another crucial factor is the privacy when dealing with sensitive data and deep learning. In this context, TimeFrame introduces two (PPML, PPML x XAI) components for direct and one (From Private to Public) component for indirect privacy. These components benchmark privacy approaches, their effect on interpretability, and the synthetic generation of data to overcome privacy concerns. TimeFrame offers a large set of interpretability and privacy components that can be combined and consider numerous different aspects. Furthermore, the novel approaches have shown to consistently outperform twenty existing state-of-the-art methods across up to 20 different datasets. To guarantee the fairness, various metrics were used including performance change, Sensitivity, Infidelity, Continuity, runtime, model dependency, compression rate, and others. This broad set of metrics makes it possible to provide guidelines for a more appropriate use of existing state-of-the-art approaches as well as the novel components included in TimeFrame.
Highly Automated Driving (HAD) vehicles represent complex and safety critical systems. They are deployed in an open context i.e., an intricate environment which undergoes continual changes. The complexity of these systems and insufficiencies in sensing and understanding the open context may result in unsafe and uncertain behaviour. The safety critical nature of the HAD vehicles requires modelling of root causes for unsafe behaviour and their mitigation to argue sufficient reduction of residual risk.
Standardization activities such as ISO 21448 provide guidelines on the Safety Of The Intended Functionality (SOTIF) and focus on the analysis of performance limitations under the influence of triggering conditions that can lead to hazardous behaviour. SOTIF references traditional safety analyses methods e.g., Failure Mode and Effect Analysis (FMEA) and Fault Tree Analysis (FTA) to perform safety analysis. These analyses methods are based on certain assumptions e.g., single point failure in FMEA and independence of basic events in FTA. Moreover, these analyses are generally based on expert knowledge i.e., data-based models or hybrid approaches (expert and data) are seldom practised. The resulting safety model is fixed i.e., it is generally seen as a one-time artefact. Open context environment may contain triggering conditions which may not be evident to the expert. Open context also evolves over time and new phenomena may emerge.
This thesis explores the applicability of the traditional safety analyses techniques to provide safety models for HAD vehicles operating in the open context, under the light of modelling assumptions taken by traditional safety analyses techniques. Moreover, incorporating uncertainties into safety analyses models is also explored. An explicit distinction between the inherent uncertainty of a probabilistic event (aleatory) and uncertainty due to lack of knowledge (epistemic) is made to formalize models to perform SOTIF analysis. A further distinction is made for conditions of complete ignorance and termed as ontological uncertainty. The distinction is important as for HAD vehicles operating in open context the ontological uncertainty can never be completely disregarded.
This thesis proposes a novel framework of SOTIF to model, estimate and dis cover triggering conditions relevant to performance limitations. The framework provides the ability to model uncertainties while also providing a hybrid approach i.e., supporting inclusion of expert knowledge as well as data driven engineering processes. Two representative algorithms are provided to support the framework. Bayesian Network (BN) and p-value hypothesis testing are utilised in this regard. The framework is implemented on a real-world case study in which LIDARs based perception systems are used as vehicle detection system.
This doctoral dissertation is comprised of nine published articles covering different
methods for ‘Fast, Robust Rigid and Non-Rigid Registration for Globally Consistent
3D Scene and Shape Reconstruction’. Overall the contributing articles are separated
and discussed in three stages – The first part of the thesis i.e., chapter 2 explains
three novel method classes of rigid point set registration namely Gravitational Approach (GA), Fast Gravitational Approach (FGA), and RPSRNet. GA was introduced as the first physics-based rigid point set registration. It includes elegant modeling of rigid by dynamics using Newtonian mechanics. The method proposed many new avenues for other types of pattern matching tasks thank point set registration. Next, FGA method, published 4 years after GA presented as an extension that breaks the algorithmic complexity of GA from O(M N ) to O(M log N ) using Barnes-Hut tree representation of point cloud. It also eliminates the requirement of heuristic optimization parameter settings by GA, and achieve state-of-the-art alignment accuracy on LiDAR odometry. Finally, RPSRNet presents deep learning version of FGA, with custom convolution layers for hierarchical point feature embedding. RPSRNet is robust and the fastest among SoA methods for LiDAR data registration. The second part, i.e., chapter 3, of the thesis introduces NRGA as the fist physics-based non-rigid point set
registration method which is computationally slow but robust against noisy and partial inputs. NRGA preserves structural consistency as it coherently regularize motion of deformable vertices. For articulated hand shape reconstruction, a tailored version of NRGA -- Articulated-NRGA -- is effective to refine final hand shape. Collision and penetration avoidance between source and target surfaces are tackled by constrained optimization in NRGA. This setting has improved hand and object interaction reconstruction. Next contribution FoldMatch method remodels the shape deformation by introducing wrinkle vector field (WVF) for capturing complex clothing and garment details while fitting body models onto 3D Scans. Quantitative evaluation of FoldMatch and NRGA shows their effectiveness in geometrically consistent surface modeling and reconstruction tasks. Finally, the third part of the thesis explains globally consistent outdoor scene reconstruciton, odometry estimation, and uncertainty guided pose-graph optimization in a novel LiDAR-based localization and map building method, called Deep Evidential LiDAR Odometry (DELO). This is the first Odometry method to use predictive uncertainty modeling for sensor pose prediction network.
From industrial fault detection to medical image analysis or financial fraud prevention: Anomaly detection—the task of identifying data points that show significant deviations from the majority of data—is critical in industrial and technological applications. For efficient and effective anomaly detection, a rich set of semantic features are required to be automatically extracted from the complex data. For example, many recent advances in image anomaly detection are based on self-supervised learning, which learns rich features from a large amount of unlabeled complex image data by exploiting data augmentations. For image data, predefined transformations such as rotations are used to generate varying views of the data. Unfortunately, for data other than images, such as time series, tabular data, graphs, or text, it is unclear what are suitable transformations. This becomes an obstacle to successful self-supervised anomaly detection on other data types.
This thesis proposes Neural Transformation Learning, a self-supervised anomaly detection method that is applicable to general data types. In contrast to previous methods relying on hand-crafted transformations, neural transformation learning learns the transformations from data and uses them for detection. The key ingredient is a novel objective that encourages learning diverse transformations while preserving the relevant semantic content of the data. We prove theoretically and empirically that it is more suited than existing objectives for transformation learning.
We also introduce the extensions of neural transformation learning for anomaly detection within time series and graph-level anomaly detection. The extensions combine transformation learning and other learning paradigms to incorporate vital prior knowledge about time series and graph data. Moreover, we propose a general training strategy for deep anomaly detection with contaminated data. The idea is to infer the unlabeled anomalies and utilize them for updating parameters alternatively. In setups where expert feedback is available, we present a diverse querying strategy based on the seeding algorithm of K-means++ for active anomaly detection.
Our extensive experiments and analysis demonstrate that neural transformation learning achieves remarkable and robust anomaly detection performance on various data types. Finally, we outline specific paths for future research.
Semi-structured data is a common data format in many domains.
It is characterized by a hierarchical structure and a schema that is not fixed.
Efficient and scalable processing of this data is therefore challenging, as many existing indexing and processing techniques are not well-suited for this data format.
This dissertation presents a novel approach to processing large JSON datasets.
We describe a new data processor, JODA, that is designed to process semi-structured data by using all available computing resources and state-of-the-art techniques.
Using a custom query language and a vertically-scaling pipeline query execution engine, JODA can process large datasets with high throughput.
We optimize JODA by using a novel optimization for iterative query workloads called delta trees, which succinctly represent the changes between two documents.
This allows us to process iterative and exploratory queries efficiently.
We improve the filtering performance of JODA by implementing a holistic adaptive indexing approach that creates and improves structural and content indices on the fly, depending on the query load.
No prior knowledge about the data is required, and the indices are automatically improved over time.
JODA is also modularized and can be extended with new user-defined predicates, functions, indices, import, and export functionalities.
These modules can be written in an external programming language and integrated into the query execution pipeline at runtime.
To evaluate this system against competitors, we introduce a benchmark generator, coined BETZE, which aims to simulate data scientists exploring unknown JSON datasets.
The generator can be tweaked to generate query workload with different characteristics, or predefined presets can be used to quickly generate a benchmark.
We see that JODA outperforms competitors in most tasks over a wide range of datasets and use-cases.
3D joint angles based human pose is needed for applications like activity recognition, musculoskeletal health, sports biomechanics and ergonomics. The microelectromechanical systems (MEMS) based magnetic-inertial measurement units (MIMUs) can estimate 3D orientation. Due to small size, MIMUs can be attached to the body as wearable sensors for obtaining full 3D human pose and this system is termed as inertial motion capture (i-Mocap). But the MIMUs suffer from sensor errors and disturbances, due to which orientation estimated from individual MIMUs can be erroneous. Accurate sensor calibration is essential and subsequently alignment of these sensors to body segments must also be precisely known, which is called sensor-to-segment calibration. Sensor fusion is employed to address the disturbances and noise in MIMUs. Many state-of-art inertial motion capture approaches ignore the magnetometer and only use IMUs to reduce the error arising from inhomogeneous magnetic field. These algorithms rely on kinematic constraints and assumptions regarding joints and are based on IMUs located on the adjacent body segments. The full body coverage requires 13-17 such units and can be quite obtrusive. The setting up and calibration of so many wearable sensors also take time.
This thesis focuses on 3D human pose estimation from a reduced number of MIMUs and deals with this problem systematically. First we propose an accurate simultaneous calibration of multiple MIMUs, which also learns the uncertainty of individual sensors. We then describe a novel sensor fusion algorithm for robust orientation estimation from an MIMU and for updating sensors calibration online. The residual errors in both sensor calibration and fusion can result in drift error in the joint angles. Therefore, we present anatomical (sensor-to-segment) calibration in which an orientation offset correction term is updated and used for online correction of residual drift in individual joint angles. Subsequently we demonstrate that 3D human joint angle constraints can be learned using a data-driven approach in a high dimensional latent space. Owing to temporal and joint angle constraints, it is possible to use only a reduced set of sensors (as opposed to one sensor per segment) and still obtain 3D human pose. But the spatial and temporal prior learning from data is often limited due to finite set of movement patterns in most datasets. This introduces uncertainty while estimating 3D human pose from sparse MIMU sensors. We propose a magnetometer robust orientation parameterization and a data-driven deep learning framework to predict 3D human pose with associated uncertainty from sparse MIMUs. The model is evaluated on real MIMU data and we show that the uncertainty predicted by the trained model is well-correlated with actual error and ambiguity.
Though Computer Aided Design (CAD) and Simulation software are mature, well established, and in wide professional use, modern design and prototyping pipelines are challenging the limits of these tools. Advances in 3D printing have brought manufacturing capability to the general public. Moreover, advancements in Machine Learning and sensor technology are enabling enthusiasts and small companies to develop their own autonomous vehicles and machines. This means that many more users are designing (or customizing) 3D objects in CAD, and many are testing machine autonomy in Simulation. Though Graphical User Interfaces (GUIs) are the de-facto standard for these tools, we find that these interfaces are not robust and flexible. For example, designs made using GUI often break when customized, and setting up large simulations can be quite tedious in GUI. Though programmatic interfaces do not suffer from these limitations, they are generally quite difficult to use, and often do not provide appropriate abstractions and language constructs.
In this Thesis, we present our work on bridging the ease of use of GUI with the robustness and flexibility of programming. For CAD, we propose an interactive framework that automatically synthesizes robust programs from GUI-based design operations. Additionally, we apply program analysis to ensure customizations do not lead to invalid objects. Finally, for simulation, we propose a novel programmatic framework that simplifies building of complex test environments, and a test generation mechanism that guarantees good coverage over test parameters. Our contributions help bring some of the advantages of programming to traditionally GUI-dominant workflows. Through novel programmatic interfaces, and without sacrificing ease of use, we show that the design and customization of 3D objects can be made more robust, and that the creation of parameterized simulations can be simplified.
Faces deliver invaluable information about people. Machine-based perception can be of a great benefit in extracting that underlying information in face images if the problem is properly modeled. Classical image processing algorithms may fail to handle the diverse data available today due to several challenges related to varying capturing locations, and conditions. Advanced machine learning methods and algorithms are now highly beneficial due to the rapid development of powerful hardware, enabling feasible advanced solutions based on data learning and summarization into powerful models. In this thesis, novel solutions are provided to the problems of head orientation estimation and gender prediction. Initially, classical machine learning algorithms were used to address head orientation estimation but were limited by their inability to handle large datasets and poor generalization. To overcome these challenges, a new highly accurate head pose dataset was acquired to tackle the identified problems. Novel trained deep neural networks have been exploited, that use the acquired data and provide novel architectures. The information about head pose is then represented in the network weights, thus, allowing predicting the head orientation angles given a new unseen face. The acquired dataset, named AutoPOSE opens the door for further studies in the field of computer vision and especially, face analysis. The problem of gender prediction has also been explored, but unlike humans who can easily identify gender from a face, computers face difficulties due to facial similarities. Therefore, hand-crafted features are not effective for generalization. To address this, a new deep learning method was developed and evaluated on multiple public datasets, with identified challenges in both still images and videos addressed. Finally, the effect of facial appearance changes due to head orientation variation has been investigated on gender prediction accuracy. A novel orientation-guided feature maps recalibration method is presented, that significantly increased the accuracy of gender prediction.
In conclusion, two problems have been addressed in this thesis, independently and joined together. Existing methods have been enhanced with intelligent pre-processing methods and new approaches have been introduced to tackle existing challenges, that arise from pose, illumination, and occlusion variations. The proposed methods have been extensively evaluated, showing that head orientation and gender prediction can be estimated with high accuracy using machine learning-based methods. Also, the evaluations showed that the use of head orientation information consistently improved the gender prediction accuracy. Scientific contributions have been presented, and the new acquired highly accurate dataset motivates the research community to push the state-of-the-art forward.
Undocumented enterprise data can easily pile up in companies in form of datasets and personal information. In absence of a data management strategy, such data becomes rather messy and may not fit for its intended use. Since there is often no documentation available, only a limited number of domain experts are aware of its contents. Therefore, for companies it becomes increasingly difficult to use such data to its full potential. To provide a solution, this PhD thesis investigates the construction of enterprise and personal knowledge graphs by semantically enriching messy data with meaning using semantic technologies. Since real world entities and their interrelations are organized in a graph, knowledge graphs serve as a semantic bridge between domain conceptualization and raw data. Spreadsheets are a prominent example of such enterprise data, since they are widely used by knowledge workers in the industrial sector. Two distinct approaches are investigated to construct knowledge graphs from them: a global extraction & annotation method and a local mapping technique. The latter is further complemented with a predictor of mapping rules on messy data. Different human-in-the-loop strategies are considered to include experts depending on their user group. Since non-technical users usually lack understanding of semantic technologies, they need appropriate tools to be able to give feedback. In case of developers, approaches are proposed to close the technology gap between industry and Semantic Web related concepts. Semantic Web practitioners participate with ontology modeling and linked data applications. Enterprise and personal data is typically confidential which is why it cannot be shared with a research community to discuss its challenges. However, for evaluation and reproducibility reasons publicly available datasets are mandatory. The thesis proposes ways to generate synthetic datasets with the goal to be as authentic as possible. Besides that, for internal evaluations a crawler of personal data on desktops is implemented. There are further contributions related to this thesis in diverse domains. One is about the motivation to support users in their daily work using personal knowledge assistants. Others are the agricultural field and the data science domain which also benefit from knowledge graph approaches. In conclusion, this PhD thesis contributes to the construction of knowledge graphs from especially messy enterprise data, while users from different groups take part in this process in various ways.
This thesis focuses on novel methods to establish the utility of wearable devices along with machine learning and pattern recognition methods for formal education and address the open research questions posed by existing methods. Firstly, state-of-the-art methods are proposed to analyse the cognitive activities in the learning process, i.e., reading, writing, and their correlation. Furthermore, this thesis presents real-time applications in wearable space as an experimental tool in Physics education, and an air-writing system.
There are two critical components in analysing the reading behaviour, i.e., WHERE a person looks at (gaze analysis) and WHAT a person looks at (content analysis). This thesis proposes novel methods to classify the reading content to address the WHAT AT component. The proposed methods are based on a hybrid approach, which fuses the traditional computer vision methods with deep neural networks. These methods, when evaluated on publicly available datasets, yield state-of-the-art results to define the structure of the document images. Moreover, extensive efforts were made to refine and correct ICDAR2017-POD dataset along with a completely new FFD dataset.
Traditionally, handwriting research focuses on character and number recognition without looking into the type of writing, i.e. text, math, and drawing. This thesis reports multiple contributions for on-line handwriting classification. First, it presents a public dataset for on-line handwriting classification OnTabWriter, collected using iPen and an iPad. In addition, a new feature set is introduced for on-line handwriting classification to establish the benchmark on the proposed dataset to classify handwriting as plain text, mathematical expression, and plot/graph. An ablation study is made to evaluate the performance of the proposed feature set in comparison to existing feature sets. Lastly, this thesis evaluates the importance of context for on-line handwriting classification.
Analysing reading and writing activities individually is not enough to provide insights to identify the student's expertise unless their correlations are analysed. This thesis presents a study where reading data from wearable eye-trackers and writing data from sensor pen are analysed together in correlation to correlate the expertise of the users in Physics education with their actual knowledge. Initial results show a strong correlation between individual's expertise and understanding of the subject.
Augmented reality & virtual applications can play a vital role in making classroom environments more interactive and engaging both for teachers and learners. To validate the hypothesis, different applications are developed and evaluated. First, smart glasses are used as an experimental tool in Physics education to help the learners perform experiments by providing assistance and feedback on head mounted display in understanding acoustics concepts. Second, a real-time application of air-writing with the finger on an imaginary canvas using a single IMU as the FAirWrite system is also presented. FAirWrite system is further equipped with DL methods to classify the air-written characters.
Due to its performance, the field of deep learning has gained a lot of attention, with neural networks succeeding in areas like \( \textit{Computer Vision} \) (CV), \( \textit{Neural Language Processing} \) (NLP), and \( \textit{Reinforcement Learning} \) (RL). However, high accuracy comes at a computational cost as larger networks require longer training time and no longer fit onto a single GPU. To reduce training costs, researchers are looking into the dynamics of different optimizers, in order to find ways to make training more efficient. Resource requirements can be limited by reducing model size during training or designing more efficient models that improve accuracy without increasing network size.
This thesis combines eigenvalue computation and high-dimensional loss surface visualization to study different optimizers and deep neural network models. Eigenvectors of different eigenvalues are computed, and the loss landscape and optimizer trajectory are projected onto the plane spanned by those eigenvectors. A new parallelization method for the stochastic Lanczos method is introduced, resulting in faster computation and thus enabling high-resolution videos of the trajectory and second-order information during neural network training. Additionally, the thesis presents the loss landscape between two minima along with the eigenvalue density spectrum at intermediate points for the first time.
Secondly, this thesis presents a regularization method for \( \textit{Generative Adversarial Networks} \) (GANs) that uses second-order information. The gradient during training is modified by subtracting the eigenvector direction of the biggest eigenvalue, preventing the network from falling into the steepest minima and avoiding mode collapse. The thesis also shows the full eigenvalue density spectra of GANs during training.
Thirdly, this thesis introduces ProxSGD, a proximal algorithm for neural network training that guarantees convergence to a stationary point and unifies multiple popular optimizers. Proximal gradients are used to find a closed-form solution to the problem of training neural networks with smooth and non-smooth regularizations, resulting in better sparsity and more efficient optimization. Experiments show that ProxSGD can find sparser networks while reaching the same accuracy as popular optimizers.
Lastly, this thesis unifies sparsity and \( \textit{neural architecture search} \) (NAS) through the framework of group sparsity. Group sparsity is achieved through \( \ell_{2,1} \)-regularization during training, allowing for filter and operation pruning to reduce model size with minimal sacrifice in accuracy. By grouping multiple operations together, group sparsity can be used for NAS as well. This approach is shown to be more robust while still achieving competitive accuracies compared to state-of-the-art methods.
In recent years, the formal methods community has made significant progress towards the development of industrial-strength static analysis tools that can check properties of real-world production code. Such tools can help developers detect potential bugs and security vulnerabilities in critical software before deployment. While the potential benefits of static analysis tools are clear, their usability and effectiveness in mainstream software development workflows often comes into question and can prevent software developers from using these tools to their full potential. In this dissertation, we focus on two major challenges that can limit their ability to be incorporated into software development workflows.
The first challenge is unintentional unsoundness. Static program analyzers are complicated tools, implementing sophisticated algorithms and performance heuristics. This makes them highly susceptible to undetected unintentional soundness issues. These issues in program analyzers can cause false negatives and have disastrous consequences e.g., when analyzing safety critical software. In this dissertation, we present novel techniques to detect unintentional unsoundness bugs in two foundational program analysis tools namely SMT solvers and Datalog engines. These tools are used extensively by the formal methods community, for instance, in software verification, systematic testing, and program synthesis. We implemented these techniques as easy-to-use open source tools that are publicly available on Github. With the proposed techniques, we were able to detect more than 55 unique and confirmed critical soundness bugs in popular and widely used SMT solvers and Datalog engines in only a few months of testing.
The second challenge is finding the right balance between soundness, precision, and perfor- mance. In an ideal world, a static analyzer should be as precise as possible while maintaining soundness and being sufficiently fast. However, to overcome undecidability issues, these tools have to employ a variety of techniques to be practical for example, compromising on the sound- ness of the analysis or approximating code behavior. Static analyzers therefore are not trivial to integrate into any usage scenario with different program sizes, resource constraints and SLAs. Most of the times, these tools also don’t scale to large industrial code bases containing millions of lines of code. This makes it extremely challenging to get the most out of these analyzers and integrate them into everyday development activities, especially for average software develop- ment teams with little to no knowledge or understanding of advanced static analysis techniques. In this dissertation we present an approach to automatically tailor an abstract interpreter to the code under analysis and any given resource constraints. We implemented our technique as an open source framework, which is publicly available on Github. The second contribution of this dissertation in this challenge area is a technique to horizontally scale analysis tools in cloud-based static analysis platforms by splitting the input to the analyzer into partitions and analyzing the partitions independently. The technique was developed in collaboration with Amazon Web Services and is now being used in production in their CodeGuru service.
The rising demand for machine learning (ML) models has become a growing concern for stakeholders who depend on automatic decisions. In today's world, black-box solutions (in particular deep neural networks) are being continuously implemented for more and more high-stake scenarios like medical diagnosis or autonomous vehicles. Unfortunately, when these opaque models make predictions that do not align with our expectations, finding a valid justification is simply not possible.
Explainable Artificial Intelligence (XAI) has emerged in response to our need for finding reasons that justify what a machine sees, but we don't. However, contributions in this field are mostly centered around local structures such as individual neurons or single input samples. Global characteristics that govern the behavior of a model are still poorly understood or have not been explored yet. An aggravating factor is the lack of a standard terminology to contextualize and compare contributions in this field. Such lack of consensus is depriving the ML community from ultimately moving away from black-boxes, and start creating systematic methods to design models that are interpretable by design.
So, what are the global patterns that govern the behavior of modern neural networks, and what can we do to make these models more interpretable from the start?
This thesis delves into both issues, unveiling patterns about existing models, and establishing strategies that lead to more interpretable architectures. These include biases coming from imbalanced datasets, quantification of model capacity, and robustness against adversarial attacks. When looking for new models that are interpretable by design, this work proposes a strategy to add more structure to neural networks, based on auxiliary tasks that are semantically related to the main objective. This strategy is the result of applying a novel theoretical framework proposed as part of this work. The XAI framework is meant to contextualize and compare contributions in XAI by providing actionable definitions for terms like "explanation" and "interpretation."
Altogether, these contributions address dire demands for understanding more about the global behavior of modern deep neural networks. More importantly, they can be used as a blueprint for designing novel, and more interpretable architectures. By tackling issues from the present and the future of XAI, results from this work are a firm step towards more interpretable models for computer vision.
With the ever-increasing amount of satellite-backed communication, constellations covering the entire world, and the rise of Software Defined Radios (SDRs), satellite signals have already become prime targets for scientific research all over the globe. However, due to logistical challenges like capture time/location and peripheral/system management for the sensors and the wide variety of protocols/encoding schemes used, no one-fits-all sniffing solution exists for capturing their wide variety of signals. Therefore, this thesis aims to analyze, design, and implement a system that makes it possible to study LEO (Low Earth Orbit) L-Band satellite signals with readily available Single Board Computers (SBCs) in a widely distributed, location, and time-aware way. The key design factors were useability, maintainability, adaptability, and security in a centrally managed client-server architecture. The research presented yielded a Satellite probe Operating System called SATOS, which aims to implement on-sensor data decoding driven by GNU Radio and secure Over The Air (OTA) updates inside the Buildroot build environment. Its intended use case is the future deployment of DISCOSAT on a university working group scale.
Processing data streams is a classical and ubiquitous problem.
A query is registered against a potentially endless data stream and continuously delivers results as tuples stream in.
Modern stream processing systems allow users to express queries in different ways.
However, when a query involves joins between multiple input streams, the order of these joins is not transparently optimized.
In this thesis, we explore ways to optimize multi-way theta joins, where the join predicates are not limited to equality and multiple inputs are referenced.
We put forward a novel operator, MultiStream, which joins multiple input streams using iterative probing and bringing minimal materialization effort in.
The order in which tuples are sent inside a MultiStream operator is optimized using a cost-based model.
Further, a query can be answered using an multi-way tree comprising multiple MultiStream operators where each inner operator represents a materialized intermediate result.
We integrate equi-joins in MultiStream to reduce communication, such that mixed queries of theta and equality predicates are supported.
Streaming queries are long-standing and thus multiple queries might be registered at the system at the same time.
Hence, we research joint answering of multiple multi-way join queries and optimize the global ordering using integer linear programming.
All these approaches are implemented in CLASH, a system for generating Apache Storm topologies including runtime components that enables users to pose queries in a declarative way and let the system craft the suitable topology.
An Efficient Automated Machine Learning Framework for Genomics and Proteomics Sequence Analysis
(2023)
Genomics and Proteomics sequence analyses are the scientific studies of understanding the language of Deoxyribonucleic Acid (DNA), Ribonucleic Acid (RNA) and protein biomolecules with an objective of controlling the production of proteins and understanding their core functionalities. It helps to detect chronic diseases in early stages, root causes of clinical changes, key genetic targets for pharmaceutical development and optimization of therapeutics for various age groups. Most Genomics and Proteomics sequence analysis work is performed using typical wet lab experimental approaches that make use of different genetic diagnostic technologies. However, these approaches are costly, time consuming, skill and labor intensive. Hence, these approaches slow down the process of developing an efficient and economical sequence analysis landscape essential to demystify a variety of cellular processes and functioning of biomolecules in living organisms. To empower manual wet lab experiment driven research, many machine learning based approaches have been developed in recent years. However, these approaches cannot be used in practical environment due to their limited performance. Considering the sensitive and inherently demanding nature of Genomics and Proteomics sequence
analysis which can have very far-reaching as well as serious repercussions on account of misdiagnosis, the main
objective of this research is to develop an efficient automated computational framework for Genomics and Proteomics sequence analysis using the predictive and prescriptive analytical powers of Artificial Intelligence (AI) to significantly improve healthcare operations.
The proposed framework is comprised of 3 main components namely sequence encoding, feature engineering and
discrete or continuous value predictor. The sequence encoding module is equipped with a variety of existing and newly developed sequence encoding algorithms that are capable of generating a rich statistical representation of DNA, RNA and protein raw sequences. The feature engineering module has diverse types of feature selection and dimensionality reduction approaches which can be used to generate the most effective feature space. Furthermore, the discrete and/or continuous value predictor module of the proposed framework contains a wide range of existing machine learning and newly developed deep learning regressors and classifiers. To evaluate the integrity and generalizability of the proposed framework, we have performed a large-scale experimentation over diverse types of Genomics and Proteomics sequence analysis tasks (i.e., DNA, RNA and proteins).
In Genomics analysis, Epigenetic modification detection is one of the key component. It helps clinical researchers and practitioners to distinguish normal cellular activities from malfunctioned ones, which can lead to diverse genetic disorders such as metabolic disorders, cancers, etc. To support this analysis, the proposed framework is used to solve the problem of DNA and Histone modification prediction where it has achieved state-of-the-art performance on 27 publicly available benchmark datasets of 17 different species with best accuracy of 97%. RNA sequence analysis is another vital component of Genomics sequence analysis where the identification of different coding and non-coding RNAs as well as their subcellular localization patterns help to demystify the functions of diverse RNAs, root causes of clinical changes, develop precision medicine and optimize therapeutics. To support this analysis, the proposed framework is utilized for non-coding RNA classification and multi-compartment RNA subcellular localization prediction. Where it achieved state-of-the-art performance on 10 publicly available benchmark datasets of Homo sapiens and Mus Musculus species with best accuracy of 98%.
Proteomics sequence analysis is essential to demystify the virus pathogenesis, host immunity responses, the way
proteins affect or are affected by the cell processes, their structure and core functionalities. To support this analysis, the proposed framework is used for host protein-protein and virus-host protein-protein interaction prediction. It has achieved state-of-the-art performance on 2 publicly available protein protein interaction datasets of Homo Sapiens and Mus Musculus species with best accuracy of 96% and 7 viral host protein protein interaction datasets of multiple hosts and viruses with best accuracy of 94%. Considering the performance and practical significance of proposed framework, we believe proposed framework will help researchers in developing cutting-edge practical applications for diverse Genomic and Proteomic sequence analyses tasks (i.e., DNA, RNA and proteins).
Scientific research plays a crucial role in the development of a society. With ever-increasing volumes of scientific publications are now making it extremely challenging to analyze and maintain insights into the scientific communities like collaboration or citation trends and evolution of interests etc. This thesis is an effort towards using scientific publications to provide detailed insights into a scientific community from a range of aspects. The contribution of this thesis is five-fold.
Firstly, this thesis proposes approaches for automatic information extraction from scientific publications. The proposed layout-based approach for this purpose is inspired by how human beings perceive individual references relying only on visual queues. The proposed approach significantly outperforms the existing text-based techniques and is independent of any domain or language.
Secondly, this thesis tackles the problem of identifying meaningful topics from a given publication as the keywords provided in the publication are not always accurate representatives of the publication topic. To rectify this problem, this thesis proposes a state-of-the-art keywords extraction approach that employs a domain ontology along with the detected keywords to perform topic modeling for a given set of publications.
Thirdly, this thesis analyses the disposition of each citation to understand its true essence. For this purpose, we proposes a transformer-based approach for analyzing the impact of each citation appearing in a scientific publication. The impact of a citation can be determined by the inherent sentiment and intent of a citation, which refers to the assessment and motive of an author towards citing a scientific publication.
Furthermore, this thesis quantifies the influence of a research contributor in a scientific community by introducing a new semantic index for researchers that takes both quantitative and qualitative aspects of a citation into account to better represent the prestige of a researcher in a scientific community. Semantic Index is also evaluated for conformity to the guidelines and recommendations of various research funding organizations to assess the impact of a researcher.
In this thesis, all of the aforementioned aspects are packaged together in a single framework called Academic Community Explorer (ACE) 2.0, which automatically extracts and analyzes information from scientific publications and visualizes the insights using several interactive visualizations. These visualizations provide an instant glimpse into the scientific communities from a wide range of aspects with different granularity levels.
The generally unsupervised nature of autoencoder models implies that the main training metric is formulated as the error between input images and their corresponding reconstructions. Different reconstruction loss variations and latent space regularization have been shown to improve model performances depending on the tasks to solve and to induce new desirable properties like disentanglement. Nevertheless, measuring the success in, or enforcing properties by, the input pixel space is a challenging endeavor. In this work, we want to make more efficient use of the available data and provide design choices to be considered in the recording or generation of future datasets to implicitly induce desirable properties during training. To this end, we propose a new sampling technique which matches semantically important parts of the image while randomizing the other parts, leading to salient feature extraction and a neglection of unimportant details. Further, we propose to recursively apply a previously trained autoencoder model, which can then be interpreted as a dynamical system with desirable properties for generalization and uncertainty estimation.
The proposed methods can be combined with any existing reconstruction loss. We give a detailed analysis of the resulting properties on various datasets and show improvements on several computer vision tasks: image and illumination normalization, invariances, synthetic to real generalization, uncertainty estimation and improved classification accuracy by means of simple classifiers in the latent space.
These investigations are adopted in the automotive application of vehicle interior rear seat occupant classification. For the latter, we release a synthetic dataset with several fine-grained extensions such that all the aforementioned topics can be investigated in isolation, or together, in a single application environment. We provide quantitative evidence that machine learning, and in particular deep learning methods cannot readily be used in industrial applications when only a limited amount of variation is available for training. The latter can, however, often be the case because of constraints enforced by the application to be considered and financial limitations.
In recent years, deep learning has made substantial improvements in various fields like image understanding, Natural Language Processing (NLP), etc. These huge advancements have led to the release of many commercial applications which aim to help users carry out their daily tasks. Personal digital assistants are one such successful application of NLP, having a diverse userbase from all age groups. NLP tasks like Natural Language Understanding (NLU) and Natural Language Generation (NLG) are core components for building these assistants. However, like any other deep learning model, the growth of NLU & NLG models is directly coupled with tremendous amounts of training examples, which are expensive to collect due to annotator costs. Therefore, this work investigates the methodologies to build NLU and NLG systems in a data-constrained setting.
We evaluate the problem of limited training data in multiple scenarios like limited or no data available when building a new system, availability of a few labeled examples when adding a new feature to an existing system, and changes in the distribution of test data during the lifetime of a deployed system.
Motivated by the standard methods to handle data-constrained settings, we propose novel approaches to generate data and exploit latent representations to overcome performance drops emerging from limited training data.We propose a framework to generate high-quality synthetic data when few training examples are available for a newly added feature for dialogue agents. Our interpretation-to-text model uses existing training data for bootstrapping new features and improves the accuracy of downstream tasks of intent classification and slot labeling. Following, we study a few-shot setting and observe that generation systems face a low semantic coverage problem. Hence, we present an unsupervised NLG algorithm that ensures that all relevant semantic information is present in the generated text.
We also study to see if we really need all training examples for learning a generalized model. We propose a data selection method that selects the most informative training examples to train Visual Question Answering (VQA) models without erosion of accuracy. We leverage the already available inter-annotator agreement and design a diagnostic tool, called (EaSe), that leverages the entropy and semantic similarity of answer patterns.
Finally, we discuss two empirical studies to understand the feature space of VQA models and show how language model pre-training and exploiting multimodal embedding space allows for building data constrained models ensuring minimal or no accuracy losses.
In recent years, Augmented Reality has made its way into everyday devices. Most smartphones are AR-enabled, providing applications like pedestrian navigation, Point of Interest highlighting, gaming, and retail. The high-tech industry has been focused on developing smartglasses to present virtual elements directly in front of the viewers’ eyes, allowing more immersive AR experiences. Smartglasses can also be deployed while driving for an enhanced and more safe experience. A 3D registered augmentation of the real world with navigation arrows, lane highlighting, or warnings can decrease the duration of inattentiveness regarding driving due to glancing at other screens. Enabling HMDs’ usage inside cars requires knowing its exact position and orientation (6-DoF pose) in the car. This necessitates sensors either built inside the AR glasses or the car. In a car, the latter option called outside-in tracking is more attractive due to two reasons. First, AR glasses containing different sensor sets exist, hampering finding one single solution for different HMDs. Second, the view from the driver’s perspective combines static interior and dynamic exterior features, complicating finding a reliable set of features. Nowadays, tracking methods utilize Deep Learning for a more generalizable and accurate derivation of the 6-DoF pose. They achieve outstanding results for head and object pose estimation. In this thesis, we present Deep Learning-based in-car 6-DoF AR glasses pose estimation approaches. The goal of the work is an exploration of accurate HMD pose estimation with the help of neural networks. The thesis achieves this by investigating numerous pose estimation techniques. Evaluations on the recorded HMDPose dataset constitute the foundation for this, consisting of infrared images of drivers wearing different HMD models. First, algorithms based on images are derived and evaluated on the dataset. For comparison, we carried out an evaluation on image-based methods considering time information. Further, pose estimation based on point clouds, generated out of infrared images, are analyzed. An investigation of various head pose estimation methods to derive its potential use are conducted. In conclusion, we introduce several highly accurate AR glasses pose estimators. The HMD pose alone achieves better results than the head pose and the combination of the head and HMD. Especially our image-based methods with optional usage of time information can efficiently and accurately regress the AR glasses pose. Our algorithms show excellent estimation results on live data when deployed inside a car, making seamless in-car HMD usage possible in the future.
The wireless spectrum is already a scarce good, shared by multiple competing technologies such as Bluetooth, ZigBee and Wi-Fi, and the hunger for traffic is only increasing. Due to the heterogeneity of the existing wireless technologies and the real threat that interference poses to network performance, sophisticated techniques must be developed to ensure acceptable levels of quality of service.
In this thesis, we present a passive channel sensing scheme based on both energy and signal detection, that primarily considers the spectrum occupation of foreign traffic while allowing for additional complementary information such as the signal-to-noise ratio. The resulting channel quality metric is first corrected for the spectrum occupation of internal transmissions and later aggregated with help of a moving average followed by an exponential weighted moving average. This aggregation keeps the metric both sufficiently stable and adaptive to significant changes in channel usage. Moreover, the channel quality metric is made volatility-aware by penalizing qualities proportionally to their downward volatility. This yields a conservative metric and allows to differentiate channels with similar aggregated qualities but different volatility behavior.
Our second main contribution is in the form of a schedule-based channel sensing protocol, in which nodes possess two network interfaces, one for communication and one for channel sensing. Channel sensing schedules are derived from communication schedules, i.e. channel hopping sequences used for communication, with help of a stochastic local search-based heuristic that attempts to minimize channel sensing bias, the channel overlap between both schedules and to maximize overlap fairness. This minimizes the effect of internal transmissions in the resulting channel quality metric, allowing nodes to derive channel quality primarily based on foreign traffic in an unbiased manner.
Finally, we propose and implement a stabilization protocol for keeping nodes in an ad-hoc network tick-synchronized and schedule-consistent w.r.t. a communication schedule. This stabilization protocol makes use of special messages, namely tick frames for synchronization, channel quality reports for sharing local views of channel conditions and schedule reports for disseminating the global communication hopping sequence. The communication schedules are computed by a master node based on an aggregation of local channel quality views and the re-computation of these schedules is triggered by significant changes in channel conditions. The resulting protocol is robust against changes in topology and channel conditions.
Several applications have emerged and benefited from the recent advancements in wireless communication technologies. In the case of industrial automation, the wireless networks substituted wired networks to control and monitor the production systems and the factory environment. In such use cases, a common requirement is communication reliability. Technologies based on IEEE 802.15.4, such as WirelessHart and ZigBee developed for industrial applications, offer deterministic guarantees using reservation-based medium access. However, it is becoming more challenging for these technologies to guarantee their sufficiently predictable behavior, as the number of consumer electronics equipped with wireless communication technologies operating in the 2.4 GHz ISM band shared by IEEE 802.15.4 is increasing day by day.
Meanwhile, developments in WiFi technology opened the opportunity to use WiFi for industrial applications. Compared to the technologies based on IEEE 802.15.4, WiFi offers significantly higher transmission rates, and the off-the-shelf commodity WiFi hardwares are available at a low cost. However, when using a contention-based technology such as WiFi for industrial applications, additional measures are required to guarantee the specified statistical reliability.
This thesis lays the foundations for developing a multi-hop wireless control network using off-the-shelf IEEE 802.11 (WiFi) hardware operating in contention mode that can satisfy the specified reliability requirements of the applications. In a multi-hop wireless network, the communication reliability between the nodes depends on the routes determined by the routing protocol and managing these routes. We introduce a novel Quality-of-Service (QoS) routing protocol for contention-based wireless technologies such as WiFi that prioritizes reliability as the QoS requirement for route selection. The proposed routing protocol relies on different aspects of the network to determine and manage the routes. For instance, it requires algorithms and protocols to monitor and measure link quality, available bandwidth, or medium overload. Further, the determined routes require certain statistical link properties for the successful operation of the routes. We develop and evaluate different protocols, algorithms, and metrics to monitor and measure different aspects of the network in this thesis.
This dissertation describes the implementation, validation, and troubleshooting of ``Digital Twins'' in assembly processes of thin structures like parts from the automotive and aerospace industry. As requirements in terms of cost, weight, and human (pedestrian) safety are increasing for modern vehicles, thinner materials are used for exterior components. By that, components become softer but less stable which is challenging for the assembly processes and impacts the resulting quality. The most critical quality measures are gap and flushness as these are affecting aesthetics, wind noise, and fuel consumption of the final vehicle. To compensate for geometrical deviations, parts have adjustable mechanical interfaces which are used to tune in gaps and flushness for each individual assembly. For the components being assembled, individual process parameters depending on the geometry of the actual physical part must be defined. This is a challenging task that cannot be solved in a straightforward manner. However, assembly quality can be predicted by setting up individual Finite Element Method (FEM) simulation models for each part being assembled. These simulation models are called Digital Twin (DTs) as they are enriched with measured properties from the actual physical part. By that, precise predictions can be made and optimal assembly parameters for automated processes are derived. The demonstration use case in this dissertation is the assembly process of exterior car components made from sheet metals. For this kind of process, the geometrical deviations of individual components are crucial and have to be considered by the DT. To capture geometrical deviations, 3D-scanning is employed which provides a high-resolution point cloud representation of the actual physical part. This point cloud is processed further to obtain the DT that preserves the measured geometry. This dissertation tackles the following challenges: (a) setting up DTs on different level of details, (b) correctly post-processing 3D-scanned data to remove systematical measurement errors, (c) automatically morphing meshes to derive simulation models from measured point clouds, and (d) troubleshooting DTs with human-in-the-loop approaches. For all approaches, validations are provided that underline applicability and benefits. All methods and results are discussed on a high-level perspective and connections as well as the interplay between methods are elaborated. Each method either improves or extends existing approaches or provides benefits, i.e. higher precision, compared to existing solutions.
This PhD thesis is concerned with the visual analysis of time-dependent scalar field ensembles as occur in climate simulations.
Modern climate projections consist of multiple simulation runs (ensemble members) that vary in parameter settings and/or initial values, which leads to variations in the resulting simulation data.
The goal of ensemble simulations is to sample the space of possible futures under the given climate model and provide quantitative information about uncertainty in the results.
The analysis of such data is challenging because apart from the spatiotemporal data, also variability has to be analyzed and communicated.
This thesis presents novel techniques to analyze climate simulation ensembles visually.
A central question is how the data can be aggregated under minimized information loss.
To address this question, a key technique applied in several places in this work is clustering.
The first part of the thesis addresses the challenge of finding clusters in the ensemble simulation data.
Various distance metrics lend themselves for the comparison of scalar fields which are explored theoretically and practically.
A visual analytics interface allows the user to interactively explore and compare multiple parameter settings for the clustering and investigate the resulting clusters, i.e. prototypical climate phenomena.
A central contribution here is the development of design principles for analyzing variability in decadal climate simulations, which has lead to a visualization system centered around the new Clustering Timeline.
This is a variant of a Sankey diagram that utilizes clustering results to communicate climatic states over time coupled with ensemble member agreement.
It can reveal
several interesting properties of the dataset, such as:
into how many inherently similar groups the ensemble can be divided at any given time,
whether the ensemble diverges in general,
whether there are different phases in the time lapse, maybe periodicity, or outliers.
The Clustering Timeline is also used to compare multiple climate simulation models and assess their performance.
The Hierarchical Clustering Timeline is an advanced version of the above.
It introduces the concept of a cluster hierarchy that may group the whole dataset down to the individual static scalar fields into clusters of various sizes and densities recording the nesting relationship between them.
One more contribution of this work in terms of visualization research is, that ways are investigated how to practically utilize a hierarchical clustering of time-dependent scalar fields to analyze the data.
To this end, a system of different views is proposed which are linked through various interaction possibilities.
The main advantage of the system is that a dataset can now be inspected at an arbitrary level of detail without having to recompute a clustering with different parameters.
Interesting branches of the simulation can be expanded to reveal smaller differences in critical clusters or folded to show only a coarse representation of the less interesting parts of the dataset.
The last building block of the suit of visual analysis methods developed for this thesis aims at a robust, (largely) automatic detection and tracking of certain features in a scalar field ensemble.
Techniques are presented that I found can identify and track super- and sub-levelsets.
And I derive “centers of action” from these sets which mark the location of extremal climate phenomena that govern the weather (e.g. Icelandic Low and Azores High).
The thesis also presents visual and quantitative techniques to evaluate the temporal change of the positions of these centers; such a displacement would be likely to manifest in changes in weather.
In a preliminary analysis with my collaborators, we indeed observed changes in the loci of the centers of action in a simulation with increased greenhouse gas concentration as compared to pre-industrial concentration levels.
One of the biggest social issues in mature societies such as Europe and Japan
is the aging population and declining birth rate. These societies have a serious
problem with the retirement of the expert workers, doctors, and engineers etc.
Especially in the sectors that require long time to make experts in fields like medicine and industry; the retirement and injuries of the experts, is a serious problem. The technology to support the training and assessment of skilled workers (like doctors, manufacturing
workers) is strongly required for the society. Although there are some solutions for
this problem, most of them are video-based which violates the privacy of the subjects.
Furthermore, they are not easy to deploy due to the need for large training data.
This thesis provides a novel framework to recognize, analyze, and assess human
skills with minimum customization cost. The presented framework tackles this problem
in two different domains, industrial setup and medical operations of catheter-based
cardiovascular interventions (CBCVI).
In particular, the contributions of this thesis are four-fold. First, it proposes an
easy-to-deploy framework for human activity recognition based on zero-shot learning
approach, which is based on learning basic actions and objects. The model recognizes
unseen activities by combinations of basic actions learned in a preliminary way and involved objects. Therefore, it is completely configurable by the user and can be used to detect completely new activities.
Second, a novel gaze-estimation model for attention driven object detection task is
presented. The key features of the model are: (i) usage of the deformable convolutional
layers to better incorporate spatial dependencies of different shapes of objects and
backgrounds, (ii) formulation of the gaze-estimation problem in two different way, as a
classification as well as a regression problem. We combine both formulations using a
joint loss that incorporates both the cross-entropy as well as the mean-squared error in
order to train our model. This enhanced the accuracy of the model from 6.8 by using only
the cross-entropy loss to 6.4 for the joint loss.
The third contribution of this thesis targets the area of quantification of quality of
i
actions using wearable sensor. To address the variety of scenarios, we have targeted two
possibilities: a) both expert and novice data is available , b) only expert data is available,
a quite common case in safety critical scenarios.
Both of the developed methods from these scenarios are deep learning based. In the
first one, we use autoencoders with OneClass SVM, and in the second one we use the
Siamese Networks. These methods allow us to encode the expert’s expertise and to learn
the differences between novice and expert workers. This enables quantification of the
performance of the novice in comparison to the expert worker.
The fourth contribution, explicitly targets medical practitioners and provides a
methodology for novel gaze-based temporal spatial analysis of CBCVI data. The developed
methodology allows continuous registration and analysis of gaze data for analysis
of the visual X-ray image processing (XRIP) strategies of expert operators in live-cases scenarios and may assist in transferring experts’ reading skills to novices.
The development of machine learning algorithms and novel sensing modalities has boosted the exploration of human activity recognition(HAR) in recent years. In this work, we explored field-based sensing solutions and different machine learning models for HAR tasks to address the shortcomings of existing HAR sensing solutions, like the weak robustness of RF-based solution, environment-dependency of the optic-based solution, etc., aiming to supply a competitive and alternative sensing approach for HAR tasks.
Field, in physics, describes a region in which each point will be affected by force. Field sensing is potentially a low-cost, low-power, non-intrusive, privacy-respecting HAR solution that is ideal for long-term, wearable activity recording. By directly/indirectly monitoring the field strength or other field variation caused variables, some unsolved HAR problems could be addressed when other sensing solutions fail. An example is the social distance monitoring problem, where the most widely adopted approach is based on the Bluetooth signal strength measurement. However, the signal is so subtle that any object surrounding the signal emitter will cause signal attenuation. To guarantee the accuracy of social distance monitoring, we developed an induced magnetic field-based social distance monitoring system with an accuracy of a sub-ten centimetre. Moreover, the system is robust and resistant to environmental variations. Like Bluetooth, other RF-wave-based sensing modalities also face the multi-path effect caused by refraction. Thus their signal is unreliable for positioning applications where higher accuracy and robustness are needed. Besides the magnetic field, we also explored a natural static passive electric field, the field between the human body and surroundings, namely the human body capacitance(HBC). HBC is a physiological parameter describing the charge distribution difference between the body and the surroundings and is seldomly explored before. We developed a few wearable, low-cost, low power consumption hardware platforms, either based on an oscillating unit or discrete components composed sensing front end followed by a high resolution analog-to-digital module, to
monitor the variation of the parameter regarding the body movement and environmental variations. Compared with the inertial sensors, the HBC could deliver full-body movement perceiving, meaning that the movement of the legs could be perceived by a wrist-worn HBC sensing unit, which is far beyond the
sensing ability of an inertial sensing unit.
To summarize, we introduced two competitive field sensing modalities for HAR tasks, the magnetic field sensing for position-related services and the passive electric field sensing for full-body action and environmental variation sensing. Both of which were still in an infant stage and not fully explored in the community. The advantages of the two field sensing modalities were demonstrated with a series of position-related and motion-related experiments.
As the usage of concurrency in software has gained importance in the last years, and is still rising, new types of defects increasingly appeared in software. One of the most prominent and critical types of such new defect types are data races. Although research resulted in an increased effectiveness of dynamic quality assurance regarding data races, the efficiency in the quality assurance process still is a factor preventing widespread practical application. First, dynamic quality assurance techniques used for the detection of data races are inefficient. Too much effort is needed for conducting dynamic quality assurance. Second, dynamic quality assurance techniques used for the analysis of reported data races are inefficient. Too much effort is needed for analyzing reported data races and identifying issues in the source code.
The goal of this thesis is to enable efficiency improvements in the process of quality assurance for data races by: (1) analyzing the representation of the dynamic behavior of a system under test. The results are used to focus instrumentation of this system, resulting in a lower runtime overhead during test execution compared to a full instrumentation of this system. (2) Analyzing characteristics and preprocessing of reported data races. The results of the preprocessing are then provided to developers and quality assurance personnel, enabling an analysis and debugging process, which is more efficient than traditional analysis of data race reports. Besides dynamic data race detection, which is complemented by the solution, all steps in the process of dynamic quality assurance for data races are discussed in this thesis.
The solution for analyzing UML Activities for nodes possibly executing in parallel to other nodes or themselves is based on a formal foundation using graph theory. A major problem that has been solved in this thesis was the handling of cycles within UML Activities. This thesis provides a dynamic limit for the number of cycle traversals, based on the elements of each UML Activity to be analyzed and their semantics. Formal proofs are provided with regard to the creation of directed acyclic graphs and with regard to their analysis concerning the identification of elements that may be executed in parallel to other elements. Based on an examination of the characteristics of data races and data race reports, the results of dynamic data race detection are preprocessed and the outcome of this preprocessing is presented to users for further analysis.
This thesis further provides an exemplary application of the solution idea, of the results of analyzing UML Activities, and an exemplary examination of the efficiency improvement of the dynamic data race detection, which showed a reduction in the runtime overhead of 44% when using the focused instrumentation compared to full instrumentation. Finally, a controlled experiment has been set up and conducted to examine the effects of the preprocessing of reported data races on the efficiency of analyzing data race reports. The results show that the solution presented in this thesis enables efficiency improvements in the analysis of data race reports between 190% and 660% compared to using traditional approaches.
Finally, opportunities for future work are shown, which may enable a broader usage of the results of this thesis and further improvements in the efficiency of quality assurance for data races.
Towards PACE-CAD Systems
(2022)
Despite phenomenal advancements in the availability of medical image datasets and the development of modern classification algorithms, Computer-Aided Diagnosis (CAD) has had limited practical exposure in the real-world clinical workflow. This is primarily because of the inherently demanding and sensitive nature of medical diagnosis that can have far-reaching and serious repercussions in case of misdiagnosis. In this work, a paradigm called PACE (Pragmatic, Accurate, Confident, & Explainable) is presented as a set of some of must-have features for any CAD. Diagnosis of glaucoma using Retinal Fundus Images (RFIs) is taken as the primary use case for development of various methods that may enrich an ordinary CAD system with PACE. However, depending on specific requirements for different methods, other application areas in ophthalmology and dermatology have also been explored.
Pragmatic CAD systems refer to a solution that can perform reliably in day-to-day clinical setup. In this research two, of possibly many, aspects of a pragmatic CAD are addressed. Firstly, observing that the existing medical image datasets are small and not representative of images taken in the real-world, a large RFI dataset for glaucoma detection is curated and published. Secondly, realising that a salient attribute of a reliable and pragmatic CAD is its ability to perform in a range of clinically relevant scenarios, classification of 622 unique cutaneous diseases in one of the largest publicly available datasets of skin lesions is successfully performed.
Accuracy is one of the most essential metrics of any CAD system's performance. Domain knowledge relevant to three types of diseases, namely glaucoma, Diabetic Retinopathy (DR), and skin lesions, is industriously utilised in an attempt to improve the accuracy. For glaucoma, a two-stage framework for automatic Optic Disc (OD) localisation and glaucoma detection is developed, which marked new state-of-the-art for glaucoma detection and OD localisation. To identify DR, a model is proposed that combines coarse-grained classifiers with fine-grained classifiers and grades the disease in four stages with respect to severity. Lastly, different methods of modelling and incorporating metadata are also examined and their effect on a model's classification performance is studied.
Confidence in diagnosing a disease is equally important as the diagnosis itself. One of the biggest reasons hampering the successful deployment of CAD in the real-world is that medical diagnosis cannot be readily decided based on an algorithm's output. Therefore, a hybrid CNN architecture is proposed with the convolutional feature extractor trained using point estimates and a dense classifier trained using Bayesian estimates. Evaluation on 13 publicly available datasets shows the superiority of this method in terms of classification accuracy and also provides an estimate of uncertainty for every prediction.
Explainability of AI-driven algorithms has become a legal requirement after Europe’s General Data Protection Regulations came into effect. This research presents a framework for easy-to-understand textual explanations of skin lesion diagnosis. The framework is called ExAID (Explainable AI for Dermatology) and relies upon two fundamental modules. The first module uses any deep skin lesion classifier and performs detailed analysis on its latent space to map human-understandable disease-related concepts to the latent representation learnt by the deep model. The second module proposes Concept Localisation Maps, which extend Concept Activation Vectors by locating significant regions corresponding to a learned concept in the latent space of a trained image classifier.
This thesis probes many viable solutions to equip a CAD system with PACE. However, it is noted that some of these methods require specific attributes in datasets and, therefore, not all methods may be applied on a single dataset. Regardless, this work anticipates that consolidating PACE into a CAD system can not only increase the confidence of medical practitioners in such tools but also serve as a stepping stone for the further development of AI-driven technologies in healthcare.
Dealing with Dependence in the End-to-End Performance Analysis in Stochastic Network Calculus
(2022)
Communication networks, in particular the Internet, have become a pivotal part of our life. Since their beginnings, a key aspect of their applicability has been the performance. Safety-critical applications, for example, can sometimes only be implemented in a responsible manner if guarantees about their end-to-end delay can be made. A mathematical modeling and performance evaluation of communication networks requires a powerful set of tools that is able to incorporate their increasing complexity.
The stochastic network calculus (SNC) is a versatile, mathematical framework that allows for a calculation of probabilistic end-to-end performance bounds of distributed systems. Its flexibility enables to incorporate a large class of different schedulers as well as different models of traffic processes beyond the assumption of Poisson arrivals that is predominant in queueing theory-based analyses. It originates in the so-called deterministic network analysis (DNC) in the 90's of the 20th century that was introduced to provide deterministic, ``hard'' guarantees that are of relevance, e.g., in the context of real-time systems. While the DNC of today can be used to calculate fast and accurate delay bounds of arbitrary feedforward networks, the SNC is still in a significantly earlier stage. In particular, method-pertinent dependencies, i.e., a phenomenon that occurs when independent flows become stochastically dependent after sharing resources in the network, can be considered a major challenge in the SNC with moment-generating functions (MGFs).
This thesis argues to contribute to the SNC in several ways. First, we show that the ``pay multiplexing only once'' (PMOO) analysis known from the DNC is also possible in the SNC. Not only does it significantly improve end-to-end delay bounds, it also needs to consider less method-pertinent dependencies. Therefore, complexity and runtimes of the calculation are greatly reduced. Second, we introduce the concept of negative dependence to the SNC with MGFs and give numerical evidence that this can further lead to better performance bounds. Third, for the larger problem of end-to-end performance bounds of tree networks, we introduce so-called ''h-mitigators'', a modification in the calculation of MGF output bounds. It is minimally invasive, all existing results and procedures are still applicable, and improves performance bounds. As a fourth contribution, we conduct extensive numerical evaluations to substantiate our claims. Moreover, we made the respective code, the ''SNC MGF toolbox'', publicly available to ensure that the results are reproducible. At last, we conduct different stochastic analyses of a popular fair scheduler, generalized processor sharing (GPS). We give an overview of the state-of-the-art analyses in the SNC and substantiate the comparison through numerical evaluations.
To support scientific work with large and complex data the field of scientific visualization emerged in computer science and produces images through computational analysis of the data. Frameworks for combination of different analysis and visualization modules allow the user to create flexible pipelines for this purpose and set the standard for interactive scientific visualization used by domain scientists.
Existing frameworks employ a thread-parallel message-passing approach to parallel and distributed scalability, leaving the field of scientific visualization in high performance computing to specialized ad-hoc implementations. The task-parallel programming paradigm proves promising to improve scalability and portability in high performance computing implementations and thus, this thesis aims towards the creation of a framework for distributed, task-based visualization modules and pipelines.
The major contribution of the thesis is the establishment of modules for Merge Tree construction and (based on the former) topological simplification. Such modules already form a necessary first step for most visualization pipelines and can be expected to increase in importance for larger and more complex data produced and/or analysed by high performance computing.
To create a task-parallel, distributed Merge Tree construction module the construction process has to be completely revised. We derive a novel property of Merge Tree saddles and introduce a novel task-parallel, distributed Merge Tree construction method that has both good performance and scalability. This forms the basis for a module for topological simplification which we extend by introducing novel alternative simplification parameters that aim to reduce the importance of prior domain knowledge to increase flexibility in typical high performance computing scenarios.
Both modules lay the groundwork for continuative analysis and visualization steps and form a fundamental step towards an extensive task-parallel visualization pipeline framework for high performance computing.
This dissertation was developed in the context of the BMBF and EU/ECSEL funded
projects GENIAL! and Arrowhead Tools. In these projects the chair examines methods
of specifications and cooperations in the automotive value chain from OEM-Tier1-Tier2.
Goal of the projects is to improve communication and collaborative planning, especially
in early development stages. Besides SysML, the use of agreed vocabularies and on-
tologies for modeling requirements, overall context, variants, and many other items, is
targeted. This thesis proposes a web database, where data from the collaborative requirements elicitation is combined with an ontology-based approach that uses reasoning
capabilities.
For this purpose, state-of-the-art ontologies have been investigated and integrated that
entail domains like hardware/software, roadmapping, IoT, context, innovation and oth-
ers. New ontologies have been designed like a HW / SW allocation ontology and a
domain-specific "eFuse ontology" as well as some prototypes. The result is a modular
ontology suite and the GENIAL! Basic Ontology that allows us to model automotive
and microelectronic functions, components, properties and dependencies based on the
ISO26262 standard among these elements. Furthermore, context knowledge that influences design decisions such as future trends in legislation, society, environment, etc. is
included. These knowledge bases are integrated in a novel tool that allows for collabo-
rative innovation planning and requirements communication along the automotive value
chain. To start off the work of the project, an architecture and prototype tool was developed. Designing ontologies and knowing how to use them proved to be a non-trivial
task, requiring a lot of context and background knowledge. Some of this background
knowledge has been selected for presentation and was utilized either in designing models
or for later immersion. Examples are basic foundations like design guidelines for ontologies, ontology categories and a continuum of expressiveness of languages and advanced
content like multi-level theory, foundational ontologies and reasoning.
Finally, at the end, we demonstrate the overall framework, and show the ontology with
reasoning, database and APPEL/SysMD (AGILA ProPErty and Dependency Descrip-
tion Language / System MarkDown) and constraints of the hardware / software knowledge base. There, by example, we explore and solve roadmap constraints that are coupled
with a car model through a constraint solver.
Controller design for continuous dynamical systems is a core algorithmic problem in the design of cyber-physical systems (CPS). When the CPS application is safety critical, additionally we require the controller to have strong correctness guarantees. One approach for this design problem is to use simpler discrete abstraction of the original continuous system, on which known reactive synthesis methods can be used to design the controller. This approach is known as the abstraction-based controller design (ABCD) paradigm.
In this thesis, we build ABCD procedures which are faster and more modular compared to the state-of-the-art, and can handle problems which were beyond the scope of the existing techniques.
Usually, existing ABCD approaches use state space discretization for computing the abstractions, for which the procedures do not scale well for larger systems. Our first contribution is a multi-layered ABCD algorithm, where we combine coarse abstractions and lazily computed fine abstractions to improve scalability. So far, we only address reach-avoid and safety specifications, for which our prototype tool (called Mascot) showed up to an order of magnitude speedup on standard benchmark examples.
Second, we consider the problem of modular design of sound local controllers for a network of local discrete abstractions communicating via discrete/boolean variables and having local specifications. We propose a sound algorithm, where the systems negotiate a pair of local assume-guarantee contracts, in order to synchronize on a set of non-conflicting and correct behaviors. As a by-product, we also obtain a set of local controllers for the systems which ensure simultaneous satisfaction of the local specifications. We show the effectiveness of the our algorithm using a prototype tool (called Agnes) on a set of discrete benchmark examples.
Our third contribution is a novel ABCD algorithm for a more expressive model of nonlinear dynamical systems with stochastic disturbances and ω-regular specifications. This part has two subparts, which are of significant merits on their own rights. First, we present an abstraction algorithm for nonlinear stochastic systems using 2.5-player games (turn-based stochastic graph games). We show that an almost sure winning strategy in this abstract 2.5-player game gives us a sound controller for the original system for satisfying the specification with probability one. Second, we present symbolic algorithms for a seemingly different class of 2-player games with certain environmental fairness assumptions, which can also be used to efficiently compute winning strategies in the aforementioned abstract 2.5-player game. Using our prototype tool (Mascot-SDS), we show that our algorithm significantly outperforms the state-of-the-art implementation on standard benchmark examples from the literature.
Comparative Uncertainty Visualization for High-Level Analysis of Scalar- and Vector-Valued Ensembles
(2022)
With this thesis, I contribute to the research field of uncertainty visualization, considering parameter dependencies in multi valued fields and the uncertainty of automated data analysis. Like uncertainty visualization in general, both of these fields are becoming more and more important due to increasing computational power, growing importance and availability of complex models and collected data, and progress in artificial intelligence. I contribute in the following application areas:
Uncertain Topology of Scalar Field Ensembles.
The generalization of topology-based visualizations to multi valued data involves many challenges. An example is the comparative visualization of multiple contour trees, complicated by the random nature of prevalent contour tree layout algorithms. I present a novel approach for the comparative visualization of contour trees - the Fuzzy Contour Tree.
Uncertain Topological Features in Time-Dependent Scalar Fields.
Tracking features in time-dependent scalar fields is an active field of research, where most approaches rely on the comparison of consecutive time steps. I created a more holistic visualization for time-varying scalar field topology by adapting Fuzzy Contour Trees to the time-dependent setting.
Uncertain Trajectories in Vector Field Ensembles.
Visitation maps are an intuitive and well-known visualization of uncertain trajectories in vector field ensembles. For large ensembles, visitation maps are not applicable, or only with extensive time requirements. I developed Visitation Graphs, a new representation and data reduction method for vector field ensembles that can be calculated in situ and is an optimal basis for the efficient generation of visitation maps. This is accomplished by bringing forward calculation times to the pre-processing.
Visually Supported Anomaly Detection in Cyber Security.
Numerous cyber attacks and the increasing complexity of networks and their protection necessitate the application of automated data analysis in cyber security. Due to uncertainty in automated anomaly detection, the results need to be communicated to analysts to ensure appropriate reactions. I introduce a visualization system combining device readings and anomaly detection results: the Security in Process System. To further support analysts I developed an application agnostic framework that supports the integration of knowledge assistance and applied it to the Security in Process System. I present this Knowledge Rocks Framework, its application and the results of evaluations for both, the original and the knowledge assisted Security in Process System. For all presented systems, I provide implementation details, illustrations and applications.
Robotic systems are entering the stage. Enabled by advances in both hardware components and software techniques, robots are increasingly able to operate outside of factories, assist humans, and work alongside them. The limiting factor of robots’ expansion remains the programming of robotic systems. Due to the many diverse skills necessary to build a multi-robot system, only the biggest organizations are able to innovate in the space of services provided by robots.
To make developing new robotic services easier, in this dissertation I propose a program- ming model in which users (programmers) give a declarative specification of what needs to be accomplished, and then a backend system makes sure that the specification is safely and reliably executed. I present Antlab, one such backend system. Antlab accepts Linear Temporal Logic (LTL) specifications from multiple users and executes them using a set of robots of different capabilities.
Building on the experience acquired implementing Antlab, I identify problems arising from the proposed programming model. These problems fall into two broad categories, specification and planning.
In the category of specification problems, I solve the problem of inferring an LTL formula from sets of positive and negative example traces, as well as from a set of positive examples only. Building on top of these solutions, I develop a method to help users transfer their intent into a formal specification. The approach taken in this dissertation is combining the intent signals from a single demonstration and a natural language description given by a user. A set of candidate specifications is inferred by encoding the problem as a satisfiability problem for propositional logic. This set is narrowed down to a single specification through interaction with the user; the user approves or declines generated simulations of the robot’s behavior in different situations.
In the category of planning problems, I first solve the problem of planning for robots that are currently executing their tasks. In such a situation, it is unclear what to take as the initial state for planning. I solve the problem by considering multiple, speculative initial states. The paths from those states are explored based on a quality function that repeatedly estimates the planning time. The second problem is a problem of reinforcement learning when the reward function is non-Markovian. The proposed solution consists of iteratively learning an automaton representing the reward function and using it to guide the exploration.
Data-driven and Sparse-to-Dense Concepts in Scene Flow Estimation for Automotive Applications
(2022)
Highly assisted driving and autonomous vehicles require a detailed and accurate perception of the environment. This includes the perception of the 3D geometry of the scene and the 3D motion of other road users. The estimation of both based on images is known as the scene flow problem in computer vision. This thesis deals with a solution to the scene flow problem that is suitable for application in autonomous vehicles. This application imposes strict requirements on accuracy, robustness, and speed. Previous work was lagging behind in at least one of these metrics. To work towards the fulfillment of those requirements, the sparse-to-dense concept for scene flow estimation is introduced in this thesis. The idea can be summarized as follows: First, scene flow is estimated for some points of the scene for which this can be done comparatively easily and reliably. Then, an interpolation is performed to obtain a dense estimate for the entire scene. Because of the separation into two steps, each part can be optimized individually. In a series of experiments, it is shown that the proposed methods achieve competitive results and are preferable to previous techniques in some aspects. As a second contribution, individual components in the sparse-to-dense pipeline are replaced by deep learning modules. These are a highly localized and highly accurate feature descriptor to represent pixels for dense matching, and a network for robust and generic sparse-to-dense interpolation. Compared to end-to-end architectures, the advantage of deep modules is that they can be trained more effciently with data from different domains. The recombination approach applies a similar concept as the sparse-to-dense approach by solving and combining less diffcult, auxiliary sub-problems. 3D geometry and 2D motion are estimated separately, the individual results are combined, and then also interpolated into a dense scene flow. As a final contribution, the thesis proposes a set of monolithic end-to-end networks for scene flow estimation.
Today’s digital world would be unthinkable without complex data sets. Whether in private, business or industrial environments, complex data provide the basis for important and critical decisions and determine many processes, some of which are automated. This is often associated with Big Data. However, often only one aspect of the usual Big Data definitions is sufficient and a human observer can no longer capture the data completely and correctly. In this thesis, different approaches are presented in order to master selected challenges in a more effective, efficient and userfriendly way. The approaches range from easier pre-processing of data sets for later analysis and the identification of design guidelines of such assistants, new visualization techniques for presenting uncertainty, extensions of existing visualizations for categorical data, concepts for time-saving selection methods for subsets of data points and faster navigation and zoom interaction–especially in the web-based area with enormous amounts of data–to new and innovative orientation-based interaction metaphors for mobile devices as well as stationary working environments. Evaluations and appropriate use case of the individual approaches show the usability also in comparison with state-of-the-art techniques.
Industrial manufacturing companies have different IT control functions that can be represented with a so-called hierarchical automation pyramid. While these conventional software systems especially support the mass production with consistent demand, the future project “Industry 4.0” focuses on customer-oriented and adaptable production processes. In order to move from conventional production systems to a factory of the future, the control levels must be redistributed. With the help of cyber-physical production systems, an interoperable architecture must be, implemented which removes the hierarchical connection of the former control levels. The accompanied digitalisation of industrial companies makes the transition to modular production possible. At the same time, the requirements for production planning and control are increasing, which can be solved with approaches such as multi-agent systems (MASs). These software solutions are autonomous and intelligent objects with a distinct collaborative ability. There are different modelling methods, communication and interaction structures, as well as different development frameworks for these new systems. Since multi-agent systems have not yet been established as an industrial standard due to their high complexity, they are usually only tested in simulations. In this bachelor thesis, a detailed literature review on the topic of MASs in the field of production planning and control is presented. In addition, selected multi-agent approaches are evaluated and compared using specific classification criteria. In addition, the applicability of using these systems in digital and modular production is assessed.
Sequence learning describes the process of understanding the spatio-temporal
relations in a sequence in order to classify it, label its elements or generate
new sequences. Due to the prevalence of structured sequences in nature
and everyday life, it has many practical applications including any language
related processing task. One particular such task that has seen recent success
using sequence learning techniques is the optical recognition of characters
(OCR).
State-of-the-art sequence learning solutions for OCR achieve high performance
through supervised training, which requires large amounts of transcribed
training data. On the other hand, few solutions have been proposed on how
to apply sequence learning in the absence of such data, which is especially
common for hard to transcribe historical documents. Rather than solving
the unsupervised training problem, research has focused on creating efficient
methods for collecting training data through smart annotation tools or generating
synthetic training data. These solutions come with various limitations
and do not solve all of the related problems.
In this work, first the use of erroneous transcriptions for supervised sequence
learning is introduced and it is described how this concept can be applied in
unsupervised training scenarios by collecting or generating such transcriptions.
The proposed OCR pipeline reduces the need of domain specific expertise
to apply OCR, with the goal of making it more accessible. Furthermore, an
approach for evaluating sequence learning OCR models in the absence of
reference transcriptions is presented and its different properties compared
to the standard method are discussed. In a second approach, unsupervised
OCR is treated as an alignment problem between the latent features of the
different language modalities. The outlined solution is to extract language
properties from both the text and image domain through adversarial training
and learn to align them by adding a cycle consistency constraint. The proposed
approach has some strict limitations on the input data, but the results
encourage future research into more widespread applications.
Recommender systems recommend items (e.g., movies, products, books) to users. In this thesis, we proposed two comprehensive and cluster-induced recommendation-based methods: Orthogonal Inductive Matrix Completion (OMIC) and Burst-induced Multi-armed Bandit (BMAB). Given the presence of side information, the first method is categorized as context-aware. OMIC is the first matrix completion method to approach the problem of incorporating biases, side information terms and a pure low-rank term into a single flexible framework with a well-principled optimization procedure. The second method, BMAB, is context-free. That is, it does not require any side data about users or items. Unlike previous context-free multi-armed bandit approaches, our method considers the temporal dynamics of human communication on the web and treats the problem in a continuous time setting. We built our models' assumptions under solid theoretical foundations. For OMIC, we provided theoretical guarantees in the form of generalization bounds by considering the distribution-free case: no assumptions about the sampling distribution are made. Additionally, we conducted a theoretical analysis of community side information when the sampling distribution is known and an adjusted nuclear norm regularization is applied. We showed that our method requires just a few entries to accurately recover the ratings matrix if the structure of the ground truth closely matches the cluster side information. For BMAB, we provided regret guarantees under mild conditions that demonstrate how the system's stability affects the expected reward. Furthermore, we conducted extensive experiments to validate our proposed methodologies. In a controlled environment, we implemented synthetic data generation techniques capable of replicating the domains for which OMIC and BMAB were designed. As a result, we were able to analyze our algorithms' performance across a broad spectrum of ground truth regimes. Finally, we replicated a real-world scenario by utilizing well-established recommender datasets. After comparing our approaches to several baselines, we observe that they achieved state-of-the-art results in terms of accuracy. Apart from being highly accurate, these methods improve interpretability by describing and quantifying features of the datasets they characterize.
In the past, information and knowledge dissemination was relegated to the
brick-and-mortar classrooms, newspapers, radio, and television. As these
processes were simple and centralized, the models behind them were well
understood and so were the empirical methods for optimizing them. In today’s
world, the internet and social media has become a powerful tool for information
and knowledge dissemination: Wikipedia gets more than 1 million edits per day,
Stack Overflow has more than 17 million questions, 25% of US population visits
Yahoo! News for articles and discussions, Twitter has more than 60 million
active monthly users, and Duolingo has 25 million users learning languages
online. These developments have introduced a paradigm shift in the process of
dissemination. Not only has the nature of the task moved from being centralized
to decentralized, but the developments have also blurred the boundary between
the creator and the consumer of the content, i.e., information and knowledge.
These changes have made it necessary to develop new models, which are better
suited to understanding and analysing the dissemination, and to develop new
methods to optimize them.
At a broad level, we can view the participation of users in the process of
dissemination as falling in one of two settings: collaborative or competitive.
In the collaborative setting, the participants work together in crafting
knowledge online, e.g., by asking questions and contributing answers, or by
discussing news or opinion pieces. In contrast, as competitors, they vie for
the attention of their followers on social media. This thesis investigates both
these settings.
The first part of the thesis focuses on the understanding and analysis of
content being created online collaboratively. To this end, I propose models for
understanding the complexity of the content of collaborative online discussions
by looking exclusively at the signals of agreement and disagreement expressed
by the crowd. This leads to a formal notion of complexity of opinions and
online discussions. Next, I turn my attention to the participants of the crowd,
i.e., the creators and consumers themselves, and propose an intuitive model for
both, the evolution of their expertise and the value of the content they
collaboratively contribute and learn from on online Q&A based forums. The
second part of the thesis explores the competitive setting. It provides methods
to help the creators gain more attention from their followers on social media.
In particular, I consider the problem of controlling the timing of the posts of
users with the aim of maximizing the attention that their posts receive under
the idealized setting of full-knowledge of timing of posts of others. To solve
it, I develop a general reinforcement learning based method which is shown to
have good performance on the when-to-post problem and which can be employed in
many other settings as well, e.g., determining the reviewing times for spaced
repetition which lead to optimal learning. The last part of the thesis looks at
methods for relaxing the idealized assumption of full knowledge. This basic
question of determining the visibility of one’s posts on the followers’ feeds
becomes difficult to answer on the internet when constantly observing the feeds
of all the followers becomes unscalable. I explore the links of this problem to
the well-studied problem of web-crawling to update a search engine’s index and
provide algorithms with performance guarantees for feed observation policies
which minimize the error in the estimate of visibility of one’s posts.
Data is the new gold and serves as a key to answer the five W’s (Who, What, Where, When, Why) and How’s of any business. Companies are now mining data more than ever and one of the most important aspects while analyzing this data is to detect anomalous patterns to identify critical patterns and points. To tackle the vital aspects of timeseries analysis, this thesis presents a novel hybrid framework that stands on three pillars: Anomaly Detection, Uncertainty Estimation,
and Interpretability and Explainability.
The first pillar is comprised of contributions in the area of time-series anomaly detection. Deep Anomaly Detection for Time-series (DeepAnT), a novel deep learning-based anomaly detection method, lies at the foundation of the proposed hybrid framework and addresses the inadequacy of traditional anomaly detection methods. To the best of the author’s knowledge, Convolutional Neural Network (CNN) was used for the first time in Deep Anomaly Detection for Time-series (DeepAnT) to robustly detect multiple types of anomalies in the tricky
and continuously changing time-series data. To further improve the anomaly detection performance, a fusion-based method, Fusion of
Statistical and Deep Learning for Anomaly Detection (FuseAD) is proposed. This method aims to combine the strengths of existing wellfounded
statistical methods and powerful data-driven methods.
In the second pillar of this framework, a hybrid approach that combines the high accuracy of the deterministic models with the posterior distribution approximation of Bayesian neural networks is proposed.
In the third pillar of the proposed framework, mechanisms to enable both HOW and WHY parts are presented.
In order to improve performance or conserve energy, modern hardware implementations have adopted weak memory models; that is, models of concurrency that allow more outcomes than the classic sequentially consistent (SC) model of execution. Modern programming languages similarly provide their own language-level memory models, which strive to allow all the behaviors allowed by the various hardware-level memory models, as well as those that can occur as a result of desired compiler optimizations.
As these weak memory models are often rather intricate, it can be difficult for programmers to keep track of all the possible behaviors of their programs. It is therefore very useful to have an abstraction layer over the model that can be used to ensure program correctness without reasoning about the underlying memory model. Program logics are a way of constructing such an abstraction—one can use their syntactic rules to reason about programs, without needing to understand the messy details of the memory model for which the logic has been proven sound.
Unfortunately, most of the work on formal verification in general, and program logics in particular, has so far assumed the SC model of execution. This means that new logics for weak memory have to be developed.
This thesis presents two such logics—fenced separation logic (FSL) and weak separation logic (Weasel)—which are sound for reasoning under two different weak memory models.
FSL considers the C/C++ concurrency memory model, supporting several of its advanced features. The soundness of FSL depends crucially on a specific strengthening of the model which eliminates a certain class of undesired behaviors (so-called out-of-thin-air behaviors) that were inadvertently allowed by the original C/C++ model.
Weasel works under weaker assumptions than FSL, considering a model which takes a more fine-grained approach to the out-of-thin-air problem. Weasel's focus is on exploring the programming constructs directly related to out-of-thin-air behaviors, and is therefore significantly less feature-rich than FSL.
Using FSL and Weasel, the thesis explores the key challenges in reasoning under weak memory models, and what effect different solutions to the out-of-thin-air problem have on such reasoning. It explains which reasoning principles are preserved when moving from a stronger to a weaker model, and develops novel proof techniques to establish soundness of logics under weaker models.
Using Enhanced Logic Programming Semantics for Extending and Optimizing Synchronous System Design
(2021)
The semantics of programming languages assign a meaning to the written program syntax.
Currently, the meaning of synchronous programming languages, which are especially designed to develop programs for reactive and embedded systems, is based on a formal semantics definition similar to Fitting`s fixpoint semantics for logic programs.
Nevertheless, it is possible to write a synchronous program code that does not evaluate to concrete values with the current semantics, which means those programs are currently seen to be not constructive.
In the last decades, the theoretical knowledge and representation of semantics for logic programming has increased, but not all theoretical results and achievements have found their way to practice and application in system design.
This thesis, in a first part, focuses on extensions to the semantics of synchronous programming languages to an evaluation similar to a well-founded semantics as defined in logic programming by van Gelder, Ross and Schlipf and to the stable model semantics as defined by Gelfond and Lifschitz. Particularly, this allows an evaluation for some of the currently not constructive programs where the semantics based on Fitting`s fixpoint fails.
It is shown that the extension to well-founded semantics is a conservative extension of Fitting`s semantics, so that the meaning for programs which were already constructive does not change. Finally, it is considered how one can still generate circuits that implement the considered synchronous programs with the well-founded semantics. Again, this is a conservative approach that does not modify the circuits generated by the so-far used synthesis procedures.
Answer set programming and the underlying stable model semantics describe problems by constraints and the related answer set solvers give all solutions to that problem as so-called answers. This allows the formulation of searching and planning problems as well as efficient solutions without having the need to develop special and possibly error-prone algorithms for every single application.
The semantics of the synchronous programming language Quartz is also extended to the stable model semantics. For this extension, two alternatives are discussed: First of all, a direct extension similar to the extension to well-founded semantics is discussed. Second, a transformation of synchronous programs to the available answer set programming languages is given, as this allows to directly use answer set solvers for the synthesis and optimization of synchronous systems.
The second part of the thesis contains further examples of the use of answer set programming in system design to emphasis their benefits for system design in general. The first example is hereby the generation of optimal/minimal interconnection-networks which allow non-blocking connections between n sources and n targets in parallel. As a second example, the stable model semantics is used to build a complete compiler chain, which transforms a given program to an optimal assembler code (called move code) for the new SCAD processor architecture which was developed at the University of Kaiserslautern. As a final part, the lessons learned from the two examples are shown by the means of some enhancement ideas for the synchronous programming language paradigm.
Deep learning has achieved significant improvements in a variety of tasks in computer vision applications with an open image dataset which has a large amount of data. However, the acquisition of a large number of the dataset is a challenge in real-world applications, especially if they are new eras for deep learning. Furthermore, the distribution of class in the dataset is often imbalanced. The data imbalance problem is frequently bottlenecks of the neural network performance in classification. Recently, the potential of generative adversarial networks (GAN) as a data augmentation method on minority data has been studied.
This dissertation investigates using GAN and transfer learning to improve the performance of the classification under imbalanced data conditions. We first propose a classification enhancement generative adversarial networks (CEGAN) to enhance the quality of generated synthetic minority data and more importantly, to improve the prediction accuracy in data imbalanced condition. Our experiments show that approximating the real data distribution using CEGAN improves the classification performance significantly in data imbalanced conditions compared with various standard data augmentation methods.
To further improve the performance of the classification, we propose a novel supervised discriminative feature generation method (DFG) for minority class dataset. DFG is based on the modified structure of Generative Adversarial Network consisting of four independent networks: generator, discriminator, feature extractor, and classifier. To augment the selected discriminative features of minority class data by adopting attention mechanism, the generator for class-imbalanced target task is trained while feature extractor and classifier are regularized with the pre-trained ones from large source data. The experimental results show that the generator of DFG enhances the augmentation of label-preserved and diverse features, and classification results are significantly improved on the target task.
In this thesis, these proposals are deployed to bearing fault detection and diagnosis of induction motor and shipping label recognition and validation for logistics. The experimental results for bearing fault detection and diagnosis conclude that the proposed GAN-based framework has good performance on the imbalanced fault diagnosis of rotating machinery. The experimental results for shipping label recognition and validation also show that the proposed method achieves better performance than many classical and state-of-the-art algorithms.
Medical cyber-physical systems (MCPS) emerged as an evolution of the relations between connected health systems, healthcare providers, and modern medical devices. Such systems combine independent medical devices at runtime in order to render new patient monitoring/control functionalities, such as physiological closed loops for controlling drug infusion or optimization of alarms. Despite the advances regarding alarm precision, healthcare providers still struggle with alarm flooding caused by the limited risk assessment models. Furthermore, these limitations also impose severe barriers on the adoption of automated supervision through autonomous actions, such as safety interlocks for avoiding overdosage. The literature has focused on the verification of safety parameters to assure the safety of treatment at runtime and thus optimize alarms and automated actions. Such solutions have relied on the definition of actuation ranges based on thresholds for a few monitored parameters. Given the very dynamic nature of the relevant context conditions (e.g., the patient’s condition, treatment details, system configurations, etc.), fixed thresholds are a weak means for assessing the current risk. This thesis presents an approach for enabling dynamic risk assessment for cooperative MCPS based on an adaptive Bayesian Networks (BN) model. The main aim of the approach is to support continuous runtime risk assessment of the current situation based on relevant context and system information. The presented approach comprises (i) a dynamic risk analysis constituent, which corresponds to the elicitation of relevant risk parameters, risk metric building, and risk metric management; and (ii) a runtime risk classification constituent, which aims to analyze the current situation risk, establish risk classes, and identify and deploy mitigation measures. The proposed approach was evaluated and its feasibility proved by means of simulated experiments guided by an international team of medical experts with a focus on the requirements of efficacy, efficiency, and availability of patient treatment.
Dataflow process networks (DPNs) consist of statically defined process nodes with First-In-First-Out (FIFO) buffered point-to-point connections. DPNs are intrinsically data-driven, i.e., node actions are not synchronized among each other and may fire whenever sufficient input operands arrived at a node. In this original form, DPNs are data-driven and therefore a suitable model of computation (MoC) for asynchronous and distributed systems. For DPNs having nodes with only static consumption/production rates, however, one can easily derive an optimal schedule that can then be used to implement the DPN in a time-driven (clock-driven) way, where each node fires according to the schedule.
Both data-driven and time-driven MoCs have their own advantages and disadvantages. For this reason, desynchronization techniques are used to convert clock-driven models into data-driven ones in order to more efficiently support distributed implementations. These techniques preserve the functional specification of the synchronous models and moreover preserve properties like deadlock-freedom and bounded memory usage that are otherwise difficult to ensure in DPNs. These desynchronized models are the starting point of this thesis.
While the general MoC of DPNs does not impose further restrictions, many different subclasses of DPNs representing different dataflow MoCs have been considered over time like Kahn process networks, cyclo-static and synchronous DPNs. These classes mainly differ in the kinds of behaviors of the processes which affect on the one hand the expressiveness of the DPN class as well as the methods for their analysis (predictability) and synthesis (efficiency). A DPN may be heterogeneous in the sense that different processes in the network may exhibit different kinds of behaviors. A heterogeneous DPN therefore can be effectively used to model and implement different components of a system with different kinds of processes and therefore different dataflow MoCs.
Design tools for modeling like Ptolemy and FERAL are used to model and to design parallel embedded systems using well-defined and precise MoCs, including different dataflow MoCs. However, there is a lack of automatic synthesis methods to analyze and to evaluate the artifacts exhibited by particular MoCs. Second, the existing design tools for synthesis are usually restricted to the weakest classes of DPNs, i.e., cyclo-static and synchronous DPNs where each tool only supports a specific dataflow MoC.
This thesis presents a model-based design based on different dataflow MoCs including their heterogeneous combinations. This model-based design covers in particular the automatic software synthesis of systems from DPN models. The main objective is to validate, evaluate and compare the artifacts exhibited by different dataflow MoCs at the implementation level of embedded systems under the supervision of a common design tool. We are mainly concerned about how these different dataflow MoCs affect the synthesis, in particular, how they affect the code generation and the final implementation on the target hardware. Moreover, this thesis also aims at offering an efficient synthesis method that targets and exploits heterogeneity in DPNs by generating implementations based on the kinds of behaviors of the processes.
The proposed synthesis design flow therefore generally starts from the desynchronized dataflow models and automatically synthesizes them for cross-vendor target hardware. In particular, it provides a synthesis tool chain, including different specialized code generators for specific dataflow MoCs, and a runtime system that finally maps models using a combination of different dataflow MoCs on the target hardware. Moreover, the tool chain offers a platform-independent code synthesis method based on the open computing language (OpenCL) that enables a more generalized synthesis targeting cross-vendor commercial off-the-shelf (COTS) heterogeneous platforms.
DeepKAF: A Knowledge Intensive Framework for Heterogeneous Case-Based Reasoning in Textual Domains
(2021)
Business-relevant domain knowledge can be found in plain text across message exchanges
among customer support tickets, employee message exchanges and other business transactions.
Decoding text-based domain knowledge can be a very demanding task since traditional
methods focus on a comprehensive representation of the business and its relevant paths. Such
a process can be highly complex, time-costly and of high maintenance effort, especially in
environments that change dynamically.
In this thesis, a novel approach is presented for developing hybrid case-based reasoning
(CBR) systems that bring together the benefits of deep learning approaches with CBR advantages.
Deep Knowledge Acquisition Framework (DeepKAF) is a domain-independent
framework that features the usage of deep neural networks and big data technologies to decode
the domain knowledge with the minimum involvement from the domain experts. While
this thesis is focusing more on the textual data because of the availability of the datasets, the
target CBR systems based on DeepKAF are able to deal with heterogeneous data where a
case can be represented by different attribute types and automatically extract the necessary
domain knowledge while keeping the ability to provide an adequate level of explainability.
The main focus within this thesis are automatic knowledge acquisition, building similarity
measures and cases retrieval.
Throughout the progress of this research, several sets of experiments have been conducted
and validated by domain experts. Past textual data produced over around 15 years have
been used for the needs of the conducted experiments. The text produced is a mixture
between English and German texts that were used to describe specific domain problems
with a lot of abbreviations. Based on these, the necessary knowledge repositories were built
and used afterwards in order to evaluate the suggested approach towards effective monitoring
and diagnosis of business workflows. Another public dataset has been used, the CaseLaw
dataset, to validate DeepKAF when dealing with longer text and cases with more attributes.
The CaseLaw dataset represents around 22 million cases from different US states.
Further work motivated by this thesis could investigate how different deep learning models
can be used within the CBR paradigm to solve some of the chronic CBR challenges and be
of benefit to large-scale multi-dimensional enterprises.
Today, information systems are often distributed to achieve high availability and low latency.
These systems can be realized by building on a highly available database to manage the distribution of data.
However, it is well known that high availability and low latency are not compatible with strong consistency guarantees.
For application developers, the lack of strong consistency on the database layer can make it difficult to reason about their programs and ensure that applications work as intended.
We address this problem from the perspective of formal verification.
We present a specification technique, which allows specifying functional properties of the application.
In addition to data invariants, we support history properties.
These let us express relations between events, including invocations of the application API and operations on the database.
To address the verification problem, we have developed a proof technique that handles concurrency using invariants and thereby reduces the problem to sequential verification.
The underlying system semantics, technique and its soundness proof are all formalized in the interactive theorem prover Isabelle/HOL.
Additionally, we have developed a tool named Repliss which uses the proof technique to enable partially automated verification and testing of applications.
For verification, Repliss generates verification conditions via symbolic execution and then uses an SMT solver to discharge them.
This work presents a visual analytics-driven workflow for an interpretable and understandable machine learning model. The model is driven by a reverse
engineering task in automotive assembly processes. The model aims
to predict the assembly parameters leading to the given displacement field
on the geometries surface. The derived model can work on both measurement
and simulation data. The proposed approach is driven by the scientific
goals from visual analytics and interpretable artificial intelligence alike. First, a concept for systematic uncertainty monitoring, an object-oriented, virtual reference scheme (OOVRS), is developed. Afterward, the prediction task is solved via a regressive machine learning model using adversarial neural networks.
A profound model parameter study is conducted and assisted with an interactive visual analytics pipeline. Further, the effects of the learned
variance in displacement fields are analyzed in detail. Therefore a visual analytics pipeline is developed, resulting in a sensitivity benchmarking tool. This allows the testing of various segmentation approaches to lower the machine learning input dimensions. The effects of the assembly parameters are
investigated in domain space to find a suitable segmentation of the training
data set’s geometry. Therefore, a sensitivity matrix visualization is developed. Further, it is shown how this concept could directly compare results
from various segmentation methods, e.g., topological segmentation, concerning the assembly parameters and their impact on the displacement field variance. The resulting databases are still of substantial size for complex simulations with large and high-dimensional parameter spaces. Finally, the applicability of video compression techniques towards compressing visualization image databases is studied.
In the increasingly competitive public-cloud marketplace, improving the efficiency of data centers is a major concern. One way to improve efficiency is to consolidate as many VMs onto as few physical cores as possible, provided that performance expectations are not violated. However, as a prerequisite for increased VM densities, the hypervisor’s VM scheduler must allocate processor time efficiently and in a timely fashion. As we show in this thesis, contemporary VM schedulers leave substantial room for improvements in both regards when facing challenging high-VM-density workloads that frequently trigger the VM scheduler. As root causes, we identify (i) high runtime overheads and (ii) unpredictable scheduling heuristics.
To better support high VM densities, we propose Tableau, a VM scheduler that guarantees a minimum processor share and a maximum bound on scheduling delay for every VM in the system. Tableau combines a low-overhead, core-local, table-driven dispatcher with a fast on-demand table-generation procedure (triggered on VM creation/teardown) that employs scheduling techniques typically used in hard real-time systems. Further, we show that, owing to its focus on efficiency and scalability, Tableau provides comparable or better throughput than existing Xen schedulers in dedicated-core scenarios as are commonly employed in public clouds today.
Tableau also extends this design by providing the ability to use idle cycles in the system to perform low-priority background work, without affecting the performance of primary VMs, a common requirement in public clouds.
Finally, VM churn and workload variations in multi-tenant public clouds result in changing interference patterns at runtime, resulting in performance variation. In particular, variation in last-level cache (LLC) interference has been shown to have a significant impact on virtualized application performance in cloud environments. Tableau employs a novel technique for dealing with dynamically changing interference, which involves periodically regenerating tables with the same guarantees on utilization and scheduling latency for all VMs in the system, but having different LLC interference characteristics. We present two strategies to mitigate LLC interference: a randomized approach, and one that uses performance counters to detect VMs running cache-intensive workloads and selectively mitigate interference.
In recent decades, there has been increasing interest in analyzing the behavior of complex systems. A popular approach for analyzing such systems is a network analytic approach where the system is represented by a graph structure (Wassermann&Faust 1994, Boccaletti et al. 2006, Brandes&Erlebach 2005, Vespignani 2018): Nodes represent the system’s entities, edges their interactions. A large toolbox of network analytic methods, such as measures for structural properties (Newman 2010), centrality measures (Koschützki et al. 2005), or methods for identifying communities (Fortunato 2010), is readily available to be applied on any network structure. However, it is often overlooked that a network representation of a system and the (technically applicable) methods contain assumptions that need to be met; otherwise, the results are not interpretable or even misleading. The most important assumption of a network representation is the presence of indirect effects: If A has an impact on B, and B has an impact on C, then A has an impact on C (Zweig 2016, Brandes et al. 2013). The presence of indirect effects can be explained by ”something” flowing through the network by moving from node to node. Such network flows (or network processes) may be the propagation of information in social networks, the spread of infections, or entities using the network as infrastructure, such as in transportation networks. Also several network measures, particularly most centrality measures, assume the presence of such a network process, but additionally assume specific properties of the network processes (Borgatti 2005). Then, a centrality value indicates a node’s importance with respect to a process with these properties.
While this has been known for several years, only recently have datasets containing real-world network flows become accessible. In this context, the goal of this dissertation is to provide a better understanding of the actual behavior of real-world network processes, with a particular focus on centrality measures: If real-world network processes turn out to show different properties than those assumed by classic centrality measures, these measures might considerably under- or overestimate the importance of nodes for the actual network flow. To the best of our knowledge, there are only very few works addressing this topic.
The contributions of this thesis are therefore as follows: (i) We investigate in which aspects real-world network flows meet the assumptions contained about them in centrality measures. (ii) Since we find that the real-world flows show considerably different properties than assumed, we test to which extent the found properties can be explained by models, i.e., models based on shortest paths or random walks. (iii) We study whether the deviations from the assumed behavior have an impact on the results of centrality measures.
To this end, we introduce flow-based variants of centrality measures which are either based on the assumed behavior or on the actual behavior of the real-world network flow. This enables systematic evaluation of the impact of each assumption on the resulting rankings of centrality measures.
While–on a large scale–we observe a surprisingly large robustness of the measures against deviations in their assumptions, there are nodes whose importance is rated very differently when the real-world network flow is taken into account. (iv) As a technical contribution, we provide a method for an efficient handling of large sets of flow trajectories by summarizing them into groups of similar trajectories. (v) We furthermore present the results of an interdisciplinary research project in which the trajectories of humans in a network were analyzed in detail. In general, we are convinced that a process-driven perspective on network analysis in which the network process is considered in addition to the network representation, can help to better understand the behavior of complex systems.
Wearable systems have been applied in various studies as a convenient and efficient solution for
monitoring health and fitness. There is a large number of commercial products in the growing market
of wearable systems that can be worn as wristbands, clasps, or in the form of clothing. However,
these systems only provide general information about the intensity and possibly the type of
user activity, which is not sufficient for monitoring strength and conditioning exercises. To achieve
optimal muscular development and reduce the risk of exercise-related injury, a wearable system
should provide reliable biomechanical details of body movements as well as real-time feedback
during training. In addition, it should be an affordable, comfortable, and easy-to-use platform for
different types of users with different levels of movement intensity and work autonomously over
long periods of time. These requirements impose many challenges on the design of such systems.
This study presents most of these challenges and proposes solutions.
In this work, a low-cost and light-weight tracking suit is designed and developed, which integrates
multiple Inertial measurement units (IMUs). A novel data acquisition approach is proposed to
improve the energy efficiency of the system without the use of additional devices.
Given a valid calibration, IMUs, comprising inertial sensors and magnetometers, can provide accurate
orientation in three dimensions (3D). Unlike the inertial sensors, magnetometer measurements
are easily disturbed by ferromagnetic materials in the vicinity of the sensor, either inside the IMU
casing or in the final mounting position. Therefore, this work proposes a practical method for
in-field magnetometer calibration and alignment to the coordinate system of an IMU. This method
is verified experimentally in terms of magnitude deviation, heading error, plane projections, and
repeatability. The results show a higher accuracy compared to the related works.
Sensor to body calibration is a critical requirement for capturing accurate body movements.
Therefore, a theoretical analysis of an existing method is carried out, showing its limited applicability
for hip and knee joints. On this basis, by applying geometric constraints, a method is
proposed for estimating the positions of three IMUs (mounted on the pelvis, upper leg, and lower
leg) simultaneously. The result of experiments with different types of movements and arbitrary
intensity shows that the proposed method outperforms the previous method.
Moreover, two real-time tracking algorithms based on the extended Kalman filter (EKF) are proposed
for lower body motion estimation. The first approach provides an estimate of the pelvis
orientation. The second approach estimates the position of IMUs and the joint angles with respect
to the pelvis by incorporating the result of body-IMU calibration. The modeling of the biomechanical
constraint compensates for lack of a reliable horizontal reference, e.g. Earth’s magnetic field.
Experiments to track strength exercises such as squat and hip abduction/adduction show promising
results.
In order to finally provide a monitoring application in which users can follow the exercises according
to the instructions and taking into account their health status, this work proposes an approach
for the identification of exercises based on an online template matching algorithm, which detects
the correct performance using a previously recorded exercise in the presence of a supervisor.
Therefore, unlike most identification algorithms, no large datasets are required for training. The
algorithm is optimized here to reduce execution time while maintaining the accuracy. Experiments
show that for the specific application of this study, i.e. squat exercise monitoring, the proposed
method outperforms the related works in optimization of online template matching.
Visual analytics has been widely studied in the past decade both in academia and industry to improve data exploration, minimize the overall cost, and improve data analysis. In this chapter, we explore the idea of visual analytics in the context of simulation data. This would then provide us with the capability to not only explore our data visually but also to apply machine learning models in order to answer high-level questions with respect to scheduling, choosing optimal simulation parameters, finding correlations, etc. More specifically, we examine state-of-the-art tools to be able to perform these above-mentioned tasks. Further, to test and validate our methodology we followed the human-centered design process to build a prototype tool called ViDAS (Visual Data Analytics of Simulated Data). Our preliminary evaluation study illustrates the intuitiveness and ease-of-use of our approach with regards to visual analysis of simulated data.
To improve efficiency of memory accesses, modern multiprocessor architectures implement a whole range of different weak memory models. The behavior of performance-critical code depends on the underlying hardware. There is a rising demand for verification tools that take the underlying memory model into account. This work examines a variety of prevalent problems in the field of program verification of increasing complexities: testing, reachability, portability and memory model synthesis.
We give efficient tools to solve these problems. What sets the presented methods apart is that they are not limited to some few given architectures. They are universal: The memory model is given as part of the input. We make use of the CAT language to succinctly describe axiomatic memory models. CAT has been used to define the semantics of assembly for x86/TSO, ARMv7, ARMv8, and POWER but also the semantics of programming languages such as C/C++, including the Linux kernel concurrency primitives.
This work shows that even the simple testing problem is NP-hard for most memory models. It does so using a general reduction technique that applies to a range of models. It examines the more difficult program verification under a memory model and introduces Dartagnan, a bounded model checker (BMC) that encodes the problem as an SMT-query and makes use of advanced encoding techniques. The program portability problem is shown to be even harder. Despite this, it is solved efficiently by the tool Porthos which uses a guided search to produce fast results for most practical instances. A memory model is synthesized by Aramis for a given set of reachability results. Concurrent program verification is generally undecidable even for sequential consistency. As an alternative to BMC, we propose a new CEGAR method for Petri net invariant synthesis. We again use SMT-queries as a back-end.
Drahtlose Kommunikationssysteme dringen in immer mehr Anwendungsbereiche vor. Für einige Anwendungsszenarien, wie etwa die Prozessautomatisierung in Fabriken und Industrieanlagen, ist die Zuverlässigkeit vieler drahtloser Kommunikationssysteme wie IEEE 802.11 (WLAN) oder Bluetooth aber noch unzureichend. Daher wurden für diese Anwendungsbereiche spezielle Kommunikationssysteme wie WirelessHART oder ISA 100.11a entwickelt. Diese basieren meist auf Time Division Multiple Access (TDMA) und erreichen durch exklusive Reservierungen deterministische Zuverlässigkeit, falls kein anderes Kommunikationssystem die genutzten Kanäle stört.
Diese Arbeit behandelt geeignete Protokolle und Algorithmen, um die Zuverlässigkeit drahtloser Kommunikationssysteme zu verbessern. Im ersten Teil der Arbeit werden Verfahren für TDMA-basierte Kommunikationssysteme betrachtet. Basierend auf IEEE 802.15.4 werden mehrere Funktionalitäten für ProNet 4.0, einem an der Arbeitsgruppe Vernetzte Systeme der TU Kaiserslautern entwickelten Kommunikations-Stack für Industrie 4.0, entworfen und auf Imote 2 Sensorknoten implementiert. Zuverlässige Kommunikation bedarf Kenntnis von sowohl der Kommunikationstopologie, über die Knoten miteinander kommunizieren können, als auch der Interferenztopologie, die angibt, wie Knoten sich gegenseitig stören können. Dazu stellt die Arbeit mit dem Automatic Topology Discovery Protocol (ATDP) ein Verfahren zur automatischen Topologieerkennung vor. Anschließend wird QoS Multicast Routing betrachtet und mit dem QoS Multicast Routing Protocol (QMRP) ein Verfahren für partiell mobile Netzwerke entwickelt. Weiterhin wird mit ProMid eine Kommunikations-Middleware beschrieben, die ein hohes Abstraktionslevel aufweist und die darunter liegenden Schichten steuert. Die dienstorientierte Architektur nutzt eine verteilte Service Registry, wobei die Auswahl der Registry-Knoten anhand eines dafür entwickelten Clustering-Algorithmus erfolgt. Das Heterogeneous Network Clustering (HNC) genannte Verfahren berücksichtigt ein heterogenes Netzwerkmodell mit Knoten, die Clusterhead bzw. Gateway werden müssen, können bzw. nicht dürfen.
Der zweite Teil der Arbeit behandelt Protokolle und Algorithmen für zuverlässige wettbewerbsbasierte Kommunikationssysteme. Die in diesem Kapitel vorgestellten Verfahren sind in einem auf WLAN basierenden Kommunikations-Stack implementiert und evaluiert worden. Zunächst wird ein Verfahren für die Topologieerkennung in WLAN-Netzwerken vorgestellt. Anschließend wird ein auf dem Token Bucket-Mechanismus basierendes Verfahren zur Verkehrskontrolle entwickelt. Daraus wird mit der Unusable Wasted Bandwidth Ratio eine Metrik abgeleitet, die es erlaubt, die Auslastung des Mediums abzuschätzen. Aufbauend auf dem Verfahren zur Verkehrskontrolle wird eine kooperative faire Bandbreitenskalierung für WLAN vorgestellt. Das Verfahren verteilt die Bandbreite fair unter den internen Knoten unter Berücksichtigung der Quality of Service (QoS) Anforderungen. Dabei reagiert es dynamisch auf Änderungen des externen Verkehrs und verhindert so Überlastsituationen. Letztlich wird ein Clustering-Protokoll vorgestellt, welches durch das Anwendungsszenario der Überwachung von Güterzügen motiviert ist und Linientopologien bildet sowie dynamisch repariert. Das auf Bluetooth LE aufbauende Verfahren dient dazu, Energie einzusparen, und wurde in einer Kooperation mit der Bosch Engineering GmbH entwickelt.
Optical Character Recognition (OCR) is one of the central problems in pattern recognition. Its
applications have played a great role in the digitization of document images collected from het-
erogeneous sources. Many of the well-known scripts have OCR systems with sufficiently high
performance that enables OCR applications in industrial/commercial settings. However, OCR sys-
tems yield very-good results only on a narrow domain and very-specific use cases. Thus, it is still
a challenging task, and there are other exotic languages with indigenous scripts, such as Amharic,
for which no well-developed OCR systems exist.
As many as 100 million people speak Amharic, and it is an official working language of Ethiopia.
Amharic script contains about 317 different alphabets derived from 34 consonants with small changes.
The change involves shortening or elongating one of its main legs or adding small diacritics to the
right, left, top, or bottom of the consonant character. Such modifications lead the characters to have
similar shapes and make the recognition task complex, but this is particularly interesting for charac-
ter recognition research. So far, Amharic script recognition models are developed based on classical
machine learning techniques, and they are very limited in addressing the issues for Amharic OCR.
The motivation of this thesis is, therefore, to explore and tailor contemporary deep learning tech-
niques for the OCR of Amharic.
This thesis addresses the challenges in Amharic OCR through two main contributions. The first
contribution is an algorithmic contribution in which we investigate deep learning approaches that
suit the demand for Amharic OCR. The second is a technical contribution that comprises several
works towards the OCR model development; thus, it introduces a new Amharic database consisting
of collections of images annotated at a character and text-line level. It also presents a novel CNN-
based framework designed by leveraging the grapheme of characters in Fidel-Gebeta (where Fidel-
Gebeta consists of the full set of Amharic characters in matrix structure) and achieves 94.97%
overall character recognition accuracy.
In addition to character level methods, text-line level methods are also investigated and devel-
oped based on sequence-to-sequence learning. These models avoid several of the pre-processing
stages used in prior works by eliminating the need to segment individual characters. In this design,
we use a stack of CNNs, before the Bi-LSTM layers and train from end-to-end. This model out-
performs the LSTM-CTC based network, on average, by a CER of 3.75% with the ADOCR test
set. Motivated by the success of attention, in addressing the problems’ of long sequences in Neural
Machine Translation (NMT), we proposed a novel attention-based methodology by blending the
attention mechanism into CTC objective function. This model performs far better than the existing
techniques with a CER of 1.04% and 0.93% on printed and synthetic text-line images respectively.
Finally, this thesis provides details on various tasks that have been performed for the development
of Amharic OCR. As per our empirical analysis, the majority of the errors are due to poor annotation
of the dataset. As future work, the methods proposed in this thesis should be further investigated
and extended to deal with handwritten and historical Amharic documents.
Multidisciplinary optimizations (MDOs) encompass optimization problems that combine different disciplines into a single optimization with the aim of converging towards a design that simultaneously fulfills multiple criteria. For example, considering both fluid and structural disciplines to obtain a shape that is not only aerodynamically efficient, but also respects structural constraints. Combined with CAD-based parametrizations, the optimization produces an improved, manufacturable shape. For turbomachinery applications, this method has been successfully applied using gradient-free optimization methods such as genetic algorithms, surrogate modeling, and others. While such algorithms can be easily applied without access to the source code, the number of iterations to converge is dependent on the number of design parameters. This results in high computational costs and limited design spaces. A competitive alternative is offered by gradient-based optimization algorithms combined with adjoint methods, where the computational complexity of the gradient calculation is no longer dependent on the number of design parameters, but rather on the number of outputs. Such methods have been extensively used in single-disciplinary aerodynamic optimizations using adjoint fluid solvers and CAD parametrizations. However, CAD-based MDOs leveraging adjoint methods are just beginning to emerge.
This thesis contributes to this field of research by setting up a CAD-based adjoint MDO framework for turbomachinery design using both fluid and structural disciplines. To achieve this, the von Kármán Institute’s existing CAD-based optimization framework cado is augmented by the development of a FEM-based structural solver which has been differentiated using the algorithmic differentiation tool CoDiPack of TU Kaiserslautern. While most of the code could be differentiated in a black-box fashion, special treatment is required for the iterative linear and eigenvalue solvers to ensure accuracy and reduce memory consumption. As a result, the solver has the capability of computing both stress and vibration gradients at a cost independent on the number of design parameters. For the presented application case of a radial turbine optimization, the full gradient calculation has a computational cost of approximately 3.14 times the cost of a primal run and the peak memory usage of approximately 2.76 times that of a primal run.
The FEM code leverages object-oriented design such that the same base structure can be reused for different purposes with minimal re-differentiation. This is demonstrated by considering a composite material test case where the gradients could be easily calculated with respect to an extended design space that includes material properties. Additionally, the structural solver is reused within a CAD-based mesh deformation, which propagates the structural FEM mesh gradients through to the CAD parameters. This closes the link between the CAD shape and FEM mesh. Finally, the MDO framework is applied by optimizing the aerodynamic efficiency of a radial turbine while respecting structural constraints.
In recent years, business intelligence applications become more real-time and traditional data warehouse tables become fresher as they are continuously refreshed by streaming ETL jobs within seconds. Besides, a new type of federated system emerged that unifies domain-specific computation engines to address a wide range of complex analytical applications, which needs streaming ETL to migrate data across computation systems.
From daily-sales reports to up-to-the-second cross-/up-sell campaign activities, we observed various latency and freshness requirements set in these analytical applications. Hence, streaming ETL jobs with regular batches are not flexible enough to fit in such a mixed workload. Jobs with small batches can cause resource overprovision for queries with low freshness needs while jobs with large batches would starve queries with high freshness needs. Therefore, we argue that ETL jobs should be self-adaptive to varying SLA demands by setting appropriate batches as needed.
The major contributions are summarized as follows.
• We defined a consistency model for “On-Demand ETL” which addresses correct batches for queries to see consistent states. Furthermore, we proposed an “Incremental ETL Pipeline” which reduces the performance impact of on-demand ETL processing.
• A distributed, incremental ETL pipeline (called HBelt) was introduced in distributed warehouse systems. HBelt aims at providing consistent, distributed snapshot maintenance for concurrent table scans across different analytics jobs.
• We addressed the elasticity property for incremental ETL pipeline to guarantee that ETL jobs with batches of varying sizes can be finished within strict deadlines. Hence, we proposed Elastic Queue Middleware and HBaqueue which replace memory-based data exchange queues with a scalable distributed store - HBase.
• We also implemented lazy maintenance logic in the extraction and the loading phases to make these two phases workload-aware. Besides, we discuss how our “On-Demand ETL” thinking can be exploited in analytic flows running on heterogeneous execution engines.
Scaling up conventional processor architectures cannot translate the ever-increasing number of transistors into comparable application performance. Although the trend is to shift from single-core to multi-core architectures, utilizing these multiple cores is not a trivial task for many applications due to thread synchronization and weak memory consistency issues. This is especially true for applications in real-time embedded systems since timing analysis becomes more complicated due to contention on shared resources. One inherent reason for the limited use of instruction-level parallelism (ILP) by conventional processors is the use of registers. Therefore, some recent processors bypass register usage by directly communicating values from producer processing units to consumer processing units. In widely used superscalar processors, this direct instruction communication is organized by hardware at runtime, adversely affecting its scalability. The exposed datapath architectures provide a scalable alternative by allowing compilers to move values directly from output ports to the input ports of processing units. Though exposed datapath architectures have already been studied in great detail, they still use registers for executing programs, thus limiting the amount of ILP they can exploit. This limitation stems from a drawback in their execution paradigm, code generator, or both.
This thesis considers a novel exposed datapath architecture named Synchronous Control Asynchronous Dataflow (SCAD) that follows a hybrid control-flow dataflow execution paradigm. The SCAD architecture employs first-in-first-out (FIFO) buffers at the output and input ports of processing units. It is programmed by move instructions that transport values from the head of output buffers to the tail of input buffers. Thus, direct instruction communication is facilitated by the architecture. The processing unit triggers the execution of an operation when operand values are available at the heads of its input buffers. We propose a code generation technique for SCAD processors inspired by classical queue machines that completely eliminates the use of registers. On this basis, we first generate optimal code by using satisfiability (SAT) solvers after establishing that optimal code generation is hard. Heuristics based on a novel buffer interference analysis are then developed to compile larger programs. The experimental results demonstrate the efficacy of the execution paradigm of SCAD using our queue-oriented code generation technique.
Kinetic models of human motion rely on boundary conditions which are defined by the interaction of the body with its environment. In the simplest case, this interaction is limited to the foot contact with the ground and is given by the so called ground reaction force (GRF). A major challenge in the reconstruction of GRF from kinematic data is the double support phase, referring to the state with multiple ground contacts. In this case, the GRF prediction is not well defined. In this work we present an approach to reconstruct and distribute vertical GRF (vGRF) to each foot separately, using only kinematic data. We propose the biomechanically inspired force shadow method (FSM) to obtain a unique solution for any contact phase, including double support, of an arbitrary motion. We create a kinematic based function, model an anatomical foot shape and mimic the effect of hip muscle activations. We compare our estimations with the measurements of a Zebris pressure plate and obtain correlations of 0.39≤r≤0.94 for double support motions and 0.83≤r≤0.87 for a walking motion. The presented data is based on inertial human motion capture, showing the applicability for scenarios outside the laboratory. The proposed approach has low computational complexity and allows for online vGRF estimation.
In today’s computer networks we see an ongoing trend towards wireless communication technologies, such as Wireless LAN, Bluetooth, ZigBee and cellular networks. As the electromagnetic spectrum usable for wireless communication is finite and largely allocated for exclusive use by respective license holders, there are only few frequency bands left for general, i.e. unlicensed, use. Subsequently, it becomes apparent, that there will be an overload situation in the unlicensed bands, up to a point where no communication is possible anymore. On the other hand, it has been observed that licensed frequency bands often go unused, at least at some places or over time. Mitola combined both observations and found the term Cognitive Radio Networks [Mit00], denoting a solution for spectrum scarcity. In this concept, so called Secondary Users are allowed to also use licensed bands (attributed to a Primary User) as long as it is vacant.
In such networks, all obligations reside with Secondary Users, especially, they must avoid any interference with the Primary User. They must therefore reliably sense the presence of Primary Users and must decide which available spectrum to use. These two functionalities are called Spectrum Sensing and Spectrum Mobility and describe 2 out of 4 core functionalities of Cognitive Radio Networks and are considered in this thesis.
Regarding Spectrum Sensing, we present our own approach for energy detection in this thesis. Energy detection essentially works by comparing measured energy levels to a threshold. The inherent problem is on how to find such thresholds. Based on existing work we found in literature, we improve techniques and assert the effectiveness of our additions by conducting real world experiments.
Regarding Spectrum Mobility, we concentrate on the point, where the Primary User shows up. At this point, nodes must not use the current channel anymore, i.e. they also have no possibility to agree on another channel to switch to. We solve this problem by employing channel switching, i.e. we change channels proactively, following a schedule shared by all nodes of the network. The main contribution of this thesis is on how to synthesize those schedules to guarantee robust operation under changing conditions. For integration, we considered three dimensions of robustness (of time, of space and of channel) and, based on our algorithms and findings, defined a network protocol, which addresses perturbation within those dimensions. In an evaluation, we showed that the protocol is actually able to maintain robust operation, even if there are large drops in channel quality.
3D hand pose and shape estimation from a single depth image is a challenging computer vision and graphics problem with many applications such as
human computer interaction and animation of a personalized hand shape in
augmented reality (AR). This problem is challenging due to several factors
for instance high degrees of freedom, view-point variations and varying hand
shapes. Hybrid approaches based on deep learning followed by model fitting
preserve the structure of hand. However, a pre-calibrated hand model limits
the generalization of these approaches. To address this limitation, we proposed a novel hybrid algorithm for simultaneous estimation of 3D hand pose
and bone-lengths of a hand model which allows training on datasets that contain varying hand shapes. On the other hand, direct joint regression methods
achieve high accuracy but they do not incorporate the structure of hand in
the learning process. Therefore, we introduced a novel structure-aware algorithm which learns to estimate 3D hand pose jointly with new structural constraints. These constraints include fingers lengths, distances of joints along
the kinematic chain and fingers inter-distances. Learning these constraints
help to maintain a structural relation between the estimated joint keypoints.
Previous methods addressed the problem of 3D hand pose estimation. We
open a new research topic and proposed the first deep network which jointly
estimates 3D hand shape and pose from a single depth image. Manually annotating real data for shape is laborious and sub-optimal. Hence, we created a
million-scale synthetic dataset with accurate joint annotations and mesh files
of depth maps. However, the performance of this deep network is restricted by
limited representation capacity of the hand model. Therefore, we proposed a
novel regression-based approach in which the 3D dense hand mesh is recovered
from sparse 3D hand pose, and weak-supervision is provided by a depth image synthesizer. The above mentioned approaches regressed 3D hand meshes
from 2D depth images via 2D convolutional neural networks, which leads to
artefacts in the estimations due to perspective distortions in the images. To
overcome this limitation, we proposed a novel voxel-based deep network with
3D convolutions trained in a weakly-supervised manner. Finally, an interesting
application is presented which is in-air signature acquisition and verification
based on deep hand pose estimation. Experiments showed that depth itself is
an important feature, which is sufficient for verification.
In the avionics domain, “ultra-reliability” refers to the practice of ensuring quantifiably negligible residual failure rates in the presence of transient and permanent hardware faults. If autonomous Cyber- Physical Systems (CPS) in other domains, e.g., autonomous vehicles, drones, and industrial automation systems, are to permeate our everyday life in the not so distant future, then they also need to become ultra-reliable. However, the rigorous reliability engineering and analysis practices used in the avionics domain are expensive and time consuming, and cannot be transferred to most other CPS domains. The increasing adoption of faster and cheaper, but less reliable, Commercial Off-The-Shelf (COTS) hardware is also an impediment in this regard.
Motivated by the goal of ultra-reliable CPS, this dissertation shows how to soundly analyze the reliability of COTS-based implementations of actively replicated Networked Control Systems (NCSs)—which are key building blocks of modern CPS—in the presence of transient hard- ware faults. When an NCS is deployed over field buses such as the Controller Area Network (CAN), transient faults are known to cause host crashes, network retransmissions, and incorrect computations. In addition, when an NCS is deployed over point-to-point networks such as Ethernet, even Byzantine errors (i.e., inconsistent broadcast transmissions) are possible. The analyses proposed in this dissertation account for NCS failures due to each of these error categories, and consider NCS failures in both time and value domains. The analyses are also provably free of reliability anomalies. Such anomalies are problematic because they can result in unsound failure rate estimates, which might lead us to believe that a system is safer than it actually is.
Specifically, this dissertation makes four main contributions. (1) To reduce the failure rate of NCSs in the presence of Byzantine errors, we present a hard real-time design of a Byzantine Fault Tolerance (BFT) protocol for Ethernet-based systems. (2) We then propose a quantitative reliability analysis of the presented design in the presence of transient faults. (3) Next, we propose a similar analysis to upper-bound the failure probability of an actively replicated CAN-based NCS. (4) Finally, to upper-bound the long-term failure rate of the NCS more accurately, we propose analyses that take into account the temporal robustness properties of an NCS expressed as weakly-hard constraints.
By design, our analyses can be applied in the context of full-system analyses. For instance, to certify a system consisting of multiple actively replicated NCSs deployed over a BFT atomic broadcast layer, the upper bounds on the failure rates of each NCS and the atomic broadcast layer can be composed using the sum-of-failure-rates model.
Learning From Networked-data: Methods and Models for Understanding Online Social Networks Dynamics
(2020)
Abstract
Nowadays, people and systems created by people are generating an unprecedented amount of
data. This data has brought us data-driven services with a variety of applications that affect
people’s behavior. One of these applications is the emergent online social networks as a method
for communicating with each other, getting and sharing information, looking for jobs, and many
other things. However, the tremendous growth of these online social networks has also led to many
new challenges that need to be addressed. In this context, the goal of this thesis is to better understand
the dynamics between the members of online social networks from two perspectives. The
first perspective is to better understand the process and the motives underlying link formation in
online social networks. We utilize external information to predict whether two members of an online
social network are friends or not. Also, we contribute a framework for assessing the strength of
friendship ties. The second perspective is to better understand the decay dynamics of online social
networks resulting from the inactivity of their members. Hence, we contribute a model, methods,
and frameworks for understanding the decay mechanics among the members, for predicting members’
inactivity, and for understanding and analyzing inactivity cascades occurring during the decay.
The results of this thesis are: (1) The link formation process is at least partly driven by interactions
among members that take place outside the social network itself; (2) external interactions might
help reduce the noise in social networks and for ranking the strength of the ties in these networks;
(3) inactivity dynamics can be modeled, predicted, and controlled using the models contributed in
this thesis, which are based on network measures. The contributions and the results of this thesis
can be beneficial in many respects. For example, improving the quality of a social network by introducing
new meaningful links and removing noisy ones help to improve the quality of the services
provided by the social network, which, e.g., enables better friend recommendations and helps to
eliminate fake accounts. Moreover, understanding the decay processes involved in the interaction
among the members of a social network can help to prolong the engagement of these members. This
is useful in designing more resilient social networks and can assist in finding influential members
whose inactivity may trigger an inactivity cascade resulting in a potential decay of a network.
Background: The use of health apps to support the treatment of chronic pain is gaining importance. Most available pain management apps are still lacking in content quality and quantity as their developers neither involve health experts to ensure target group suitability nor use gamification to engage and motivate the user. To close this gap, we aimed to develop a gamified pain management app, Pain-Mentor.
Objective: To determine whether medical professionals would approve of Pain-Mentor’s concept and content, this study aimed to evaluate the quality of the app’s first prototype with experts from the field of chronic pain management and to discover necessary improvements.
Methods: A total of 11 health professionals with a background in chronic pain treatment and 2 mobile health experts participated in this study. Each expert first received a detailed presentation of the app. Afterward, they tested Pain-Mentor and then rated its quality using the mobile application rating scale (MARS) in a semistructured interview.
Results: The experts found the app to be of excellent general (mean 4.54, SD 0.55) and subjective quality (mean 4.57, SD 0.43). The app-specific section was rated as good (mean 4.38, SD 0.75). Overall, the experts approved of the app’s content, namely, pain and stress management techniques, behavior change techniques, and gamification. They believed that the use of gamification in Pain-Mentor positively influences the patients’ motivation and engagement and thus has the potential to promote the learning of pain management techniques. Moreover, applying the MARS in a semistructured interview provided in-depth insight into the ratings and concrete suggestions for improvement.
Conclusions: The experts rated Pain-Mentor to be of excellent quality. It can be concluded that experts perceived the use of gamification in this pain management app in a positive manner. This showed that combining pain management with gamification did not negatively affect the app’s integrity. This study was therefore a promising first step in the development of Pain-Mentor.
To assess ergonomic aspects of a (future) workplace already in the design phase where no physical prototypes exist, the use of digital human models (DHMs) becomes essential. Thereby, the prediction of human motions is a key aspect when simulating human work tasks. For ergonomic assessment e.g. the resulting postures, joint angles, the duration of the motion and muscle loads are important quantities. From a physical point of view, there is an infinite number of possible ways for a human to fulfill a given goal (trajectories, velocities...), which makes human motions and behavior hard to predict. A common approach used in state of the art commercial DHMs is the manual definition of joint angles by the user, which requires expert knowledge and is limited to postural assessments. Another way is to make use of pre-recorded motions from a real human that operates on a physical prototype, which limits assessments to scenarios which have been measured before. Both approaches need further post processing and inverse dynamics calculations with other software tools to get information about inner loads and muscle data, which leads to further uncertainties concerning validity of the simulated data.
In this thesis work a DHM control and validation framework is developed, which allows to investigate in how far the implemented human like actuation and control principles directly lead to human like motions and muscle actuations. From experiments performed in the motion laboratory, motion data is captured and muscle activations are measured using surface electromyography measurements (EMG). From the EMG data, time invariant muscle synergies are extracted by the use of a non-negative Matrix Factorization algorithm (NMF). Muscle synergies are one hypothesis from neuroscience to explain how the human central nervous system might reduce control complexity: instead of activating each muscle separately, muscles are grouped into functional units, whereas each muscle is present in each unit with a fixed amplitude. The measured experiment is then simulated in an optimal control framework. The used framework allows to build up DHMs as multibody system (MBS): bones are modeled as rigid bodies connected via joints, actuated by joint torques or by Hill type muscle models (1D string elements transferring fundamental characteristics of muscle force generation in humans). The OC code calculates the actuation signals for the modeled DHM in a way that a certain goal is fulfilled (e.g. reach for an object) while minimizing some cost function (e.g. minimizing time) and considering the side constraints that the equations of motion of the MBS are fulfilled. Therefore, three different Actuation Modes (AM) can be used (joint torques (AM-T), direct muscle actuation (AM-M) and muscle synergy actuation (AM-S), using the before extracted synergies as control parameters)). Simulation results are then compared with measured data, to investigate the influence of the different Actuation Modes and the solved OC cost function. The approach is applied to three different experiments, the basic reaching test, the weight lift test and a box lifting task, where a human arm model actuated by 29 Hill muscles is used for simulation. It is shown that, in contrast to a joint torque actuation (AM-T), using muscles as actuators (AM-M & AM-S) leads to very human like motion trajectories. Muscle synergies as control parameters, resulted in smoother velocity profiles, which were closer to those measured and appeared to be more robust, concerning the underlying muscle activation signals (compared to AM-M). In combination with a developed biomechanical cost function (a mix of different OC cost functions), the approach showed promising results, concerning the simulation of valid, human like motions, in a predictive manner.
In recent years, the Internet has become a major source of visual information exchange. Popular social platforms have reported an average of 80 million photo uploads a day. These images, are often accompanied with a user provided text one-liner, called an image caption. Deep Learning techniques have made significant advances towards automatic generation of factual image captions. However, captions generated by humans are much more than mere factual image descriptions. This work takes a step towards enhancing a machine's ability to generate image captions with human-like properties. We name this field as Affective Image Captioning, to differentiate it from the other areas of research focused on generating factual descriptions.
To deepen our understanding of human generated captions, we first perform a large-scale Crowd-Sourcing study on a subset of Yahoo Flickr Creative Commons 100 Million Dataset (YFCC100M). Three thousand random image-caption pairs were evaluated by native English speakers w.r.t different dimensions like focus, intent, emotion, meaning, and visibility. Our findings indicate three important underlying properties of human captions: subjectivity, sentiment, and variability. Based on these results, we develop Deep Learning models to address each of these dimensions.
To address the subjectivity dimension, we propose the Focus-Aspect-Value (FAV) model (along with a new task of aspect-detection) to structure the process of capturing subjectivity. We also introduce a novel dataset, aspects-DB, following this way of modeling. To implement the model, we propose a novel architecture called Tensor Fusion. Our experiments show that Tensor Fusion outperforms the state-of-the-art cross residual networks (XResNet) in aspect-detection.
Towards the sentiment dimension, we propose two models:Concept & Syntax Transition Network (CAST) and Show & Tell with Emotions (STEM). The CAST model uses a graphical structure to generate sentiment. The STEM model uses a neural network to inject adjectives into a neutral caption. Achieving a high score of 93% with human evaluation, these models were selected as the top-3 at the ACMMM Grand Challenge 2016.
To address the last dimension, variability, we take a generative approach called Generative Adversarial Networks (GAN) along with multimodal fusion. Our modified GAN, with two discriminators, is trained using Reinforcement Learning. We also show that it is possible to control the properties of the generated caption-variations with an external signal. Using sentiment as the external signal, we show that we can easily outperform state-of-the-art sentiment caption models.
B-spline surfaces are a well-established tool to analytically describe objects. They are commonly used in various fields, e.g., mechanical and aerospace engineering, computer aided design, and computer graphics. Obtaining and using B-spline surface models of real-life objects is an intricate process. Initial virtual representations are usually obtained via scanning technologies in the form of discrete data, e.g., point clouds, surface meshes, or volume images. This data often requires pre-processing to remove noise and artifacts. Even with high quality data, obtaining models of complex or very large structures needs specialized solutions that are viable for the available hardware. Once B-spline models are constructed, their properties can be utilized and combined with application-specific knowledge to provide efficient solutions for practical problems.
This thesis contributes to various aspects of the processing pipeline. It addresses pre-processing, creating B-spline models of large and topologically challenging data, and the use of such models within the context of visual surface inspection. Proposed methods improve existing solutions in terms of efficiency, hardware restrictions, and quality of the results. The following contributions are presented:
Fast and memory-efficient quantile filter: Quantile filters are widely used operations in image processing. The most common instance is the median filter which is a standard solution to treat noise while preserving the shape of an object. Various implementations of such filters are available offering either high performance or low memory complexity, but not both. This thesis proposes a generalization of two existing algorithms: one that favors speed and one that favors low memory usage. An adaptable hybrid algorithm is introduced. It can be tuned for optimal performance on the available hardware. Results show that it outperforms both state-of-the-art reference methods for most practical filter sizes.
Robust B-spline reconstructions of isosurfaces in volume images: The micro-structure of wood-based thermal insulation materials is analyzed to research heat conductivity properties. High-quality scans reveal a complex system of cellulose fibers. B-spline models of individual fibers are highly desirable to conduct simulations. Due to the physical processing of the material, the surfaces of those fibers consist of challenging elements like loose filaments, holes, and tunnels. Standard solutions fail to partition the data into a small number of quadrilateral cells which is required for the B-spline construction step. A novel approach is presented that splits up the data processing into separate topology and geometry pipelines. This robust method is demonstrated by constructing B-spline models with 236 to 676 surfaces from triangulated isosurfaces with 423628 to 1203844 triangles.
Local method for smooth B-spline surface approximations: Constructing smooth B-spline models to approximate discrete data is a challenging task. Various standard solutions exist, often imposing restrictions to knot vectors, spline order, or available degrees of freedom for the data approximation. This thesis presents a local approach with less restrictions aiming for approximate \(G^1\)-continuity. Nonlinear terms are added to standard minimization problems. The local design of the algorithm compensates for the higher computational complexity. Results are shown and evaluated for objects of varying complexity. A comparison with an exact \(G^1\)-continuous method shows that the novel method improves approximation accuracy on average by a factor of 10 at the cost of having small discontinuities in normal vectors of less than 1 degree.
Model-based viewpoint generation for surface inspection: Within modern and flexible factories, surface inspection of products is still a very rigid process. An automated inspection system requires the definition of viewpoints from which a robot then takes pictures during the inspection process. Setting up such a system is a time-intensive process which is primarily done manually by experts. This work presents a purely virtual approach for the generation of viewpoints. Based on an intuitive definition of analytic feature functionals, a non-uniform sampling with respect to inspection-specific criteria is performed on given B-spline models. This leads to the definition of a low number of viewpoint candidates. Results of applying this method to several test objects with varying parameters indicate that good viewpoints can be obtained through a fast process that can be performed fully automatically or interactively through the use of meaningful parameters.
With the technological advancement in the field of robotics, it is now quite practical to acknowledge the actuality of social robots being a part of human's daily life in the next decades. Concerning HRI, the basic expectations from a social robot are to perceive words, emotions, and behaviours, in order to draw several conclusions and adapt its behaviour to realize natural HRI. Henceforth, assessment of human personality traits is essential to bring a sense of appeal and acceptance towards the robot during interaction.
Knowledge of human personality is highly relevant as far as natural and efficient HRI is concerned. The idea is taken from human behaviourism, with humans behaving differently based on the personality trait of the communicating partners. This thesis contributes to the development of personality trait assessment system for intelligent human-robot interaction.
The personality trait assessment system is organized in three separate levels. The first level, known as perceptual level, is responsible for enabling the robot to perceive, recognize and understand human actions in the surrounding environment in order to make sense of the situation. Using psychological concepts and theories, several percepts have been extracted. A study has been conducted to validate the significance of these percepts towards personality traits.
The second level, known as affective level, helps the robot to connect the knowledge acquired in the first level to make higher order evaluations such as assessment of human personality traits. The affective system of the robot is responsible for analysing human personality traits. To the best of our knowledge, this thesis is the first work in the field of human-robot interaction that presents an automatic assessment of human personality traits in real-time using visual information. Using psychology and cognitive studies, many theories has been studied. Two theories have been been used to build the personality trait assessment system: Big Five personality traits assessment and temperament framework for personality traits assessment.
By using the information from the perceptual and affective level, the last level, known as behavioural level, enables the robot to synthesize an appropriate behaviour adapted to human personality traits. Multiple experiments have been conducted with different scenarios. It has been shown that the robot, ROBIN, assesses personality traits correctly during interaction and uses the similarity-attraction principle to behave with similar personality type. For example, if the person is found out to be extrovert, the robot also behaves like an extrovert. However, it also uses the complementary attraction theory to adapt its behaviour and complement the personality of the interaction partner. For example, if the person is found out to be self-centred, the robot behaves like an agreeable in order to flourish human-robot interaction.
Interconnection networks enable fast data communication between components of a digital system. The selection of an appropriate interconnection network and its architecture plays an important role in the development process of the system. The selection of a bad network architecture may significantly delay the communication between components and decrease the overall system performance.
There are various interconnection networks available. Most of them are blocking networks. Blocking means that even though a pair of source and target components may be free, a connection between them might still not be possible due to limited capabilities of the network. Moreover, routing algorithms of blocking networks have to avoid deadlocks and livelocks, which typically does only allow poor real-time guarantees for delivering a message. Nonblocking networks can always manage all requests that are coming from their input components and can therefore deliver all messages in guaranteed time, i.e, with strong real-time guarantees. However, only a few networks are nonblocking and easy to implement. The simplest one is the crossbar network which is a comparably simple circuit with also a simple routing algorithm. However, while its circuit depth of O(log(n)) is optimal, its size increases with O(n^2) and quickly becomes infeasible for large networks. Therefore, the construction of nonblocking networks with a quasipolynomial size O(nlog(n)^a) and polylogarithmic depth O(log(n)^b) turned out as a research problem.
Benes [Clos53; Bene65] networks were the first non blocking networks having an optimal size of O(nlog(n)) and an optimal depth of O(log(n)), but their routing algorithms are quite complicated and require circuits of depth O(log(n)^2) [NaSa82].
Other nonblocking interconnection networks are derived from sorting networks. Essentially, there are merge-based (MBS) and radix-based (RBS) sorting networks. MBS and RBS networks can be both implemented in a pipelined fashion which leads to a big advantage for their circuit implementation. While these networks are nonblocking and can implement all n! permutations, they cannot directly handle partial permutations that frequently occur in practice since not every input component communicates at every point of time with an output component. For merge-based sorting networks, there is a well known general solution called the Batcher-Banyan network. However, for the larger class of radix-based sorting networks this does not work, and there is only one solution known for a particular permutation network.
In this thesis, new nonblocking radix-based interconnection networks are presented. In particular, for a certain permutation network, three routing algorithms are developed and their circuit implementations are evaluated concerning their size, depth, and power consumption. A special extension of these networks allows them to route also partial permutations. Moreover, three general constructions to convert any binary sorter into a ternary split module were presented which is the key to construct a radix-based interconnection network that can cope with partial permutations. The thesis compares also chip designs of these networks with other radix-based sorting networks as well as with the Batcher-Banyan networks as competitors. As a result, it turns out that the proposed radix-based networks are superior and could form the basis of larger manycore architectures.
Although today’s bipeds are capable of demonstrating impressive locomotion skills, in many aspects, there’s still a big gap compared to the capabilities observed in humans. Partially, this is due to the deployed control paradigms that are mostly based on analytical approaches. The analytical nature of those approaches entails strong model dependencies – regarding the robotic platform as well as the environment – which makes them prone to unknown disturbances. Recently, an increasing number of biologically-inspired control approaches have been presented from which a human-like bipedal gait emerges. Although the control structures only rely on proprioceptive sensory information, the smoothness of the motions and the robustness against external disturbances is impressive. Due to the lack of suitable robotic platforms, until today the controllers have been mostly applied to
simulations.
Therefore, as the first step towards a suitable platform, this thesis presents the Compliant Robotic Leg (CARL) that features mono- as well as biarticular actuation. The design is driven by a set of core-requirements that is primarily derived from the biologically-inspired behavior-based bipedal locomotion control (B4LC) and complemented by further functional aspects from biomechanical research. Throughout the design process, CARL is understood as a unified dynamic system that emerges from the interplay of the mechanics, the electronics, and the control. Thus, having an explicit control approach and the respective gait in mind, the influence of each subsystem on the characteristics of the overall system is considered
carefully.
The result is a planar robotic leg whose three joints are driven by five highly integrated linear SEAs– three mono- and two biarticular actuators – with minimized reflected inertia. The SEAs are encapsulated by FPGA-based embedded nodes that are designed to meet the hard application requirements while enabling the deployment of a full-featured robotic framework. CARL’s foot is implemented using a COTS prosthetic foot; the sensor information is obtained from the deformation of its main structure. Both subsystems are integrated into a leg structure that matches the proportions of a human with a size of 1.7 m.
The functionality of the subsystems, as well as the overall system, is validated experimentally. In particular, the final experiment demonstrates a coordinated walking motion and thereby confirms that CARL can produce the desired behavior – a natural looking, human-like gait is emerging from the interplay of the behavior-based walking control and the mechatronic system. CARL is robust regarding impacts, the redundant actuation system can render the desired joint torques/impedances, and the foot system supports the walking structurally while it provides the necessary sensory information. Considering that there is no movement of the upper trunk, the angle and torque profiles are comparable to the ones found in humans.
This thesis investigates how smart sensors can quantify the process of learning. Traditionally, human beings have obtained various skills by inventing technologies. Those who integrate technologies into daily life and enhance their capabilities are called augmented humans. While most existing augmenting human technologies focus on directly assisting specific skills, the objective of this thesis is to assist learning -- the meta-skill to master new skills -- with the aim of long-term augmentations.
Learning consists of cognitive activities such as reading, writing, and watching. It has been considered that tracking them by motion sensors (in the same way as the recognition of physical activities) is a challenging task because dynamic body movements could not be observed during cognitive activities. I have solved this problem with smart sensors monitoring eye movements and physiological signals.
I propose activity recognition methods using sensors built into eyewear computers. Head movements and eye blinks measured by an infrared proximity sensor on Google Glass could classify five activities including reading with 82% accuracy. Head and eye movements measured by electrooculography on JINS MEME could classify four activities with 70% accuracy. In a wild experiment involving seven participants who wore JINS MEME more than two weeks, deep neural networks could detect natural reading activities with 74% accuracy. I demonstrate Wordometer 2.0, an application to estimate the number of rear words on JINS MEME, which was evaluated in a dataset involving five readers with 11% error rate.
Smart sensors can recognize not only activities but also internal states during the activities. I present an expertise recognition method using an eye tracker which performs 70% classification accuracy into three classes using one minute data of reading a textbook, a positive correlation between interest and pupil diameter (p < 0.01), a negative correlation between mental workload and nose temperature measured by an infrared thermal camera (p < 0.05), an interest detection on newspaper articles, and effective gaze and physiological features to estimate self-confidence while solving multiple choice questions and spelling tests of English vocabulary.
The quantified learning process can be utilized for feedback to each learner on the basis of the context. I present HyperMind, an interactive intelligent digital textbook. It can be developed on HyperMind Builder which may be employed to augment any electronic text by multimedia aspects activated via gaze.
Applications mentioned above have already been deployed at several laboratories including Immersive Quantified Learning Lab (iQL-Lab) at the German Research Center for Artificial Intelligence (DFKI).
As visualization as a field matures, the discussion about the development of a
theory of the field becomes increasingly vivid. Despite some voices claiming that
visualization applications would be too different from each other to generalize,
there is a significant push towards a better understanding of the principles underlying
visual data analysis. As of today, visualization is primarily data-driven.
Years of experience in the visalization of all kinds of different data accumulated
a vast reservoir of implicit knowledge in the community of how to best represent
data according to its shape, its format, and what it is meant to express.
This knowledge is complemented by knowledge imported to visualization from
a variety of other fields, for example psychology, vision science, color theory,
and information theory. Yet, a theory of visualization is still only nascent. One
major reason for that is the field's too strong focus on the quantitative aspects
of data analysis. Although when designing visualizations major design decisions
also consider perception and other human factors, the overall appearance
of visualizations as of now is determined primarily by the type and format of
the data to be visualized and its quantitative attributes like scale, range, or
density. This is also reflected by the current approaches in theoretical work on
visualization. The models developed in this regard also concentrate primarily
on perceptual and quantitative aspects of visual data analysis. Qualitative considerations
like the interpretations made by viewers and the conclusions drawn
by analysts currently only play a minor role in the literature. This Thesis contributes
to the nascent theory of visualization by investigating approaches to
the explicit integration of qualitative considerations into visual data analysis.
To this end, it promotes qualitative visual analysis, the explicit discussion of
the interpretation of artifacts and structures in the visualization, of efficient
workflows designed to optimally support an analyst's reasoning strategy and
capturing information about insight provenance, and of design methodology
tailoring visualizations towards the insights they are meant to provide rather
than to the data they show. Towards this aim, three central qualitative principles
of visual information encodings are identified during the development of
a model for the visual data analysis process that explicitly includes the anticipated
reasoning structure into the consideration. This model can be applied
throughout the whole life cycle of a visualization application, from the early
design phase to the documentation of insight provenance during analysis using
the developed visualization application. The three principles identified inspire
novel visual data analysis workflows aiming for an insight-driven data analysis
process. Moreover, two case studies prove the benefit of following the qualitative
principles of visual information encodings for the design of visualization
applications. The formalism applied to the development of the presented theoretical
framework is founded in formal logics, mathematical set theory, and the
theory of formal languages and automata. The models discussed in this Thesis
and the findings derived from them are therefore based on a mathematically
well-founded theoretical underpinning. This Thesis establishes a sound theoretical
framework for the design and description of visualization applications and
the prediction of the conclusions an analyst is capable of drawing from working
with the visualization. Thereby, it contributes an important piece to the yet
unsolved puzzle of developing a visualization theory.