006 Spezielle Computerverfahren
Refine
H-BRS Bibliography
- yes (43)
Departments, institutes and facilities
- Fachbereich Informatik (43) (remove)
Document Type
- Conference Object (24)
- Article (9)
- Preprint (3)
- Doctoral Thesis (2)
- Report (2)
- Part of a Book (1)
- Contribution to a Periodical (1)
- Research Data (1)
Year of publication
Has Fulltext
- no (43) (remove)
Keywords
- Augmented Reality (3)
- Machine Learning (2)
- Robotics (2)
- Virtual Reality (2)
- guidance (2)
- 3D user interface (1)
- 450 MHz (1)
- Altenhilfe (1)
- Auditory Cueing (1)
- Bayesian Deep Learning (1)
- Behaviour-Driven Development (1)
- Bioinformatics (1)
- Blasendiagramm (1)
- Business Process Intelligence (1)
- Classifiers (1)
- Collaborating industrial robots (1)
- Compliant fingers (1)
- Computergrafik (1)
- Concurrent repeated failure prognosis (1)
- Conformation (1)
- Crossmedia (1)
- Crystal structure (1)
- Curriculum (1)
- Cybersickness (1)
- Data Fusion (1)
- Datenanalyse (1)
- Demenz (1)
- Diagnostic bond graph-based online fault diagnosis (1)
- Disco (1)
- Distance Perception (1)
- Educational Data Mining (1)
- Educational Process Mining (1)
- Embedded system (1)
- Emotion (1)
- Facial Emotion Recognition (1)
- Fallbeschreibung (1)
- Fluency (1)
- Forests (1)
- Functional safety (1)
- Fuzzy Mining (1)
- Games and Simulations for Learning (1)
- Geschäftsprozess (1)
- Head-mounted Display (1)
- Higher education (1)
- Human factors (1)
- Hyperspectral image (1)
- IEC 104 (1)
- IEC 61850 (1)
- Increasing fault magnitude (1)
- Inductive Logic Programming (1)
- Inductive Visual Mining (1)
- Information Security (1)
- Intermittent faults (1)
- Knowledge Graphs (1)
- LTE-M (1)
- Language learning (1)
- Langzeitbehandlung (1)
- Ligands (1)
- Locomotion (1)
- MQTT (1)
- Mathematical methods (1)
- Microgravity (1)
- Model-driven engineering (1)
- Molecular structure (1)
- Motion Sickness (1)
- NIR-point sensor (1)
- Natural Language Processing (1)
- Object-Based Image Analysis (OBIA) (1)
- Out-of-view Objects (1)
- Pflegepersonal (1)
- ProM (1)
- Process Mining (1)
- Pronunciation (1)
- Raman microscopy (1)
- RapidMiner (1)
- Ray tracing (1)
- Reasoning (1)
- Remaining Useful Life (RUL) estimates (1)
- Requirements (1)
- Review (1)
- Robust grasping (1)
- Serious Games (1)
- Skin detection (1)
- Slippage detection (1)
- Smart Grid (1)
- Smart InGaAs camera-system (1)
- Studenten (1)
- Studienverlauf (1)
- Survey (1)
- Technologie (1)
- Traffic Simulations (1)
- Transformers (1)
- Travel Techniques (1)
- Tree Stumps (1)
- UAV (1)
- Ultrasonic array (1)
- Uncertainty Quantification (1)
- Underwater (1)
- Unmanned Aerial Vehicle (UAV) (1)
- Unterstützung (1)
- Virtual Agents (1)
- Virtuelle Realität (1)
- Visual Cueing (1)
- Visual Discrimination (1)
- Visuelle Wahrnehmung (1)
- Vulnerable Groups (1)
- aerodynamics (1)
- audio-tactile feedback (1)
- authoring tools (1)
- brightfield microscopy (1)
- component analyses (1)
- depth perception (1)
- dynamic vector fields (1)
- flight zone (1)
- geofence (1)
- haptics (1)
- image fusion (1)
- multisensory (1)
- neutral buoyancy (1)
- optic flow (1)
- pansharpening (1)
- remote sensing (1)
- self-motion perception (1)
- sensory perception (1)
- vection (1)
- virtual reality (1)
Selection Performance and Reliability of Eye and Head Gaze Tracking Under Varying Light Conditions
(2024)
This research investigates the efficacy of multisensory cues for locating targets in Augmented Reality (AR). Sensory constraints can impair perception and attention in AR, leading to reduced performance due to factors such as conflicting visual cues or a restricted field of view. To address these limitations, the research proposes head-based multisensory guidance methods that leverage audio-tactile cues to direct users' attention towards target locations. The research findings demonstrate that this approach can effectively reduce the influence of sensory constraints, resulting in improved search performance in AR. Additionally, the thesis discusses the limitations of the proposed methods and provides recommendations for future research.
This paper addresses the classification of Arabic text data in the field of Natural Language Processing (NLP), with a particular focus on Natural Language Inference (NLI) and Contradiction Detection (CD). Arabic is considered a resource-poor language, meaning that there are few data sets available, which leads to limited availability of NLP methods. To overcome this limitation, we create a dedicated data set from publicly available resources. Subsequently, transformer-based machine learning models are being trained and evaluated. We find that a language-specific model (AraBERT) performs competitively with state-of-the-art multilingual approaches, when we apply linguistically informed pre-training methods such as Named Entity Recognition (NER). To our knowledge, this is the first large-scale evaluation for this task in Arabic, as well as the first application of multi-task pre-training in this context.
Vection underwater
(2022)
The majority of biomedical knowledge is stored in structured databases or as unstructured text in scientific publications. This vast amount of information has led to numerous machine learning-based biological applications using either text through natural language processing (NLP) or structured data through knowledge graph embedding models (KGEMs). However, representations based on a single modality are inherently limited. To generate better representations of biological knowledge, we propose STonKGs, a Sophisticated Transformer trained on biomedical text and Knowledge Graphs. This multimodal Transformer uses combined input sequences of structured information from KGs and unstructured text data from biomedical literature to learn joint representations. First, we pre-trained STonKGs on a knowledge base assembled by the Integrated Network and Dynamical Reasoning Assembler (INDRA) consisting of millions of text-triple pairs extracted from biomedical literature by multiple NLP systems. Then, we benchmarked STonKGs against two baseline models trained on either one of the modalities (i.e., text or KG) across eight different classification tasks, each corresponding to a different biological application. Our results demonstrate that STonKGs outperforms both baselines, especially on the more challenging tasks with respect to the number of classes, improving upon the F1-score of the best baseline by up to 0.083. Additionally, our pre-trained model as well as the model architecture can be adapted to various other transfer learning applications. Finally, the source code and pre-trained STonKGs models are available at https://github.com/stonkgs/stonkgs and https://huggingface.co/stonkgs/stonkgs-150k.
Using Visual and Auditory Cues to Locate Out-of-View Objects in Head-Mounted Augmented Reality
(2021)
Low-Cost In-Hand Slippage Detection and Avoidance for Robust Robotic Grasping with Compliant Fingers
(2021)
Facial emotion recognition is the task to classify human emotions in face images. It is a difficult task due to high aleatoric uncertainty and visual ambiguity. A large part of the literature aims to show progress by increasing accuracy on this task, but this ignores the inherent uncertainty and ambiguity in the task. In this paper we show that Bayesian Neural Networks, as approximated using MC-Dropout, MC-DropConnect, or an Ensemble, are able to model the aleatoric uncertainty in facial emotion recognition, and produce output probabilities that are closer to what a human expects. We also show that calibration metrics show strange behaviors for this task, due to the multiple classes that can be considered correct, which motivates future work. We believe our work will motivate other researchers to move away from Classical and into Bayesian Neural Networks.