006 Spezielle Computerverfahren
Refine
Departments, institutes and facilities
- Fachbereich Informatik (24)
- Institute of Visual Computing (IVC) (11)
- Fachbereich Wirtschaftswissenschaften (5)
- Institut für Sicherheitsforschung (ISF) (4)
- Institut für Technik, Ressourcenschonung und Energieeffizienz (TREE) (4)
- Institut für Verbraucherinformatik (IVI) (4)
- Fachbereich Ingenieurwissenschaften und Kommunikation (2)
- Institut für KI und Autonome Systeme (A2S) (2)
- Fachbereich Angewandte Naturwissenschaften (1)
- Institut für Cyber Security & Privacy (ICSP) (1)
Document Type
- Conference Object (40) (remove)
Year of publication
Language
- English (40) (remove)
Keywords
- Augmented Reality (3)
- Machine Learning (2)
- Robotics (2)
- authoring tools (2)
- prototyping (2)
- 450 MHz (1)
- AR design (1)
- AR development (1)
- AR/VR (1)
- Applications in Energy Transport (1)
- Automatic Differentiation (1)
- Behaviour-Driven Development (1)
- Bioinformatics (1)
- Collaborating industrial robots (1)
- Complex Systems Modeling and Simulation (1)
- Compliant fingers (1)
- Computational fluid dynamics (1)
- Concurrent repeated failure prognosis (1)
- Crossmedia (1)
- Data Fusion (1)
- Design Recommendations (1)
- Design Theory and Practice (1)
- Diagnostic bond graph-based online fault diagnosis (1)
- Embedded system (1)
- Explainable artificial intelligence (1)
- Flow control (1)
- Forests (1)
- Functional safety (1)
- Games and Simulations for Learning (1)
- Guidelines (1)
- Higher education (1)
- Human-Food-Interaction (1)
- Hyperspectral image (1)
- IEC 104 (1)
- IEC 61850 (1)
- Increasing fault magnitude (1)
- Inductive Logic Programming (1)
- Information Security (1)
- Intermittent faults (1)
- Kalman filter (1)
- Knowledge Graphs (1)
- LTE-M (1)
- Lattice Boltzmann Method (1)
- MQTT (1)
- MR (1)
- Mixed Reality (1)
- Model-driven engineering (1)
- NIR-point sensor (1)
- Natural Language Processing (1)
- Negotiation of Taste (1)
- Non-linear systems (1)
- Object-Based Image Analysis (OBIA) (1)
- Pytorch (1)
- Raman microscopy (1)
- Reasoning (1)
- Remaining Useful Life (RUL) estimates (1)
- Requirements (1)
- Review (1)
- Robust grasping (1)
- Serious Games (1)
- Skin detection (1)
- Slippage detection (1)
- Smart Grid (1)
- Smart InGaAs camera-system (1)
- Survey (1)
- Taste (1)
- Traffic Simulations (1)
- Transformers (1)
- Tree Stumps (1)
- Ultrasonic array (1)
- Unmanned Aerial Vehicle (UAV) (1)
- User Interface Design (1)
- User-Centered Design (1)
- Virtual Agents (1)
- Virtual Reality (1)
- Vulnerable Groups (1)
- XR (1)
- action unit recognition (1)
- audio-tactile feedback (1)
- augmented reality (1)
- authoring (1)
- brightfield microscopy (1)
- co-design (1)
- component analyses (1)
- deep learning (1)
- depth perception (1)
- emotion recognition (1)
- facial expression analysis (1)
- guidance (1)
- image fusion (1)
- interaction design (1)
- interface design (1)
- mixed reality (1)
- neural networks (1)
- pain recognition (1)
- pansharpening (1)
- practitioners (1)
- remote sensing (1)
- state constraints (1)
- virtual reality, XR (1)
Towards self-explaining social robots. Verbal explanation strategies for a needs-based architecture
(2019)
In order to establish long-term relationships with users, social companion robots and their behaviors need to be comprehensible. Purely reactive behavior such as answering questions or following commands can be readily interpreted by users. However, the robot's proactive behaviors, included in order to increase liveliness and improve the user experience, often raise a need for explanation. In this paper, we provide a concept to produce accessible “why-explanations” for the goal-directed behavior an autonomous, lively robot might produce. To this end we present an architecture that provides reasons for behaviors in terms of comprehensible needs and strategies of the robot, and we propose a model for generating different kinds of explanations.
Towards explaining deep learning networks to distinguish facial expressions of pain and emotions
(2018)
Deep learning networks are successfully used for object and face recognition in images and videos. In order to be able to apply such networks in practice, for example in hospitals as a pain recognition tool, the current procedures are only suitable to a limited extent. The advantage of deep learning methods is that they can learn complex non-linear relationships between raw data and target classes without limiting themselves to a set of hand-crafted features provided by humans. However, the disadvantage is that due to the complexity of these networks, it is not possible to interpret the knowledge that is stored inside the network. It is a black-box learning procedure. Explainable Artificial Intelligence (AI) approaches mitigate this problem by extracting explanations for decisions and representing them in a human-interpretable form. The aim of this paper is to investigate the explainable AI method Layer-wise Relevance Propagation (LRP) and apply it to explain how a deep learning network distinguishes facial expressions of pain from facial expressions of emotions such as happiness and disgust.
Towards an Interaction-Centered and Dynamically Constructed Episodic Memory for Social Robots
(2020)
Selection Performance and Reliability of Eye and Head Gaze Tracking Under Varying Light Conditions
(2024)
Over the last decades, different kinds of design guides have been created to maintain consistency and usability in interactive system development. However, in the case of spatial applications, practitioners from research and industry either have difficulty finding them or perceive such guides as lacking relevance, practicability, and applicability. This paper presents the current state of scientific research and industry practice by investigating currently used design recommendations for mixed reality (MR) system development. We analyzed and compared 875 design recommendations for MR applications elicited from 89 scientific papers and documentation from six industry practitioners in a literature review. In doing so, we identified differences regarding four key topics: Focus on unique MR design challenges, abstraction regarding devices and ecosystems, level of detail and abstraction of content, and covered topics. Based on that,we contribute to the MR design research by providing three factors for perceived irrelevance and six main implications for design recommendations that are applicable in scientific and industry practice.
In this paper, we introduce an optical sensor system, which is integrated into an industrial push-button. The sensor allows to classify the type of material that is in contact with the button when pressed into different material categories on the basis of the material's so called "spectral signature". An approach for a safety sensor system at circular table saws on the same base has been introduced previously on SIAS-2007. This contactless working sensor is able to distinguish reliably between skin, textiles, leather and various other kinds of materials. A typical application for this intelligent push-button is the use at possibly dangerous machines, whose operating instructions include either the prohibition or the obligation to wear gloves during the work at the machine. An exemple of machines at which no gloves are allowed are pillar drilling machines, because of the risk of getting caught in the drill chuck and being turned in by the machine. In many cases this causes very serious hand injuries. Depending on the application needs, the sensor system integrated into the push-button can be configured flexibly by software to prevent the operator from accidentally starting a machine with or without gloves, which can decrease the risk of severe accidents significantly. Especially two-hand controls are incentive to manipulation for easier handling. By equipping both push-buttons of a two-hand control with material classification properties, the user is forced to operate the controls with his bare fingers. That limitation disallows the manipulation of a two-hand control by a simple rodding device.
ProtSTonKGs: A Sophisticated Transformer Trained on Protein Sequences, Text, and Knowledge Graphs
(2022)
While most approaches individually exploit unstructured data from the biomedical literature or structured data from biomedical knowledge graphs, their union can better exploit the advantages of such approaches, ultimately improving representations of biology. Using multimodal transformers for such purposes can improve performance on context dependent classication tasks, as demonstrated by our previous model, the Sophisticated Transformer Trained on Biomedical Text and Knowledge Graphs (STonKGs). In this work, we introduce ProtSTonKGs, a transformer aimed at learning all-encompassing representations of protein-protein interactions. ProtSTonKGs presents an extension to our previous work by adding textual protein descriptions and amino acid sequences (i.e., structural information) to the text- and knowledge graph-based input sequence used in STonKGs. We benchmark ProtSTonKGs against STonKGs, resulting in improved F1 scores by up to 0.066 (i.e., from 0.204 to 0.270) in several tasks such as predicting protein interactions in several contexts. Our work demonstrates how multimodal transformers can be used to integrate heterogeneous sources of information, paving the foundation for future approaches that use multiple modalities for biomedical applications.
Taste is a complex phenomenon that depends on the individual experience and is a matter of collective negotiation and mediation. On the contrary, it is uncommon to include taste and its many facets in everyday design, particularly online shopping for fresh food products. To realize this unused potential, we conducted two Co-Design workshops. Based on the participants’ results in the workshops, we prototyped and evaluated a click-dummy smart-phone app to explore consumers’ needs for digital taste depiction. We found that emphasizing the natural qualities of food products, external reviews, and personalizing features lead to a reflection on the individual taste experience. The self-reflection through our design enables consumers to develop their taste competencies and thus strengthen their autonomy in decision-making. Ultimately, exploring taste as a social experience adds to a broader understanding of taste beyond a sensory phenomenon.