Refine
H-BRS Bibliography
- yes (65) (remove)
Departments, institutes and facilities
- Fachbereich Informatik (65) (remove)
Document Type
- Conference Object (23)
- Article (18)
- Preprint (9)
- Report (4)
- Book (monograph, edited volume) (2)
- Part of a Book (2)
- Conference Proceedings (2)
- Research Data (2)
- Doctoral Thesis (2)
- Master's Thesis (1)
Year of publication
- 2022 (65) (remove)
Keywords
- Machine Learning (4)
- Knowledge Graphs (3)
- virtual reality (3)
- 3D user interface (2)
- Bioinformatics (2)
- Computer Vision (2)
- Control Systems and Automation (2)
- Deep Learning (2)
- Electrical Machines and Power Electronics (2)
- Intelligent controls (2)
- Natural Language Processing (2)
- Renewable Energy Systems (2)
- Security (2)
- Transformers (2)
- Usable Security (2)
- embedded systems (2)
- haptics (2)
- usable privacy (2)
- 3D Segmentation (1)
- AI usage in sports (1)
- ANSYS (1)
- Abstract Syntax Tree (1)
- Air Pollution Monitoring (1)
- Artificial Intelligence (1)
- Artificial Intelligence (cs.AI) (1)
- Augmented Reality (1)
- Autonomous Driving (1)
- Ballastless track (1)
- Beacon Chain (1)
- Bounding box explanations (1)
- Classification explanations (1)
- Cloud computing (1)
- Complexity (1)
- Conformation (1)
- Corporate Social Responsibility (1)
- Crystal structure (1)
- Current research information systems (1)
- Cypher (1)
- Data Generation (1)
- Digitaler Stress (1)
- Distance Perception (1)
- Employee Privacy (1)
- Environment Perception (1)
- Ethereum (1)
- Experiment (1)
- Explainable Artificial Intelligence (XAI) (1)
- FOS: Computer and information sciences (1)
- Feedback (1)
- GDPR (1)
- Gradient-based explanation methods (1)
- Graph Convolutional Neural Networks (1)
- Graph embeddings (1)
- Graph theory (1)
- HDBR (1)
- High-speed track (1)
- Highly Automated Driving (1)
- Human orientation perception (1)
- Human-Centered Design (1)
- IaaS (1)
- Image representation (1)
- Information Privacy (1)
- IoT (1)
- LSTM (1)
- Ligands (1)
- LoRa (1)
- LoRaWAN (1)
- MOX gas sensors (1)
- Machine Learning (cs.LG) (1)
- Malware (1)
- Markov Cluster Algorithm (1)
- Mathematical methods (1)
- Microgravity (1)
- Molecular structure (1)
- Multi-object visualization (1)
- NLP (1)
- Neural Machine Translation (1)
- Noise reduction (1)
- OCT (1)
- Object Segmentation (1)
- Object detectors (1)
- Open Access (1)
- PAD (1)
- PaaS (1)
- Parametric study (1)
- Part Segmentation (1)
- Perception (1)
- Perceptual Upright (1)
- Performance (1)
- Persönlichkeit (1)
- Persönlichkeitseigenschaften (1)
- Persönlichkeitsfaktor (1)
- Point Cloud Segmentation (1)
- Proof of Stake (1)
- Proximity (1)
- Quantitative analysis of explanations (1)
- Right to Informational Self-Determination (1)
- SMPA loop (1)
- SQL (1)
- SaaS (1)
- Safety (1)
- Saliency maps (1)
- Sanity checks for explaining detectors (1)
- Scalability (1)
- Self-supervised learning (1)
- Semantic Segmentation (1)
- Semantic search (1)
- Service-based cloud computing (1)
- Simulator (1)
- Software Supply Chain (1)
- Sonar (1)
- Supervised learning (1)
- TLS (1)
- Transfer learning (1)
- Underwater (1)
- Usable Privacy (1)
- adaptive trigger (1)
- authentication (1)
- automatic measurement validation (1)
- automatic music generation (1)
- biometrics (1)
- blockchain (1)
- collision (1)
- controller design (1)
- cybersickness (1)
- digital co-creation (1)
- elite sports (1)
- employee privacy (1)
- explainable AI (1)
- factor analysis (1)
- feature (1)
- fingerprint (1)
- flying (1)
- guidance (1)
- head down bed rest (1)
- human-centred design (1)
- hybrid system (1)
- intervention mechanisms (1)
- latent class analysis (1)
- low-cost air sensor (1)
- motion sickness (1)
- multisensory cues (1)
- music analysis (1)
- neutral buoyancy (1)
- optic flow (1)
- optical coherence tomography (1)
- presentation attack detection (1)
- prioritizable ranking (1)
- privacy by design (1)
- psychophysics (1)
- representation learning (1)
- security and privacy literacy (1)
- self-motion perception (1)
- situation awareness (1)
- slope based signature (1)
- space flight analog (1)
- structural equation modeling (1)
- subjective visual vertical (1)
- time series analysis (1)
- traffic surveillance (1)
- transparency-enhancing technologies (1)
- travel techniques (1)
- vection (1)
- vibration (1)
- view management (1)
- weight perception (1)
Robots applied in therapeutic scenarios, for instance in the therapy of individuals with Autism Spectrum Disorder, are sometimes used for imitation learning activities in which a person needs to repeat motions by the robot. To simplify the task of incorporating new types of motions that a robot can perform, it is desirable that the robot has the ability to learn motions by observing demonstrations from a human, such as a therapist. In this paper, we investigate an approach for acquiring motions from skeleton observations of a human, which are collected by a robot-centric RGB-D camera. Given a sequence of observations of various joints, the joint positions are mapped to match the configuration of a robot before being executed by a PID position controller. We evaluate the method, in particular the reproduction error, by performing a study with QTrobot in which the robot acquired different upper-body dance moves from multiple participants. The results indicate the method's overall feasibility, but also indicate that the reproduction quality is affected by noise in the skeleton observations.
Trojanized software packages used in software supply chain attacks constitute an emerging threat. Unfortunately, there is still a lack of scalable approaches that allow automated and timely detection of malicious software packages and thus most detections are based on manual labor and expertise. However, it has been observed that most attack campaigns comprise multiple packages that share the same or similar malicious code. We leverage that fact to automatically reproduce manually identified clusters of known malicious packages that have been used in real world attacks, thus, reducing the need for expert knowledge and manual inspection. Our approach, AST Clustering using MCL to mimic Expertise (ACME), yields promising results with a 𝐹1 score of 0.99. Signatures are automatically generated based on characteristic code fragments from clusters and are subsequently used to scan the whole npm registry for unreported malicious packages. We are able to identify and report six malicious packages that have been removed from npm consequentially. Therefore, our approach can support the detection by reducing manual labor and hence may be employed by maintainers of package repositories to detect possible software supply chain attacks through trojanized software packages.
State-of-the-art object detectors are treated as black boxes due to their highly non-linear internal computations. Even with unprecedented advancements in detector performance, the inability to explain how their outputs are generated limits their use in safety-critical applications. Previous work fails to produce explanations for both bounding box and classification decisions, and generally make individual explanations for various detectors. In this paper, we propose an open-source Detector Explanation Toolkit (DExT) which implements the proposed approach to generate a holistic explanation for all detector decisions using certain gradient-based explanation methods. We suggests various multi-object visualization methods to merge the explanations of multiple objects detected in an image as well as the corresponding detections in a single image. The quantitative evaluation show that the Single Shot MultiBox Detector (SSD) is more faithfully explained compared to other detectors regardless of the explanation methods. Both quantitative and human-centric evaluations identify that SmoothGrad with Guided Backpropagation (GBP) provides more trustworthy explanations among selected methods across all detectors. We expect that DExT will motivate practitioners to evaluate object detectors from the interpretability perspective by explaining both bounding box and classification decisions.
21 pages, with supplementary
Deployment of modern data-driven machine learning methods, most often realized by deep neural networks (DNNs), in safety-critical applications such as health care, industrial plant control, or autonomous driving is highly challenging due to numerous model-inherent shortcomings. These shortcomings are diverse and range from a lack of generalization over insufficient interpretability and implausible predictions to directed attacks by means of malicious inputs. Cyber-physical systems employing DNNs are therefore likely to suffer from so-called safety concerns, properties that preclude their deployment as no argument or experimental setup can help to assess the remaining risk. In recent years, an abundance of state-of-the-art techniques aiming to address these safety concerns has emerged. This chapter provides a structured and broad overview of them. We first identify categories of insufficiencies to then describe research activities aiming at their detection, quantification, or mitigation. Our work addresses machine learning experts and safety engineers alike: The former ones might profit from the broad range of machine learning topics covered and discussions on limitations of recent methods. The latter ones might gain insights into the specifics of modern machine learning methods. We hope that this contribution fuels discussions on desiderata for machine learning systems and strategies on how to help to advance existing approaches accordingly.
This open access book brings together the latest developments from industry and research on automated driving and artificial intelligence.
Environment perception for highly automated driving heavily employs deep neural networks, facing many challenges. How much data do we need for training and testing? How to use synthetic data to save labeling costs for training? How do we increase robustness and decrease memory usage? For inevitably poor conditions: How do we know that the network is uncertain about its decisions? Can we understand a bit more about what actually happens inside neural networks? This leads to a very practical problem particularly for DNNs employed in automated driving: What are useful validation techniques and how about safety?
This book unites the views from both academia and industry, where computer vision and machine learning meet environment perception for highly automated driving. Naturally, aspects of data, robustness, uncertainty quantification, and, last but not least, safety are at the core of it. This book is unique: In its first part, an extended survey of all the relevant aspects is provided. The second part contains the detailed technical elaboration of the various questions mentioned above.
In the field of automatic music generation, one of the greatest challenges is the consistent generation of pieces continuously perceived positively by the majority of the audience since there is no objective method to determine the quality of a musical composition. However, composing principles, which have been refined for millennia, have shaped the core characteristics of today's music. A hybrid music generation system, mlmusic, that incorporates various static, music-theory-based methods, as well as data-driven, subsystems, is implemented to automatically generate pieces considered acceptable by the average listener. Initially, a MIDI dataset, consisting of over 100 hand-picked pieces of various styles and complexities, is analysed using basic music theory principles, and the abstracted information is fed into explicitly constrained LSTM networks. For chord progressions, each individual network is specifically trained on a given sequence length, while phrases are created by consecutively predicting the notes' offset, pitch and duration. Using these outputs as a composition's foundation, additional musical elements, along with constrained recurrent rhythmic and tonal patterns, are statically generated. Although no survey regarding the pieces' reception could be carried out, the successful generation of numerous compositions of varying complexities suggests that the integration of these fundamentally distinctive approaches might lead to success in other branches.
Vietnam requires a sustainable urbanization, for which city sensing is used in planning and de-cision-making. Large cities need portable, scalable, and inexpensive digital technology for this purpose. End-to-end air quality monitoring companies such as AirVisual and Plume Air have shown their reliability with portable devices outfitted with superior air sensors. They are pricey, yet homeowners use them to get local air data without evaluating the causal effect. Our air quality inspection system is scalable, reasonably priced, and flexible. Minicomputer of the sys-tem remotely monitors PMS7003 and BME280 sensor data through a microcontroller processor. The 5-megapixel camera module enables researchers to infer the causal relationship between traffic intensity and dust concentration. The design enables inexpensive, commercial-grade hardware, with Azure Blob storing air pollution data and surrounding-area imagery and pre-venting the system from physically expanding. In addition, by including an air channel that re-plenishes and distributes temperature, the design improves ventilation and safeguards electrical components. The gadget allows for the analysis of the correlation between traffic and air quali-ty data, which might aid in the establishment of sustainable urban development plans and poli-cies.
Fatigue strength estimation is a costly manual material characterization process in which state-of-the-art approaches follow a standardized experiment and analysis procedure. In this paper, we examine a modular, Machine Learning-based approach for fatigue strength estimation that is likely to reduce the number of experiments and, thus, the overall experimental costs. Despite its high potential, deployment of a new approach in a real-life lab requires more than the theoretical definition and simulation. Therefore, we study the robustness of the approach against misspecification of the prior and discretization of the specified loads. We identify its applicability and its advantageous behavior over the state-of-the-art methods, potentially reducing the number of costly experiments.
Safety-critical applications like autonomous driving use Deep Neural Networks (DNNs) for object detection and segmentation. The DNNs fail to predict when they observe an Out-of-Distribution (OOD) input leading to catastrophic consequences. Existing OOD detection methods were extensively studied for image inputs but have not been explored much for LiDAR inputs. So in this study, we proposed two datasets for benchmarking OOD detection in 3D semantic segmentation. We used Maximum Softmax Probability and Entropy scores generated using Deep Ensembles and Flipout versions of RandLA-Net as OOD scores. We observed that Deep Ensembles out perform Flipout model in OOD detection with greater AUROC scores for both datasets.
Login Data Set for Risk-Based Authentication
Synthesized login feature data of >33M login attempts and >3.3M users on a large-scale online service in Norway. Original data collected between February 2020 and February 2021.
This data sets aims to foster research and development for <a href="https://riskbasedauthentication.org">Risk-Based Authentication (RBA) systems. The data was synthesized from the real-world login behavior of more than 3.3M users at a large-scale single sign-on (SSO) online service in Norway.
Robust Identification and Segmentation of the Outer Skin Layers in Volumetric Fingerprint Data
(2022)
Despite the long history of fingerprint biometrics and its use to authenticate individuals, there are still some unsolved challenges with fingerprint acquisition and presentation attack detection (PAD). Currently available commercial fingerprint capture devices struggle with non-ideal skin conditions, including soft skin in infants. They are also susceptible to presentation attacks, which limits their applicability in unsupervised scenarios such as border control. Optical coherence tomography (OCT) could be a promising solution to these problems. In this work, we propose a digital signal processing chain for segmenting two complementary fingerprints from the same OCT fingertip scan: One fingerprint is captured as usual from the epidermis (“outer fingerprint”), whereas the other is taken from inside the skin, at the junction between the epidermis and the underlying dermis (“inner fingerprint”). The resulting 3D fingerprints are then converted to a conventional 2D grayscale representation from which minutiae points can be extracted using existing methods. Our approach is device-independent and has been proven to work with two different time domain OCT scanners. Using efficient GPGPU computing, it took less than a second to process an entire gigabyte of OCT data. To validate the results, we captured OCT fingerprints of 130 individual fingers and compared them with conventional 2D fingerprints of the same fingers. We found that both the outer and inner OCT fingerprints were backward compatible with conventional 2D fingerprints, with the inner fingerprint generally being less damaged and, therefore, more reliable.
Ziel der achten Auflage des wissenschaftlichen Workshops “Usable Security and Privacy” auf der Mensch und Computer 2022 ist es, aktuelle Forschungs- und Praxisbeiträge zu präsentieren und anschließend mit den Teilnehmenden zu diskutieren. Der Workshop soll ein etabliertes Forum fortführen und weiterentwickeln, in dem sich Experten aus verschiedenen Bereichen, z. B. Usability und Security Engineering, transdisziplinär austauschen können.
The visual and auditory quality of computer-mediated stimuli for virtual and extended reality (VR/XR) is rapidly improving. Still, it remains challenging to provide a fully embodied sensation and awareness of objects surrounding, approaching, or touching us in a 3D environment, though it can greatly aid task performance in a 3D user interface. For example, feedback can provide warning signals for potential collisions (e.g., bumping into an obstacle while navigating) or pinpointing areas where one’s attention should be directed to (e.g., points of interest or danger). These events inform our motor behaviour and are often associated with perception mechanisms associated with our so-called peripersonal and extrapersonal space models that relate our body to object distance, direction, and contact point/impact. We will discuss these references spaces to explain the role of different cues in our motor action responses that underlie 3D interaction tasks. However, providing proximity and collision cues can be challenging. Various full-body vibration systems have been developed that stimulate body parts other than the hands, but can have limitations in their applicability and feasibility due to their cost and effort to operate, as well as hygienic considerations associated with e.g., Covid-19. Informed by results of a prior study using low-frequencies for collision feedback, in this paper we look at an unobtrusive way to provide spatial, proximal and collision cues. Specifically, we assess the potential of foot sole stimulation to provide cues about object direction and relative distance, as well as collision direction and force of impact. Results indicate that in particular vibration-based stimuli could be useful within the frame of peripersonal and extrapersonal space perception that support 3DUI tasks. Current results favor the feedback combination of continuous vibrotactor cues for proximity, and bass-shaker cues for body collision. Results show that users could rather easily judge the different cues at a reasonably high granularity. This granularity may be sufficient to support common navigation tasks in a 3DUI.