Refine
H-BRS Bibliography
- yes (51) (remove)
Departments, institutes and facilities
- Fachbereich Informatik (41)
- Institute of Visual Computing (IVC) (9)
- Fachbereich Wirtschaftswissenschaften (4)
- Institut für Verbraucherinformatik (IVI) (3)
- Fachbereich Ingenieurwissenschaften und Kommunikation (2)
- Fachbereich Sozialpolitik und Soziale Sicherung (1)
- Zentrum für Ethik und Verantwortung (ZEV) (1)
Document Type
- Report (51) (remove)
Year of publication
Language
- English (51) (remove)
Keywords
- Robotik (5)
- Deep Learning (2)
- Machine Learning (2)
- Virtuelle Realität (2)
- 3D Segmentation (1)
- 3D-Scanner (1)
- 802.11 (1)
- Adaptive Behavior (1)
- Adaptive Case Management (1)
- Agents (1)
- Apprenticeship Learning (1)
- Assistenzsystem (1)
- Benchmark (1)
- Bildverarbeitung (1)
- Bioinformatics (1)
- Blockchain (1)
- Calibration (1)
- Centrifuge (1)
- Cloud Computing (1)
- Comparative Analysis (1)
- Concurrent Kleene Algebra (1)
- Convexity (1)
- Created Gravity (1)
- Cryptography (1)
- CyberGlove (1)
- DCF (1)
- Declarative Process Modeling (1)
- Domain-Specific Language (1)
- Domestic Robots (1)
- Dynamic Case Management (1)
- Emotion (1)
- EnOcean (1)
- FPGA (1)
- FS20 (1)
- Five Factor Model (1)
- Forschungsbericht (1)
- GDDL (1)
- Gabor filters (1)
- Graph Convolutional Neural Networks (1)
- Grasp Domain Definition Language (1)
- Grasp Planner (1)
- Grasping (1)
- Gravitation (1)
- Healthcare logistics (1)
- HomeMatic (1)
- Human-Centered Robotics (1)
- ICF (1)
- ISO9999 (1)
- Image Classification (1)
- Instantaneous assignment (1)
- Interaktion (1)
- KNX (1)
- Knowledge Graphs (1)
- Knowledge Worker (1)
- Knowledge-intensive Process (1)
- Kosovo (1)
- LBP (1)
- LDP (1)
- LSTM (1)
- Laws of programming (1)
- Learning and Adaptive Systems (1)
- Long-Term Autonomy (1)
- METEOR score (1)
- Method of lines (1)
- Mobiler Roboter (1)
- Multi-robot systems (1)
- Naive physics (1)
- Natural Language Processing (1)
- Object Segmentation (1)
- Out Of Distribution (OOD) data (1)
- Outer Space Research (1)
- Part Segmentation (1)
- Peer methods (1)
- Perception (1)
- Perceptual Upright (1)
- Personality (1)
- Point Cloud Segmentation (1)
- Point Clouds (1)
- Process Automation (1)
- Proof-of-Stake (1)
- Proof-of-Work (1)
- Prozessautomation (1)
- Qualitative reasoning (1)
- RGB-D (1)
- ROPOD (1)
- Raumwahrnehmung (1)
- Refinement (1)
- Robotic faults (1)
- Scene understanding through Deep Learning (1)
- Segmentation (1)
- Semantic Segmentation (1)
- Semantic models (1)
- Shallow water equations (1)
- Spatio-Temporal (1)
- Spectral Analysis (1)
- Spectral Clustering (1)
- Task allocation (1)
- Temporal constraints (1)
- Time extended assignment (1)
- Trace algebra (1)
- Transformers (1)
- Uncertainty Estimation (1)
- Unifying theories (1)
- Virtual Reality (1)
- WENO-schemes (1)
- ZWave (1)
- ZigBee (1)
- assistive robots (1)
- automatic music generation (1)
- binary classification (1)
- building automation (1)
- camera (1)
- computer vision (1)
- constraint relaxation (1)
- control (1)
- control architectures (1)
- convex optimization (1)
- data glove (1)
- database (1)
- dynamics (1)
- energy (1)
- energy saving (1)
- external faults (1)
- facial expression recognition (1)
- fiducial marker (1)
- grasp motions (1)
- grasping (1)
- hybrid dynamics solver (1)
- hybrid system (1)
- image captioning (1)
- infrared pattern (1)
- interaction (1)
- long-distance modeling (1)
- migration (1)
- mobile manipulators (1)
- mobility assistance system (1)
- motion capture (1)
- music analysis (1)
- optical character recognition (1)
- optical tracking (1)
- prehensile motions (1)
- representation learning (1)
- road (1)
- robot control (1)
- robot dynamics (1)
- robotic arm (1)
- robotic evaluation (1)
- robotics (1)
- scene-segmentation (1)
- scenes (1)
- security (1)
- static friction (1)
- task models (1)
- taxonomie (1)
- technology mapping (1)
- text detection (1)
- text localization (1)
- traffic sign detection (1)
- traffic sign localization (1)
- user input (1)
- user interaction (1)
- vocational education (1)
The Global Compact for Safe, Orderly and Regular Migration defines Global Skill Partnerships (GSP) as an innovative means of strengthen skills development among origin countries and countries of destination in mutually beneficial manner. However, GSPs are very limited in number and scope, and empirical analyses of them are, to date, relatively rare. This study helps fill this gap in data by presenting and examining existing GSPs or GSP-like approaches (e.g., transnational training partnerships). The aim of the study is to take stock of the various conceptual discourses on and practical experience with transnational skill partnerships. Using Kosovo as a case study, the study details the structure of such partnerships and the processes they entail. It documents the experience of those involved and catalogues the factors contributing to success. On this basis, the authors propose a means of categorizing the various practices that will help structure the empirical diversity of such approaches and render them conceptually feasible: Transnational Skills and Mobility Partnerships (TSMP).
Neuromorphic computing aims to mimic the computational principles of the brain in silico and has motivated research into event-based vision and spiking neural networks (SNNs). Event cameras (ECs) capture local, independent changes in brightness, and offer superior power consumption, response latencies, and dynamic ranges compared to frame-based cameras. SNNs replicate neuronal dynamics observed in biological neurons and propagate information in sparse sequences of ”spikes”. Apart from biological fidelity, SNNs have demonstrated potential as an alternative to conventional artificial neural networks (ANNs), such as in reducing energy expenditure and inference time in visual classification. Although potentially beneficial for robotics, the novel event-driven and spike-based paradigms remain scarcely explored outside the domain of aerial robots.
To investigate the utility of brain-inspired sensing and data processing in a robotics application, we developed a neuromorphic approach to real-time, online obstacle avoidance on a manipulator with an onboard camera. Our approach adapts high-level trajectory plans with reactive maneuvers by processing emulated event data in a convolutional SNN, decoding neural activations into avoidance motions, and adjusting plans in a dynamic motion primitive formulation. We conducted simulated and real experiments with a Kinova Gen3 arm performing simple reaching tasks involving static and dynamic obstacles. Our implementation was systematically tuned, validated, and tested in sets of distinct task scenarios, and compared to a non-adaptive baseline through formalized quantitative metrics and qualitative criteria.
The neuromorphic implementation facilitated reliable avoidance of imminent collisions in most scenarios, with 84% and 92% median success rates in simulated and real experiments, where the baseline consistently failed. Adapted trajectories were qualitatively similar to baseline trajectories, indicating low impacts on safety, predictability and smoothness criteria. Among notable properties of the SNN were the correlation of processing time with the magnitude of perceived motions (captured in events) and robustness to different event emulation methods. Preliminary tests with a DAVIS346 EC showed similar performance, validating our experimental event emulation method. These results motivate future efforts to incorporate SNN learning, utilize neuromorphic processors, and target other robot tasks to further explore this approach.
The increasing ubiquity of Artificial Intelligence (AI) poses significant political consequences. The rapid proliferation of AI over the past decade has prompted legislators and regulators to attempt to contain AI’s technological consequences. For Germany, relevant design requirements have been expressed by the European Commission’s High-Level Expert Group on Artificial Intelligence (HLEG AI), and, at the national level, by the German government’s Data Ethics Commission (DEK) as well as the German Bundestag’s Commission of Inquiry on Artificial Intelligence (EKKI).
In the field of automatic music generation, one of the greatest challenges is the consistent generation of pieces continuously perceived positively by the majority of the audience since there is no objective method to determine the quality of a musical composition. However, composing principles, which have been refined for millennia, have shaped the core characteristics of today's music. A hybrid music generation system, mlmusic, that incorporates various static, music-theory-based methods, as well as data-driven, subsystems, is implemented to automatically generate pieces considered acceptable by the average listener. Initially, a MIDI dataset, consisting of over 100 hand-picked pieces of various styles and complexities, is analysed using basic music theory principles, and the abstracted information is fed into explicitly constrained LSTM networks. For chord progressions, each individual network is specifically trained on a given sequence length, while phrases are created by consecutively predicting the notes' offset, pitch and duration. Using these outputs as a composition's foundation, additional musical elements, along with constrained recurrent rhythmic and tonal patterns, are statically generated. Although no survey regarding the pieces' reception could be carried out, the successful generation of numerous compositions of varying complexities suggests that the integration of these fundamentally distinctive approaches might lead to success in other branches.
Effective Neighborhood Feature Exploitation in Graph CNNs for Point Cloud Object-Part Segmentation
(2022)
Part segmentation is the task of semantic segmentation applied on objects and carries a wide range of applications from robotic manipulation to medical imaging. This work deals with the problem of part segmentation on raw, unordered point clouds of 3D objects. While pioneering works on deep learning for point clouds typically ignore taking advantage of local geometric structure around individual points, the subsequent methods proposed to extract features by exploiting local geometry have not yielded significant improvements either. In order to investigate further, a graph convolutional network (GCN) is used in this work in an attempt to increase the effectiveness of such neighborhood feature exploitation approaches. Most of the previous works also focus only on segmenting complete point cloud data. Considering the impracticality of such approaches, taking into consideration the real world scenarios where complete point clouds are scarcely available, this work proposes approaches to deal with partial point cloud segmentation.
In the attempt to better capture neighborhood features, this work proposes a novel method to learn regional part descriptors which guide and refine the segmentation predictions. The proposed approach helps the network achieve state-of-the-art performance of 86.4% mIoU on the ShapeNetPart dataset for methods which do not use any preprocessing techniques or voting strategies. In order to better deal with partial point clouds, this work also proposes new strategies to train and test on partial data. While achieving significant improvements compared to the baseline performance, the problem of partial point cloud segmentation is also viewed through an alternate lens of semantic shape completion.
Semantic shape completion networks not only help deal with partial point cloud segmentation but also enrich the information captured by the system by predicting complete point clouds with corresponding semantic labels for each point. To this end, a new network architecture for semantic shape completion is also proposed based on point completion network (PCN) which takes advantage of a graph convolution based hierarchical decoder for completion as well as segmentation. In addition to predicting complete point clouds, results indicate that the network is capable of reaching within a margin of 5% to the mIoU performance of dedicated segmentation networks for partial point cloud segmentation.
Recent advances in Natural Language Processing have substantially improved contextualized representations of language. However, the inclusion of factual knowledge, particularly in the biomedical domain, remains challenging. Hence, many Language Models (LMs) are extended by Knowledge Graphs (KGs), but most approaches require entity linking (i.e., explicit alignment between text and KG entities). Inspired by single-stream multimodal Transformers operating on text, image and video data, this thesis proposes the Sophisticated Transformer trained on biomedical text and Knowledge Graphs (STonKGs). STonKGs incorporates a novel multimodal architecture based on a cross encoder that uses the attention mechanism on a concatenation of input sequences derived from text and KG triples, respectively. Over 13 million so-called text-triple pairs, coming from PubMed and assembled using the Integrated Network and Dynamical Reasoning Assembler (INDRA), were used in an unsupervised pre-training procedure to learn representations of biomedical knowledge in STonKGs. By comparing STonKGs to an NLP- and a KG-baseline (operating on either text or KG data) on a benchmark consisting of eight fine-tuning tasks, the proposed knowledge integration method applied in STonKGs was empirically validated. Specifically, on tasks with a comparatively small dataset size and a larger number of classes, STonKGs resulted in considerable performance gains, beating the F1-score of the best baseline by up to 0.083. Both the source code as well as the code used to implement STonKGs are made publicly available so that the proposed method of this thesis can be extended to many other biomedical applications.
High-dimensional and multi-variate data from dynamical systems such as turbulent flows and wind turbines can be analyzed with deep learning due to its capacity to learn representations in lower-dimensional manifolds. Two challenges of interest arise from data generated from these systems, namely, how to anticipate wind turbine failures and how to better understand air flow through car ventilation systems. There are deep neural network architectures that can project data into a lower-dimensional space with the goal of identifying and understanding patterns that are not distinguishable in the original dimensional space. Learning data representations in lower dimensions via non-linear mappings allows one to perform data compression, data clustering (for anomaly detection), data reconstruction and synthetic data generation.
In this work, we explore the potential that variational autoencoders (VAE) have to learn low-dimensional data representations in order to tackle the problems posed by the two dynamical systems mentioned above. A VAE is a neural network architecture that combines the mechanisms of the standard autoencoder and variational bayes. The goal here is to train a neural network to minimize a loss function defined by a reconstruction term together with a variational term defined as a Kulback-Leibler (KL) divergence.
The report discusses the results obtained for the two different data domains: wind turbine time series and turbulence data from computational fluid dynamics (CFD) simulations.
We report on the reconstruction, clustering and unsupervised anomaly detection of wind turbine multi-variate time series data using a variant of a VAE called Variational Recurrent Autoencoder (VRAE). We trained a VRAE to cluster normal and abnormal wind turbine series (two class problem) as well as normal and multiple abnormal series (multi-class problem). We found that the model is capable of distinguishing between normal and abnormal cases by reducing the dimensionality of the input data and projecting it to two dimensions using techniques such as Principal Component Analysis (PCA) and t-distributed stochastic neighbor embedding (t-SNE). A set of anomaly scoring methods is applied on top of these latent vectors in order to compute unsupervised clustering. We have achieved an accuracy of up to 96% with the KM eans + + algorithm.
We also report the data reconstruction and generation results of two dimensional turbulence slices corresponding to CFD simulation of a HVAC air duct. For this, we have trained a Convolutional Variational Autoencoder (CVAE). We have found that the model is capable of reconstructing laminar flows up to a certain degree of resolution as well generating synthetic turbulence data from the learned latent distribution.
A Comparative Study of Uncertainty Estimation Methods in Deep Learning Based Classification Models
(2020)
Deep learning models produce overconfident predictions even for misclassified data. This work aims to improve the safety guarantees of software-intensive systems that use deep learning based classification models for decision making by performing comparative evaluation of different uncertainty estimation methods to identify possible misclassifications.
In this work, uncertainty estimation methods applicable to deep learning models are reviewed and those which can be seamlessly integrated to existing deployed deep learning architectures are selected for evaluation. The different uncertainty estimation methods, deep ensembles, test-time data augmentation and Monte Carlo dropout with its variants, are empirically evaluated on two standard datasets (CIFAR-10 and CIFAR-100) and two custom classification datasets (optical inspection and RoboCup@Work dataset). A relative ranking between the methods is provided by evaluating the deep learning classifiers on various aspects such as uncertainty quality, classifier performance and calibration. Standard metrics like entropy, cross-entropy, mutual information, and variance, combined with a rank histogram based method to identify uncertain predictions by thresholding on these metrics, are used to evaluate uncertainty quality.
The results indicate that Monte Carlo dropout combined with test-time data augmentation outperforms all other methods by identifying more than 95% of the misclassifications and representing uncertainty in the highest number of samples in the test set. It also yields a better classifier performance and calibration in terms of higher accuracy and lower Expected Calibration Error (ECE), respectively. A python based uncertainty estimation library for training and real-time uncertainty estimation of deep learning based classification models is also developed.
Human and robot tasks in household environments include actions such as carrying an object, cleaning a surface, etc. These tasks are performed by means of dexterous manipulation, and for humans, they are straightforward to accomplish. Moreover, humans perform these actions with reasonable accuracy and precision but with much less energy and stress on the actuators (muscles) than the robots do. The high agility in controlling their forces and motions is actually due to "laziness", i.e. humans exploit the existing natural forces and constraints to execute the tasks.
The above-mentioned properties of the human lazy strategy motivate us to relax the problem of controlling robot motions and forces, and solve it with the help of the environment. Therefore, in this work, we developed a lazy control strategy, i.e. task specification models and control architectures that relax several aspects of robot control by exploiting prior knowledge about the task and environment. The developed control strategy is realized in four different robotics use cases. In this work, the Popov-Vereshchagin hybrid dynamics solver is used as one of the building blocks in the proposed control architectures. An extension of the solver’s interface with the artificial Cartesian force and feed-forward joint torque task-drivers is proposed in this thesis.
To validate the proposed lazy control approach, an experimental evaluation was performed in a simulation environment and on a real robot platform.
This work provides a short but technical introduction to the main building blocks of a blockchain. It argues that a blockchain is not a revolutionary technology but rather a clever combination of three fields: cryptography, decentralization and game theory. In addition, it summaries the differences between a public, private and federate blockchain model and the two prominent consensus mechanism Proof-of-Work (POW) and Proof-of-Stake (POS).