Refine
Departments, institutes and facilities
- Fachbereich Informatik (48)
- Institut für Technik, Ressourcenschonung und Energieeffizienz (TREE) (42)
- Fachbereich Ingenieurwissenschaften und Kommunikation (9)
- Internationales Zentrum für Nachhaltige Entwicklung (IZNE) (3)
- Institut für KI und Autonome Systeme (A2S) (1)
- Institut für Sicherheitsforschung (ISF) (1)
- Institute of Visual Computing (IVC) (1)
Document Type
- Conference Object (39)
- Article (11)
- Preprint (5)
- Report (5)
- Book (monograph, edited volume) (1)
- Part of a Book (1)
- Diploma Thesis (1)
- Doctoral Thesis (1)
Year of publication
Keywords
- Quality Diversity (4)
- Quality diversity (4)
- Bayesian optimization (3)
- MAP-Elites (3)
- Aerodynamics (2)
- Autoencoder (2)
- Evolutionary Computation (2)
- Evolutionary computation (2)
- Heart Rate Prediction (2)
- Lattice Boltzmann Method (2)
AErOmAt Abschlussbericht
(2020)
Das Projekt AErOmAt hatte zum Ziel, neue Methoden zu entwickeln, um einen erheblichen Teil aerodynamischer Simulationen bei rechenaufwändigen Optimierungsdomänen einzusparen. Die Hochschule Bonn-Rhein-Sieg (H-BRS) hat auf diesem Weg einen gesellschaftlich relevanten und gleichzeitig wirtschaftlich verwertbaren Beitrag zur Energieeffizienzforschung geleistet. Das Projekt führte außerdem zu einer schnelleren Integration der neuberufenen Antragsteller in die vorhandenen Forschungsstrukturen.
In complex, expensive optimization domains we often narrowly focus on finding high performing solutions, instead of expanding our understanding of the domain itself. But what if we could quickly understand the complex behaviors that can emerge in said domains instead? We introduce surrogate-assisted phenotypic niching, a quality diversity algorithm which allows to discover a large, diverse set of behaviors by using computationally expensive phenotypic features. In this work we discover the types of air flow in a 2D fluid dynamics optimization problem. A fast GPU-based fluid dynamics solver is used in conjunction with surrogate models to accurately predict fluid characteristics from the shapes that produce the air flow. We show that these features can be modeled in a data-driven way while sampling to improve performance, rather than explicitly sampling to improve feature models. Our method can reduce the need to run an infeasibly large set of simulations while still being able to design a large diversity of air flows and the shapes that cause them. Discovering diversity of behaviors helps engineers to better understand expensive domains and their solutions.
Abschlussbericht zum BMBF-Fördervorhaben Enabling Infrastructure for HPC-Applications (EI-HPC)
(2020)
The MAP-Elites algorithm produces a set of high-performing solutions that vary according to features defined by the user. This technique has the potential to be a powerful tool for design space exploration, but is limited by the need for numerous evaluations. The Surrogate-Assisted Illumination algorithm (SAIL), introduced here, integrates approximative models and intelligent sampling of the objective function to minimize the number of evaluations required by MAP-Elites.
The ability of SAIL to efficiently produce both accurate models and diverse high performing solutions is illustrated on a 2D airfoil design problem. The search space is divided into bins, each holding a design with a different combination of features. In each bin SAIL produces a better performing solution than MAP-Elites, and requires several orders of magnitude fewer evaluations. The CMA-ES algorithm was used to produce an optimal design in each bin: with the same number of evaluations required by CMA-ES to find a near-optimal solution in a single bin, SAIL finds solutions of similar quality in every bin.
Surrogate-assistance approaches have long been used in computationally expensive domains to improve the data-efficiency of optimization algorithms. Neuroevolution, however, has so far resisted the application of these techniques because it requires the surrogate model to make fitness predictions based on variable topologies, instead of a vector of parameters. Our main insight is that we can sidestep this problem by using kernel-based surrogate models, which require only the definition of a distance measure between individuals. Our second insight is that the well-established Neuroevolution of Augmenting Topologies (NEAT) algorithm provides a computationally efficient distance measure between dissimilar networks in the form of "compatibility distance", initially designed to maintain topological diversity. Combining these two ideas, we introduce a surrogate-assisted neuroevolution algorithm that combines NEAT and a surrogate model built using a compatibility distance kernel. We demonstrate the data-efficiency of this new algorithm on the low dimensional cart-pole swing-up problem, as well as the higher dimensional half-cheetah running task. In both tasks the surrogate-assisted variant achieves the same or better results with several times fewer function evaluations as the original NEAT.
In der vorliegenden Arbeit werden Verfahren vorgestellt, die geeignet sind, Modelle des menschlichen kardiovaskulären Systems an individuelle Kreislaufreaktionen anzupassen. Allgemeine Kreislaufmodelle des menschlichen kardiovaskulären Systems sind in der Regel nichtlineare Differentialgleichungssysteme, für die es keine effizienten Optimierungsverfahren gibt. Durch die Einschränkung auf relevante Aspekte (bzgl. der Individualisierungsaufgabe) wird ein solches Modell auf Modelle einfacherer Struktur projiziert, die eine Approximation durch Funktionsapproximatoren erlauben, für welche wiederum effiziente Optimierungsalgorithmen existieren. In Abhängigkeit von der Struktur der Individualisierungsaufgabe kommt zusätzlich ein modifiziertes BFGS-Verfahren zum Einsatz, das Approximationen solcher Modellaspekte verwendet um die Konvergenz der Modellindividualisierung zu verbessern.
We present a model checking algorithm for ∀CTL (and full CTL) which uses an iterative abstraction refinement strategy.
It terminates at least for all transition systems M that have a finite simulation or bisimulation quotient. In contrast to other abstraction refinement algorithms, we always work with abstract models whose sizes depend only on the length of the formula θ (but not on the size of the system, which might be infinite).
A new method for design space exploration and optimization, Surrogate-Assisted Illumination (SAIL), is presented. Inspired by robotics techniques designed to produce diverse repertoires of behaviors for use in damage recovery, SAIL produces diverse designs that vary according to features specified by the designer. By producing high-performing designs with varied combinations of user-defined features a map of the design space is created. This map illuminates the relationship between the chosen features and performance, and can aid designers in identifying promising design concepts. SAIL is designed for use with compu-tationally expensive design problems, such as fluid or structural dynamics, and integrates approximative models and intelligent sampling of the objective function to minimize the number of function evaluations required. On a 2D airfoil optimization problem SAIL is shown to produce hundreds of diverse designs which perform competitively with those found by state-of-the-art black box optimization. Its capabilities are further illustrated in a more expensive 3D aerodynamic optimization task.
The MAP-Elites algorithm produces a set of high-performing solutions that vary according to features defined by the user. This technique to 'illuminate' the problem space through the lens of chosen features has the potential to be a powerful tool for exploring design spaces, but is limited by the need for numerous evaluations. The Surrogate-Assisted Illumination (SAIL) algorithm, introduced here, integrates approximative models and intelligent sampling of the objective function to minimize the number of evaluations required by MAP-Elites.
The ability of SAIL to efficiently produce both accurate models and diverse high-performing solutions is illustrated on a 2D airfoil design problem. The search space is divided into bins, each holding a design with a different combination of features. In each bin SAIL produces a better performing solution than MAP-Elites, and requires several orders of magnitude fewer evaluations. The CMA-ES algorithm was used to produce an optimal design in each bin: with the same number of evaluations required by CMA-ES to find a near-optimal solution in a single bin, SAIL finds solutions of similar quality in every bin.
The encoding of solutions in black-box optimization is a delicate, handcrafted balance between expressiveness and domain knowledge between exploring a wide variety of solutions, and ensuring that those solutions are useful. Our main insight is that this process can be automated by generating a dataset of high-performing solutions with a quality diversity algorithm (here, MAP-Elites), then learning a representation with a generative model (here, a Varia-tional Autoencoder) from that dataset. Our second insight is that this representation can be used to scale quality diversity optimization to higher dimensions-but only if we carefully mix solutions generated with the learned representation and those generated with traditional variation operators. We demonstrate these capabilities by learning an low-dimensional encoding for the inverse kinemat-ics of a thousand joint planar arm. The results show that learned representations make it possible to solve high-dimensional problems with orders of magnitude fewer evaluations than the standard MAP-Elites, and that, once solved, the produced encoding can be used for rapid optimization of novel, but similar, tasks. The presented techniques not only scale up quality diversity algorithms to high dimensions, but show that black-box optimization encodings can be automatically learned, rather than hand designed.
This paper describes the development of a Pedelec controller whose performance level (PL) conforms to European standard on safety of machinery [9] and whose soft- ware is verified to conform to EPAC standard [6] by means of a software verification technique called model checking. In compliance with the standard [9] the hardware needs to implement the required properties corresponding to categories “C” and “D”. The latter is used if the breaks are not able to bring the velomobile with a broken motor controller to a full stop. Therefore the controller needs to implement a test unit, which verifies the functionality of the components and, in case of an emergency, shuts the whole hardware down to prevent injuries of the cyclist. The MTTFd can be measured through a failure graph, which is the result of a FMEA analysis, and can be used to proof that the Pedelec controller meets the regulations of the system specification. The analysis of the system in compliance with [9] usually treats the software as a black box thus ignoring its inner workings and validating its correctness by means of testing. In this paper we present a temporal logic specification according to [6], based on which the software for the Pedelec controller is implemented, and verify instead of only testing its functionality. By means of model checking [1] we proof that the software fulfills all requirements which are regulated by its specification.
During exercise, heart rate has proven to be a good measure in planning workouts. It is not only simple to measure but also well understood and has been used for many years for workout planning. To use heart rate to control physical exercise, a model which predicts future heart rate dependent on a given strain can be utilized. In this paper, we present a mathematical model based on convolution for predicting the heart rate response to strain with four physiologically explainable parameters. This model is based on the general idea of the Fitness-Fatigue model for performance analysis, but is revised here for heart rate analysis. Comparisons show that the Convolution model can compete with other known heart rate models. Furthermore, this new model can be improved by reducing the number of parameters. The remaining parameter seems to be a promising indicator of the actual subject’s fitness.
Analyzing training performance in sport is usually based on standardized test protocols and needs laboratory equipment, e.g., for measuring blood lactate concentration or other physiological body parameters. Avoiding special equipment and standardized test protocols, we show that it is possible to reach a quality of performance simulation comparable to the results of laboratory studies using training models with nothing but training data. For this purpose, we introduce a fitting concept for a performance model that takes the peculiarities of using training data for the task of performance diagnostics into account. With a specific way of data preprocessing, accuracy of laboratory studies can be achieved for about 50% of the tested subjects, while lower correlation of the other 50% can be explained.
The Fitness-Fatigue model (Calvert et al. 1976) is widely used for performance analysis. This antagonistic model is based on a fitness-term, a fatigue-term, and an initial basic level of performance. Instead of generic parameter values, individualizing the model needs a fitting of parameters. With fitted parameters, the model adapts to account for individual responses to strain. Even though in most cases fitting of recorded training data shows useful results, without modification the model cannot be simply used for prediction.