Refine
Departments, institutes and facilities
- Institut für Technik, Ressourcenschonung und Energieeffizienz (TREE) (481) (remove)
Document Type
- Conference Object (216)
- Article (178)
- Part of a Book (26)
- Preprint (17)
- Report (11)
- Doctoral Thesis (8)
- Contribution to a Periodical (6)
- Research Data (6)
- Book (monograph, edited volume) (4)
- Part of Periodical (4)
- Lecture (2)
- Other (1)
- Patent (1)
- Working Paper (1)
Year of publication
Keywords
- lignin (7)
- Quality diversity (6)
- West Africa (6)
- advanced applications (5)
- modeling of complex systems (5)
- stem cells (5)
- Hydrogen storage (4)
- Lattice Boltzmann Method (4)
- Lignin (4)
- additive (4)
Das sogenannte „Deutschlandstipendium“ ist 2010 ins Leben gerufen worden. Gemäß den gesetzlichen Vorgaben sollen die Stipendien nach Begabung und Leistung vergeben werden. Darüber hinaus sollen auch gesellschaftliches Engagement oder besondere soziale, familiäre oder persönliche Umstände berücksichtigt werden. Bei der Finanzierung sind die Hochschulen zunächst auf das Einwerben privater Fördermittel angewiesen, die von Bund und Land um denselben Betrag aufgestockt werden. Die privaten Mittelgeber können für die von ihnen anteilig finanzierten Stipendien festlegen, aus welchen Studiengängen ihre Stipendiaten ausgewählt werden sollen. Die Hochschulen haben jedoch darauf zu achten, dass ein Drittel aller zu vergebenden Stipendien ohne eine entsprechende Zweckbindung vergeben werden. Einen direkten Einfluss auf die Auswahl einzelner Kandidaten dürfen die Förderer nicht haben. Vor diesem Hintergrund sind die Hochschulen angehalten, Anreize für private Förderer zu schaffen und parallel Bewerbungs- und Auswahlverfahren zu konzipieren, die die genannten gesetzlichen Vorgaben einhalten. Dadurch entsteht bei den Hochschulen ein erheblicher Verwaltungsaufwand. Zu dessen Reduzierung wird in diesem Artikel ein transparenter, nachvollziehbarer, zeit- und kostensparender Prozess durch einen programmierten Workflow beschrieben.
Sie sind im Bereich Qualitätsmanagement tätig und haben die Aufgabe bekommen, ein Problem systematisch zu untersuchen und methodisch zu lösen? Sie haben zu viele Aufgaben und wissen nicht, wie Sie diese priorisieren sollen? Oder haben Sie zu begrenzte Ressourcen, um alle Reklamationen gleichzeitig bearbeiten zu können? Oder wissen nicht, wie Sie einen bestimmten Prozess in seinen Grenzen zielführend verbessern können?
Animal models are often needed in cancer research but some research questions may be answered with other models, e.g., 3D replicas of patient-specific data, as these mirror the anatomy in more detail. We, therefore, developed a simple eight-step process to fabricate a 3D replica from computer tomography (CT) data using solely open access software and described the method in detail. For evaluation, we performed experiments regarding endoscopic tumor treatment with magnetic nanoparticles by magnetic hyperthermia and local drug release. For this, the magnetic nanoparticles need to be accumulated at the tumor site via a magnetic field trap. Using the developed eight-step process, we printed a replica of a locally advanced pancreatic cancer and used it to find the best position for the magnetic field trap. In addition, we described a method to hold these magnetic field traps stably in place. The results are highly important for the development of endoscopic tumor treatment with magnetic nanoparticles as the handling and the stable positioning of the magnetic field trap at the stomach wall in close proximity to the pancreatic tumor could be defined and practiced. Finally, the detailed description of the workflow and use of open access software allows for a wide range of possible uses.
During exercise, heart rate has proven to be a good measure in planning workouts. It is not only simple to measure but also well understood and has been used for many years for workout planning. To use heart rate to control physical exercise, a model which predicts future heart rate dependent on a given strain can be utilized. In this paper, we present a mathematical model based on convolution for predicting the heart rate response to strain with four physiologically explainable parameters. This model is based on the general idea of the Fitness-Fatigue model for performance analysis, but is revised here for heart rate analysis. Comparisons show that the Convolution model can compete with other known heart rate models. Furthermore, this new model can be improved by reducing the number of parameters. The remaining parameter seems to be a promising indicator of the actual subject’s fitness.
Computers can help us to trigger our intuition about how to solve a problem. But how does a computer take into account what a user wants and update these triggers? User preferences are hard to model as they are by nature vague, depend on the user’s background and are not always deterministic, changing depending on the context and process under which they were established. We pose that the process of preference discovery should be the object of interest in computer aided design or ideation. The process should be transparent, informative, interactive and intuitive. We formulate Hyper-Pref, a cyclic co-creative process between human and computer, which triggers the user’s intuition about what is possible and is updated according to what the user wants based on their decisions. We combine quality diversity algorithms, a divergent optimization method that can produce many, diverse solutions, with variational autoencoders to both model that diversity as well as the user’s preferences, discovering the preference hypervolume within large search spaces.
Electrical signal transmission in power electronic devices takes place through high-purity aluminum bonding wires. Cyclic mechanical and thermal stresses during operation lead to fatigue loads, resulting in premature failure of the wires, which cannot be reliably predicted. The following work presents two fatigue lifetime models calibrated and validated based on experimental fatigue results of an aluminum bonding wire and subsequently transferred and applied to other wire types. The lifetime modeling of Wöhler curves for different load ratios shows good but limited applicability for the linear model. The model can only be applied above 10,000 cycles and within the investigated load range of R = 0.1 to R = 0.7. The nonlinear model shows very good agreement between model prediction and experimental results over the entire investigated cycle range. Furthermore, the predicted Smith diagram is not only consistent in the investigated load range but also in the extrapolated load range from R = −1.0 to R = 0.8. A transfer of both model approaches to other wire types by using their tensile strengths can be implemented as well, although the nonlinear model is more suitable since it covers the entire load and cycle range.
Integrating physical simulation data into data ecosystems challenges the compatibility and interoperability of data management tools. Semantic web technologies and relational databases mostly use other data types, such as measurement or manufacturing design data. Standardizing simulation data storage and harmonizing the data structures with other domains is still a challenge, as current standards such as the ISO standard STEP (ISO 10303 ”Standard for the Exchange of Product model data”) fail to bridge the gap between design and simulation data. This challenge requires new methods, such as ontologies, to rethink simulation results integration. This research describes a new software architecture and application methodology based on the industrial standard ”Virtual Material Modelling in Manufacturing” (VMAP). The architecture integrates large quantities of structured simulation data and their analyses into a semantic data structure. It is capable of providing data permeability from the global digital twin level to the detailed numerical values of data entries and even new key indicators in a three-step approach: It represents a file as an instance in a knowledge graph, queries the file’s metadata, and finds a semantically represented process that enables new metadata to be created and instantiated.