Refine
H-BRS Bibliography
- yes (73)
Departments, institutes and facilities
- Fachbereich Angewandte Naturwissenschaften (21)
- Fachbereich Informatik (18)
- Fachbereich Ingenieurwissenschaften und Kommunikation (12)
- Institut für Technik, Ressourcenschonung und Energieeffizienz (TREE) (12)
- Fachbereich Wirtschaftswissenschaften (11)
- Institut für funktionale Gen-Analytik (IFGA) (9)
- Internationales Zentrum für Nachhaltige Entwicklung (IZNE) (6)
- Institut für Sicherheitsforschung (ISF) (4)
- Institute of Visual Computing (IVC) (4)
- Stabsstelle Kommunikation und Marketing (3)
Document Type
- Article (44)
- Report (5)
- Conference Object (4)
- Master's Thesis (4)
- Bachelor Thesis (3)
- Part of a Book (3)
- Part of Periodical (3)
- Preprint (3)
- Working Paper (3)
- Book (monograph, edited volume) (1)
Year of publication
- 2020 (73) (remove)
Has Fulltext
- yes (73) (remove)
Keywords
- Digitalisierung (2)
- Hochschule Bonn-Rhein-Sieg (2)
- Inborn error of metabolism (2)
- Ketone body (2)
- Metabolic acidosis (2)
- Organic aciduria (2)
- Usable Security (2)
- bank customers (2)
- ACAT1 (1)
- ADP release (1)
"Durchdringen - Klarheit schaffen, Barrieren überwinden, Gehör finden" lautet diesmal das Motto des Jahresberichts. Er zeigt, wie die Hochschule nach Antworten auf die vielschichtigen, komplexen Fragen der Zeit sucht. Ob Digitalisierung, Klimawandel oder gesellschaftliche Verantwortung - Wissenschaftlerinnen und Wissenschaftler durchdringen ihre Themengebiete, und sie müssen am Ende mit ihren Erkenntnissen Gehör finden.
"Diffusion - create clarity, overcome barriers, be heard" is the title of this year's annual report. It shows how the university is searching for answers to the multilayered, complex questions of our time. Whether digitalisation, climate change or social responsibility - scientists are getting through their subject areas and in the end they have to make their findings heard.
An essential measure of autonomy in service robots designed to assist humans is adaptivity to the various contexts of human-oriented tasks. These robots may have to frequently execute the same action, but subject to subtle variations in task parameters that determine optimal behaviour. Such actions are traditionally executed by robots using pre-determined, generic motions, but a better approach could utilize robot arm maneuverability to learn and execute different trajectories that work best in each context.
In this project, we explore a robot skill acquisition procedure that allows incorporating contextual knowledge, adjusting executions according to context, and improvement through experience, as a step towards more adaptive service robots. We propose an apprenticeship learning approach to achieving context-aware action generalisation on the task of robot-to-human object hand-over. The procedure combines learning from demonstration, with which a robot learns to imitate a demonstrator’s execution of the task, and a reinforcement learning strategy, which enables subsequent experiential learning of contextualized policies, guided by information about context that is integrated into the learning process. By extending the initial, static hand-over policy to a contextually adaptive one, the robot derives and executes variants of the demonstrated action that most appropriately suit the current context. We use dynamic movement primitives (DMPs) as compact motion representations, and a model-based Contextual Relative Entropy Policy Search (C-REPS) algorithm for learning policies that can specify hand-over position, trajectory shape, and execution speed, conditioned on context variables. Policies are learned using simulated task executions, before transferring them to the robot and evaluating emergent behaviours.
We demonstrate the algorithm’s ability to learn context-dependent hand-over positions, and new trajectories, guided by suitable reward functions, and show that the current DMP implementation limits learning context-dependent execution speeds. We additionally conduct a user study involving participants assuming different postures and receiving an object from the robot, which executes hand-overs by either exclusively imitating a demonstrated motion, or selecting hand-over positions based on learned contextual policies and adapting its motion accordingly. The results confirm the hypothesized improvements in the robot’s perceived behaviour when it is context-aware and adaptive, and provide useful insights that can inform future developments.
YAWL (Yet Another Workflow Language) is an open source Business Process Management System, first released in 2003. YAWL grew out of a university research environment to become a unique system that has been deployed worldwide as a laboratory environment for research in Business Process Management and as a productive system in other scientific domains.
Multiwalled carbon nanotubes (MWCNTs) were easily and efficiently functionalised with highly cross-linked polyamines. The radical polymerisation of two bis-vinylimidazolium salts in the presence of pristine MWCNTs and azobisisobutyronitrile (AIBN) as a radical initiator led to the formation of materials with a high functionalisation degree. The subsequent treatment with sodium borohydride gave rise to the reduction of imidazolium moieties with the concomitant formation of secondary and tertiary amino groups. The obtained materials were characterised by thermogravimetric analysis (TGA), elemental analysis, solid state 13C-NMR, Fourier-transform infrared spectroscopy (FT-IR), transmission electron microscopy (TEM), potentiometric titration, and temperature programmed desorption of carbon dioxide (CO2-TPD). One of the prepared materials was tested as a heterogeneous base catalyst in C–C bond forming reactions such as the Knoevenagel condensation and Henry reaction. Furthermore, two examples concerning a sequential one-pot approach involving two consecutive reactions, namely Knoevenagel and Michael reactions, were reported.
The ongoing coronavirus disease 2019 (COVID-19) pandemic threatens global health thereby causing unprecedented social, economic, and political disruptions. One way to prevent such a pandemic is through interventions at the human-animal-environment interface by using an integrated One Health (OH) approach. This systematic literature review documented the three coronavirus outbreaks, i.e. SARS, MERS, COVID-19, to evaluate the evolution of the OH approach, including the identification of key OH actions taken for prevention, response, and control.
The OH understandings identified were categorized into three distinct patterns: institutional coordination and collaboration, OH in action/implementation, and extended OH (i.e. a clear involvement of the environmental domain). Across all studies, OH was most often framed as OH in action/implementation and least often in its extended meaning. Utilizing OH as institutional coordination and collaboration and the extended OH both increased over time. OH actions were classified into twelve sub-groups and further categorized as classical OH actions (i.e. at the human-animal interface), classical OH actions with outcomes to the environment, and extended OH actions.
The majority of studies focused on human-animal interaction, giving less attention to the natural and built environment. Different understandings of the OH approach in practice and several practical limitations might hinder current efforts to achieve the operationalization of OH by combining institutional coordination and collaboration with specific OH actions. The actions identified here are a valuable starting point for evaluating the stage of OH development in different settings. This study showed that by moving beyond the classical OH approach and its actions towards a more extended understanding, OH can unfold its entire capacity thereby improving preparedness and mitigating the impacts of the next outbreak.
Studi ini bertujuan untuk memvalidasi perangkat penilaian efikasi diri yang berkaitan dengan kesehatan kerja yang dikembangkan pada tahap studi pendahuluan. Skala Efikasi Diri untuk Kesehatan Kerja (SEDKK) berlandaskan konsep efikasi diri pada teori kognitif sosial yang mengukur empat faktor yang berpengaruh pada kesehatan setiap individu yang bekerja, seperti: perilaku makan dan minum, tidur, keamanan dan kesehatan kerja, serta kegiatan pemulihan dari stres bekerja. Hasil analisis faktor eksploratori menunjukan bahwa ada empat faktor yang terefleksikan dari butir-butir SEDKK. Validitas konstruk SEDKK dapat dibuktikan dengan korelasi positif antara SEDKK dan skala Efikasi Diri Umum yang sangat signifikan. Pengujian validitas kriteria dapat ditelusuri melalui efek SEDKK terhadap kondisi kesehatan umum, kepuasan akan kesehatan pribadi, keseimbangan kehidupan kerja/KKK (work life balance), perilaku sehat, dan perilaku berisiko. Namun demikian, asumsi mengenai reliabilitas tes berulang (test-retest) pada penelitian ini ditolak. Implikasi dan saran-saran untuk penelitian selanjutnya didiskusikan pada artikel ini.
Any political phenomenon can only be properly understood in its broader con-text. Questions of international cooperation are thus necessarily framed by his-torical processes and relations of power. We therefore start our first discussion with an examination of the global ‘status quo’ and embed the topic of this pub-lication, ODA graduation, into the shifting world order, analysing current roles and settings in international relations and identifying changes in positions, sta-tus and categories. What are the overarching issues determining world politics and who are the old and the new actors driving them? What is the impact of these global shifts on international cooperation, especially development coop-eration? Of what relevance are roles, status and categories and what is the im-pact of changes in positions and relations? What challenges face multilateralism and what ways exist to maintain and renew strategic partnerships and shared values?
Am Beispiel einer jahrelang in Präsenz gelehrten Veranstaltung mit Vorlesungen, Übungen und Laborpraktika wird gezeigt, wie die Vermittlung prüfungsrelevanter Kompetenzen auch „online“ gelang. Das passende „Setting“ des Lehr- und Lernprozesses unter Beachtung von Handlungsempfehlungen ist auch für die Zukunft relevant.
The ability to finely segment different instances of various objects in an environment forms a critical tool in the perception tool-box of any autonomous agent. Traditionally instance segmentation is treated as a multi-label pixel-wise classification problem. This formulation has resulted in networks that are capable of producing high-quality instance masks but are extremely slow for real-world usage, especially on platforms with limited computational capabilities. This thesis investigates an alternate regression-based formulation of instance segmentation to achieve a good trade-off between mask precision and run-time. Particularly the instance masks are parameterized and a CNN is trained to regress to these parameters, analogous to bounding box regression performed by an object detection network.
In this investigation, the instance segmentation masks in the Cityscape dataset are approximated using irregular octagons and an existing object detector network (i.e., SqueezeDet) is modified to regresses to the parameters of these octagonal approximations. The resulting network is referred to as SqueezeDetOcta. At the image boundaries, object instances are only partially visible. Due to the convolutional nature of most object detection networks, special handling of the boundary adhering object instances is warranted. However, the current object detection techniques seem to be unaffected by this and handle all the object instances alike. To this end, this work proposes selectively learning only partial, untainted parameters of the bounding box approximation of the boundary adhering object instances. Anchor-based object detection networks like SqueezeDet and YOLOv2 have a discrepancy between the ground-truth encoding/decoding scheme and the coordinate space used for clustering, to generate the prior anchor shapes. To resolve this disagreement, this work proposes clustering in a space defined by two coordinate axes representing the natural log transformations of the width and height of the ground-truth bounding boxes.
When both SqueezeDet and SqueezeDetOcta were trained from scratch, SqueezeDetOcta lagged behind the SqueezeDet network by a massive ≈ 6.19 mAP. Further analysis revealed that the sparsity of the annotated data was the reason for this lackluster performance of the SqueezeDetOcta network. To mitigate this issue transfer-learning was used to fine-tune the SqueezeDetOcta network starting from the trained weights of the SqueezeDet network. When all the layers of the SqueezeDetOcta were fine-tuned, it outperformed the SqueezeDet network paired with logarithmically extracted anchors by ≈ 0.77 mAP. In addition to this, the forward pass latencies of both SqueezeDet and SqueezeDetOcta are close to ≈ 19ms. Boundary adhesion considerations, during training, resulted in an improvement of ≈ 2.62 mAP of the baseline SqueezeDet network. A SqueezeDet network paired with logarithmically extracted anchors improved the performance of the baseline SqueezeDet network by ≈ 1.85 mAP.
In summary, this work demonstrates that if given sufficient fine instance annotated data, an existing object detection network can be modified to predict much finer approximations (i.e., irregular octagons) of the instance annotations, whilst having the same forward pass latency as that of the bounding box predicting network. The results justify the merits of logarithmically extracted anchors to boost the performance of any anchor-based object detection network. The results also showed that the special handling of image boundary adhering object instances produces more performant object detectors.
Modern Monte-Carlo-based rendering systems still suffer from the computational complexity involved in the generation of noise-free images, making it challenging to synthesize interactive previews. We present a framework suited for rendering such previews of static scenes using a caching technique that builds upon a linkless octree. Our approach allows for memory-efficient storage and constant-time lookup to cache diffuse illumination at multiple hitpoints along the traced paths. Non-diffuse surfaces are dealt with in a hybrid way in order to reconstruct view-dependent illumination while maintaining interactive frame rates. By evaluating the visual fidelity against ground truth sequences and by benchmarking, we show that our approach compares well to low-noise path-traced results, but with a greatly reduced computational complexity, allowing for interactive frame rates. This way, our caching technique provides a useful tool for global illumination previews and multi-view rendering.
Unsachgemäß entsorgte Zigarettenkippen stellen aufgrund der in ihnen enthaltenen Giftstoffe ein relevantes, ökologisches Problem dar. Diese Forschungsarbeit untersucht den Einsatz von Nudging zur Bekämpfung der Problematik. In einer quantiativen Online-Befragung wurden zunächst die Gründe für das umweltschädliche Verhalten untersucht (N = 96). Hierbei konnte die Gegenwartstendenz von Personen als statistisch signifikanter Hauptgrund ermittelt werden. Viele Personen gaben an, die langfristigen ökologischen Kosten einer unsachgemäßen Entsorgung aufgrund des kurzfristigen persönlichen Nutzens zu ignorieren. Dieser entsteht durch die Gemütlichkeit des „Wegschnipsens“ einer Zigarettenkippe. Im Anschluss wurde ein auf die Gegenwartstendenz von Personen fokussierter Nudge entwickelt und in einem Feldexperiment auf seine Wirksamkeit überprüft, indem die Relation von unsachgemäß zu sachgemäß entsorgten Zigarettenkippen vor und nach dem Einsatz des Nudges dokumentiert wurde. Ohne Einsatz des Nudges (N = 92) wurden am Erhebungsort 64.1 Prozent und mit Einsatz des Nudges (N = 142) lediglich 38.0 Prozent der Zigarettenkippen unsachgemäß entsorgt. In dem Feldexperiment konnte der Nudge effektiv zur Förderung von nachhaltigem Verhalten eingesetzt werden.
Telepresence robots allow users to be spatially and socially present in remote environments. Yet, it can be challenging to remotely operate telepresence robots, especially in dense environments such as academic conferences or workplaces. In this paper, we primarily focus on the effect that a speed control method, which automatically slows the telepresence robot down when getting closer to obstacles, has on user behaviors. In our first user study, participants drove the robot through a static obstacle course with narrow sections. Results indicate that the automatic speed control method significantly decreases the number of collisions. For the second study we designed a more naturalistic, conference-like experimental environment with tasks that require social interaction, and collected subjective responses from the participants when they were asked to navigate through the environment. While about half of the participants preferred automatic speed control because it allowed for smoother and safer navigation, others did not want to be influenced by an automatic mechanism. Overall, the results suggest that automatic speed control simplifies the user interface for telepresence robots in static dense environments, but should be considered as optionally available, especially in situations involving social interactions.
An internal model of self-motion provides a fundamental basis for action in our daily lives, yet little is known about its development. The ability to control self-motion develops in youth and often deteriorates with advanced age. Self-motion generates relative motion between the viewer and the environment. Thus, the smoothness of the visual motion created will vary as control improves. Here, we study the influence of the smoothness of visually simulated self-motion on an observer's ability to judge how far they have travelled over a wide range of ages. Previous studies were typically highly controlled and concentrated on university students. But are such populations representative of the general public? And are there developmental and sex effects? Here, estimates of distance travelled (visual odometry) during visually induced self-motion were obtained from 466 participants drawn from visitors to a public science museum. Participants were presented with visual motion that simulated forward linear self-motion through a field of lollipops using a head-mounted virtual reality display. They judged the distance of their simulated motion by indicating when they had reached the position of a previously presented target. The simulated visual motion was presented with or without horizontal or vertical sinusoidal jitter. Participants' responses indicated that they felt they travelled further in the presence of vertical jitter. The effectiveness of the display increased with age over all jitter conditions. The estimated time for participants to feel that they had started to move also increased slightly with age. There were no differences between the sexes. These results suggest that age should be taken into account when generating motion in a virtual reality environment. Citizen science studies like this can provide a unique and valuable insight into perceptual processes in a truly representative sample of people.
The development of metals tailored to the metallurgical conditions of laser-based additive manufacturing is crucial to advance the maturity of these materials for their use in structural applications. While efforts in this regard are being carried out around the globe, the use of high strength eutectic alloys have, so far, received minor attention, although previous works showed that rapid solidification techniques can result in ultrafine microstructures with excellent mechanical performance, albeit for small sample sizes. In the present work, a eutectic Ti-32.5Fe alloy has been produced by laser powder bed fusion aiming at exploiting rapid solidification and the capability to produce bulk ultrafine microstructures provided by this processing technique.
Process energy densities between 160 J/mm³ and 180 J/mm³ resulted in a dense and crack-free material with an oxygen content of ~ 0.45 wt.% in which a hierarchical microstructure is formed by µm-sized η-Ti4Fe2Ox dendrites embedded in an ultrafine eutectic β-Ti/TiFe matrix. The microstructure was studied three-dimensionally using near-field synchrotron ptychographic X-ray computed tomography with an actual spatial resolution down to 39 nm to analyse the morphology of the eutectic and dendritic structures as well as to quantify their mass density, size and distribution. Inter-lamellar spacings down to ~ 30–50 nm were achieved, revealing the potential of laser-based additive manufacturing to generate microstructures smaller than those obtained by classical rapid solidification techniques for bulk materials. The alloy was deformed at 600 °C under compressive loading up to a strain of ~ 30% without damage formation, resulting in a compressive yield stress of ~ 800 MPa.
This study provides a first demonstration of the feasibility to produce eutectic Ti-Fe alloys with ultrafine microstructures by laser powder bed fusion that are suitable for structural applications at elevated temperature.
Long-term variability of solar irradiance and its implications for photovoltaic power in West Africa
(2020)
This paper addresses long-term changes in solar irradiance for West Africa (3° N to 20° N and 20° W to 16° E) and its implications for photovoltaic power systems. Here we use satellite irradiance (Surface Solar Radiation Data Set-Heliosat, Edition 2.1, SARAH-2.1) to derive photovoltaic yields. Based on 35 years of data (1983–2017) the temporal and regional variability as well as long-term trends of global and direct horizontal irradiance are analyzed. Furthermore, at four locations a detailed time series analysis is undertaken. The dry and the wet season are considered separately.
In der heutigen Zeit nimmt die Bedeutung schlanker und effektiver Prozesse in Unternehmen vor dem Hintergrund des Wettbewerbs sowie Kostendrucks stetig zu. Um dieser Herausforderung entgegenzuwirken, fokussieren sich Unternehmen auf die Identifikation neuer innovativer Potenziale. Aufgrund der Tatsache, dass monotone und regelbasierte Prozesse durch Softwareroboter automatisiert werden können, ist das Interesse an Robotic Process Automation (RPA) in den letzten Jahren stetig gestiegen. Bevor sich Unternehmen allerdings für oder gegen den Einsatz von RPA entscheiden, ist es zunächst notwendig, dass die Entscheidungsträger ein Verständnis von RPA erlangen sowie die entsprechenden Einsatzpotenziale und Risiken einschätzen können. Dieser Artikel trägt diesem Bedürfnis Rechnung, indem es diese auf Basis einer Literaturrecherche ermittelt und bewertet. Im Ausblick wird das zukünftige Potenzial von RPA eingeschätzt.
Background: 3-hydroxy-3-methylglutaryl-coenzyme A lyase deficiency (HMGCLD) is an autosomal recessive disorder of ketogenesis and leucine degradation due to mutations in HMGCL.
Method: We performed a systematic literature search to identify all published cases. Two hundred eleven patients of whom relevant clinical data were available were included in this analysis. Clinical course, biochemical findings and mutation data are highlighted and discussed. An overview on all published HMGCL variants is provided.
Results: More than 95% of patients presented with acute metabolic decompensation. Most patients manifested within the first year of life, 42.4% already neonatally. Very few individuals remained asymptomatic. The neurologic long-term outcome was favorable with 62.6% of patients showing normal development.
Conclusion: This comprehensive data analysis provides a systematic overview on all published cases with HMGCLD including a list of all known HMGCL mutations.
Computers can help us to trigger our intuition about how to solve a problem. But how does a computer take into account what a user wants and update these triggers? User preferences are hard to model as they are by nature vague, depend on the user’s background and are not always deterministic, changing depending on the context and process under which they were established. We pose that the process of preference discovery should be the object of interest in computer aided design or ideation. The process should be transparent, informative, interactive and intuitive. We formulate Hyper-Pref, a cyclic co-creative process between human and computer, which triggers the user’s intuition about what is possible and is updated according to what the user wants based on their decisions. We combine quality diversity algorithms, a divergent optimization method that can produce many, diverse solutions, with variational autoencoders to both model that diversity as well as the user’s preferences, discovering the preference hypervolume within large search spaces.
Until recently, studies regarding e-banking transactions have focused more on motivational factors that trigger the intention to accept and use the e-banking transaction, rather than the de-motivational factors that propel the action. However, in the developing countries like Sub-Sahara economies, the factors associated with the former have not been explored and are still rudimentary in the literature. Drawing from the Technology Threat Avoidance Theory (TTAT), the study seeks to examine the impact of online identity theft on customers’ willingness to engage in e-banking transactions in Ghana. A quantitative survey of 393 valid responses from retail bank customers amongst two leading commercial banks in Ghana for the analyses. Results from the PLS-SEM showed that the research constructs; perceived online identity theft’ positively and significantly predict “fear of financial loss”, “fear of reputational damage”, and “security and privacy concern” whilst the former has a negative mediated-relationship between perceived online identity theft and the intention to engage in e-banking transaction. This study is the first of its kind that has extended the application of the TTAT framework into the study of e-banking transactions. The study serves as a practical tool that will enable the banks in their quest to assess customers’ restriction/aversion towards the use of Fintech while ensuring sustainable growth of e-banking transactions in an emerging economy context. The study is limited to only banking institutions in Ghana without considering other players in the financial sub-sector. Future research direction has been suggested in the concluding part of the paper.
Do socio-economic factors impede the engagement in online banking transactions? Evidence from Ghana
(2020)
Researchers have long pondered on the online banking transaction adoption. Some of these studies focus primarily on the motivating factors that affect customers’ intention to adopt/accept these services (technologies). However, research into the constraining factors, in particular socio-economic factors, barely exist in the literature, especially in the context of sub-Saharan Africa. Against this background, the paper seeks to fill in this gap by: (1) assessing the socio-economic factors impeding the engagement of e-banking transactions among retail bank customers in Ghana, and (2) examining the moderating effect of ‘customer experience of Internet’ on the identified factors that inhibit the engagement in online banking in Ghana. The paper used a quantitative research approach to obtain data from two leading Ghanaian banks. Out of the 450 questionnaires distributed, 393 were valid for analysis. Data were analyzed with the aid of PLS-SEM (partial least squares and structural equation modeling). Findings revealed that perceived knowledge gap and the price of digital devices were directly important to the intention to disembark on e-banking transactions among Ghanaian bank customers. Whilst customer experience (frequent use of the Internet), as a moderator variable, has a significant effect on the interaction between perceived knowledge gap and the intent to disembark on e-banking transactions; and finance charges and the intent to disembark on e-banking transactions. Study implications and directions for future research are discussed in the paper.
4GREAT is an extension of the German Receiver for Astronomy at Terahertz frequencies (GREAT) operated aboard the Stratospheric Observatory for Infrared Astronomy (SOFIA). The spectrometer comprises four different detector bands and their associated subsystems for simultaneous and fully independent science operation. All detector beams are co-aligned on the sky. The frequency bands of 4GREAT cover 491-635, 890-1090, 1240-1525 and 2490-2590 GHz, respectively. This paper presents the design and characterization of the instrument, and its in-flight performance. 4GREAT saw first light in June 2018, and has been offered to the interested SOFIA communities starting with observing cycle 6.
2-methylacetoacetyl-coenzyme A thiolase (beta-ketothiolase) deficiency: one disease - two pathways
(2020)
Background: 2-methylacetoacetyl-coenzyme A thiolase deficiency (MATD; deficiency of mitochondrial acetoacetyl-coenzyme A thiolase T2/ “beta-ketothiolase”) is an autosomal recessive disorder of ketone body utilization and isoleucine degradation due to mutations in ACAT1.
Methods: We performed a systematic literature search for all available clinical descriptions of patients with MATD. Two hundred forty-four patients were identified and included in this analysis. Clinical course and biochemical data are presented and discussed.
Results: For 89.6% of patients at least one acute metabolic decompensation was reported. Age at first symptoms ranged from 2 days to 8 years (median 12 months). More than 82% of patients presented in the first 2 years of life, while manifestation in the neonatal period was the exception (3.4%). 77.0% (157 of 204 patients) of patients showed normal psychomotor development without neurologic abnormalities. Conclusion: This comprehensive data analysis provides a systematic overview on all cases with MATD identified in the literature. It demonstrates that MATD is a rather benign disorder with often favourable outcome, when compared with many other organic acidurias.
Intelligente Dialogsysteme – Chatbots – werden immer häufiger als virtuelle Ansprechpartner von Unternehmen und Institutionen eingesetzt. Auf Basis einer Wissensdatenbank können Chatbots einen größeren Anteil von Kundenanfragen automatisiert beantworten. Analog ist der Einsatz von Chatbots als digitaler Ansprechpartner öffentlicher Verwaltungen denkbar. Sie könnten Bürgern helfen, sich innerhalb der behördlichen Strukturen zu orientieren und Verwaltungsleistungen effizient und effektiv in Anspruch zu nehmen.
Diese Arbeit überprüft den Einsatz eines Chatbots in der öffentlichen Verwaltung hinsichtlich der entstehenden Kosten und des erwartbaren Nutzens. Auf Basis einer umfangreichen Literaturauswertung und der prototypischen Realisierung eines Chatbots für ein Stadtportal werden dabei Herausforderungen dieser Anwendungsdomäne herausgearbeitet, konkrete Funktionsweise und Implementierungsstrategien von Chatbots erörtert und einige Erfolgsfaktoren formuliert, die den Kern einer Handlungsempfehlung für Entscheidungsträger öffentlicher Verwaltungen bilden.
Object detectors have improved considerably in the last years by using advanced Convolutional Neural Networks (CNNs) architectures. However, many detector hyper-parameters are not generally tuned, and they are used with values set by the detector authors. Blackbox optimization methods have gained more attention in recent years because of its ability to optimize the hyper-parameters of various machine learning algorithms and deep learning models. However, these methods are not explored in improving CNN-based object detector's hyper-parameters. In this research work, we propose the use of blackbox optimization methods such as Gaussian Process based Bayesian Optimization (BOGP), Sequential Model-based Algorithm Configuration (SMAC), and Covariance Matrix Adaptation Evolution Strategy (CMA-ES) to tune the hyper-parameters in Faster R-CNN and Single Shot MultiBox Detector (SSD). In Faster R-CNN, tuning the input image size, prior box anchor scales and ratios using BOGP, SMAC, and CMA-ES has increased the performance around 1.5% in terms of Mean Average Precision (mAP) on PASCAL VOC. Tuning the anchor scales of SSD has increased the mAP by 3% on PASCAL VOC and marine debris datasets. On the COCO dataset with SSD, mAP improvement is observed in the medium and large objects, but mAP decreases by 1% in small objects. The experimental results show that the blackbox optimization methods have proved to increase the mAP performance by optimizing the object detectors. Moreover, it has achieved better results than the hand-tuned configurations in most of the cases.
The motor protein myosin drives a wide range of cellular and muscular functions by generating directed movement and force, fueled through adenosine triphosphate (ATP) hydrolysis. Release of the hydrolysis product adenosine diphosphate (ADP) is a fundamental and regulatory process during force production. However, details about the molecular mechanism accompanying ADP release are scarce due to the lack of representative structures. Here we solved a novel blebbistatin-bound myosin conformation with critical structural elements in positions between the myosin pre-power stroke and rigor states. ADP in this structure is repositioned towards the surface by the phosphate-sensing P-loop, and stabilized in a partially unbound conformation via a salt-bridge between Arg131 and Glu187. A 5 Å rotation separates the mechanical converter in this conformation from the rigor position. The crystallized myosin structure thus resembles a conformation towards the end of the two-step power stroke, associated with ADP release. Computationally reconstructing ADP release from myosin by means of molecular dynamics simulations further supported the existence of an equivalent conformation along the power stroke that shows the same major characteristics in the myosin motor domain as the resolved blebbistatin-bound myosin-II·ADP crystal structure, and identified a communication hub centered on Arg232 that mediates chemomechanical energy transduction.
In recent years, a plethora of observations with high spectral resolution of sub-millimetre and far-infrared transitions of methylidene (CH), conducted with Herschel and SOFIA, have demonstrated this radical to be a valuable proxy for molecular hydrogen that can be used for characterising molecular gas within the interstellar medium on a Galactic scale, including the CO-dark component. We report the discovery of the 13CH isotopologue in the interstellar medium using the upGREAT receiver on board SOFIA. We have detected the three hyperfine structure components of the ≈2 THz frequency transition from its X2Π1∕2 ground-state towards the high-mass star-forming regions Sgr B2(M), G34.26+0.15, W49(N), and W51E and determined 13CH column densities. The ubiquity of molecules containing carbon in the interstellar medium has turned the determination of the ratio between the abundances of the two stable isotopes of carbon, 12C/13C, into a cornerstone for Galactic chemical evolution studies. Whilst displaying a rising gradient with galactocentric distance, this ratio, when measured using observations of different molecules (CO, H2CO, and others), shows systematic variations depending on the tracer used. These observed inconsistencies may arise from optical depth effects, chemical fractionation, or isotope-selective photo-dissociation. Formed from C+ either through UV-driven or turbulence-driven chemistry, CH reflects the fractionation of C+, and does not show any significant fractionation effects, unlike other molecules that were previously used to determine the 12C/13C isotopic ratio. This makes it an ideal tracer for the 12C/13C ratio throughout the Galaxy. By comparing the derived column densities of 13CH with previously obtained SOFIA data of the corresponding transitions of the main isotopologue 12CH, we therefore derive 12C/13C isotopic ratios toward Sgr B2(M), G34.26+0.15, W49(N) and W51E. Adding our values derived from 12∕13CH to previous calculations of the Galactic isotopic gradient, we derive a revised value of 12C/13C = 5.87(0.45)RGC + 13.25(2.94).
Green infrastructure improves environmental health in cities, benefits human health, and provides habitat for wildlife. Increasing urbanization has demanded the expansion of urban areas and transformation of existing cities. The adoption of compact design in urban planning is a recommended strategy to minimize environmental impacts; however, it may undermine green infrastructure networks within cities as it sets a battleground for urban space. Under this scenario, multifunctionality of green spaces is highly desirable but reconciling human needs and biodiversity conservation in a limited space is still a challenge. Through a systematic review, we first compiled urban green space's characteristics that affect mental health and urban wildlife support, and then identified potential synergies and trade-offs between these dimensions. A framework based on the One Health approach is proposed, synthesizing the interlinkages between green space quality, mental health, and wildlife support; providing a new holistic perspective on the topic. Looking at the human-wildlife-environment relationships simultaneously may contribute to practical guidance on more effective green space design and management that benefit all dimensions.
Surface-enhanced Raman spectroscopy (SERS) with subsequent chemometric evaluation was performed for the rapid and non-destructive differentiation of seven important meat-associated microorganisms, namely Brochothrix thermosphacta DSM 20171, Pseudomonas fluorescens DSM 4358, Salmonella enterica subsp. enterica sv. Enteritidis DSM 14221, Listeria monocytogenes DSM 19094, Micrococcus luteus DSM 20030, Escherichia coli HB101 and Bacillus thuringiensis sv. israelensis DSM 5724. A simple method for collecting spectra from commercial paper-based SERS substrates without any laborious pre-treatments was used. In order to prepare the spectroscopic data for classification at genera level with a subsequent chemometric evaluation consisting of principal component analysis and discriminant analysis, a pre-processing method with spike correction and sum normalisation was performed. Because of the spike correction rather than exclusion, and therefore the use of a balanced data set, the multivariate analysis of the data is significantly resilient and meaningful. The analysis showed that the differentiation of meat-associated microorganisms and thereby the detection of important meat-related pathogenic bacteria was successful on genera level and a cross-validation as well as a classification of ungrouped data showed promising results, with 99.5 % and 97.5 %, respectively.
This work provides a short but technical introduction to the main building blocks of a blockchain. It argues that a blockchain is not a revolutionary technology but rather a clever combination of three fields: cryptography, decentralization and game theory. In addition, it summaries the differences between a public, private and federate blockchain model and the two prominent consensus mechanism Proof-of-Work (POW) and Proof-of-Stake (POS).
The temperature of photovoltaic modules is modelled as a dynamic function of ambient temperature, shortwave and longwave irradiance and wind speed, in order to allow for a more accurate characterisation of their efficiency. A simple dynamic thermal model is developed by extending an existing parametric steady-state model using an exponential smoothing kernel to include the effect of the heat capacity of the system. The four parameters of the model are fitted to measured data from three photovoltaic systems in the Allgäu region in Germany using non-linear optimisation. The dynamic model reduces the root-mean-square error between measured and modelled module temperature to 1.58 K on average, compared to 3.03 K for the steady-state model, whereas the maximum instantaneous error is reduced from 20.02 to 6.58 K.
Are There Extended Cognitive Improvements from Different Kinds of Acute Bouts of Physical Activity?
(2020)
Acute bouts of physical activity of at least moderate intensity have shown to enhance cognition in young as well as older adults. This effect has been observed for different kinds of activities such as aerobic or strength and coordination training. However, only few studies have directly compared these activities regarding their effectiveness. Further, most previous studies have mainly focused on inhibition and have not examined other important core executive functions (i.e., updating, switching) which are essential for our behavior in daily life (e.g., staying focused, resisting temptations, thinking before acting), as well. Therefore, this study aimed to directly compare two kinds of activities, aerobic and coordinative, and examine how they might affect executive functions (i.e., inhibition, updating, and switching) in a test-retest protocol. It is interesting for practical implications, as coordinative exercises, for example, require little space and would be preferable in settings such as an office or a classroom. Furthermore, we designed our experiment in such a way that learning effects were controlled. Then, we tested the influence of acute bouts of physical activity on the executive functioning in both young and older adults (young 16–22 years, old 65–80 years). Overall, we found no differences between aerobic and coordinative activities and, in fact, benefits from physical activities occurred only in the updating tasks in young adults. Additionally, we also showed some learning effects that might influence the results. Thus, it is important to control cognitive tests for learning effects in test-retest studies as well as to analyze effects from physical activity on a construct level of executive functions.
This work thoroughly investigates a semi-Lagrangian lattice Boltzmann (SLLBM) solver for compressible flows. In contrast to other LBM for compressible flows, the vertices are organized in cells, and interpolation polynomials up to fourth order are used to attain the off-vertex distribution function values. Differing from the recently introduced Particles on Demand (PoD) method , the method operates in a static, non-moving reference frame. Yet the SLLBM in the present formulation grants supersonic flows and exhibits a high degree of Galilean invariance. The SLLBM solver allows for an independent time step size due to the integration along characteristics and for the use of unusual velocity sets, like the D2Q25, which is constructed by the roots of the fifth-order Hermite polynomial. The properties of the present model are shown in diverse example simulations of a two-dimensional Taylor-Green vortex, a Sod shock tube, a two-dimensional Riemann problem and a shock-vortex interaction. It is shown that the cell-based interpolation and the use of Gauss-Lobatto-Chebyshev support points allow for spatially high-order solutions and minimize the mass loss caused by the interpolation. Transformed grids in the shock-vortex interaction show the general applicability to non-uniform grids.
Turbulent compressible flows are traditionally simulated using explicit Eulerian time integration applied to the Navier-Stokes equations. However, the associated Courant-Friedrichs-Lewy condition severely restricts the maximum time step size. Exploiting the Lagrangian nature of the Boltzmann equation's material derivative, we now introduce a feasible three-dimensional semi-Lagrangian lattice Boltzmann method (SLLBM), which elegantly circumvents this restriction. Previous lattice Boltzmann methods for compressible flows were mostly restricted to two dimensions due to the enormous number of discrete velocities needed in three dimensions. In contrast, this Rapid Communication demonstrates how cubature rules enhance the SLLBM to yield a three-dimensional velocity set with only 45 discrete velocities. Based on simulations of a compressible Taylor-Green vortex we show that the new method accurately captures shocks or shocklets as well as turbulence in 3D without utilizing additional filtering or stabilizing techniques, even when the time step sizes are up to two orders of magnitude larger compared to simulations in the literature. Our new method therefore enables researchers for the first time to study compressible turbulent flows by a fully explicit scheme, whose range of admissible time step sizes is only dictated by physics, while being decoupled from the spatial discretization.
Cytokine-induced killer (CIK) cells are heterogeneous, major histocompatibility complex (MHC)-unrestricted T lymphocytes that have acquired the expression of several natural killer (NK) cell surface markers following the addition of interferon gamma (IFN-γ), OKT3 and interleukin-2 (IL-2). Treatment with CIK cells demonstrates a practical approach in cancer immunotherapy with limited, if any, graft versus host disease (GvHD) toxicity. CIK cells have been proposed and tested in many clinical trials in cancer patients by autologous, allogeneic or haploidentical administration. The possibility of combining them with specific monoclonal antibodies nivolumab and ipilimumab will further expand the possibility of their clinical utilization. Initially, phenotypic analysis was performed to explore CD3, CD4, CD56, PD-1 and CTLA-4 expression on CIK cells and PD-L1/PD-L2 expression on tumor cells. We further treated CIK cells with nivolumab and ipilimumab and measured the cytotoxicity of CIK cells cocultured to renal carcinoma cell lines, A-498 and Caki-2. We observed a significant decrease in viability of renal cell lines after treating with CIK cells (p < 0.0001) in comparison to untreated renal cell lines and anti-PD-1 or anti-CTLA-4 treatment had no remarkable effect on the viability of tumor cells. Using CCK-8, Precision Count Beads™ and Cell Trace™ violet proliferation assays, we proved significant increased proliferation of CIK cells in the presence of a combination of anti-PD-1 and anti-CTLA-4 antibodies compared to untreated CIK cells. The IFN-γ secretion increased significantly in the presence of A-498 and combinatorial blockade of PD-1 and CTLA-4 compared to nivolumab or ipilimumab monotreatment (p < 0.001). In conclusion, a combination of immune checkpoint inhibition with CIK cells augments cytotoxicity of CIK cells against renal cancer cells.
Human and robot tasks in household environments include actions such as carrying an object, cleaning a surface, etc. These tasks are performed by means of dexterous manipulation, and for humans, they are straightforward to accomplish. Moreover, humans perform these actions with reasonable accuracy and precision but with much less energy and stress on the actuators (muscles) than the robots do. The high agility in controlling their forces and motions is actually due to "laziness", i.e. humans exploit the existing natural forces and constraints to execute the tasks.
The above-mentioned properties of the human lazy strategy motivate us to relax the problem of controlling robot motions and forces, and solve it with the help of the environment. Therefore, in this work, we developed a lazy control strategy, i.e. task specification models and control architectures that relax several aspects of robot control by exploiting prior knowledge about the task and environment. The developed control strategy is realized in four different robotics use cases. In this work, the Popov-Vereshchagin hybrid dynamics solver is used as one of the building blocks in the proposed control architectures. An extension of the solver’s interface with the artificial Cartesian force and feed-forward joint torque task-drivers is proposed in this thesis.
To validate the proposed lazy control approach, an experimental evaluation was performed in a simulation environment and on a real robot platform.
A Comparative Study of Uncertainty Estimation Methods in Deep Learning Based Classification Models
(2020)
Deep learning models produce overconfident predictions even for misclassified data. This work aims to improve the safety guarantees of software-intensive systems that use deep learning based classification models for decision making by performing comparative evaluation of different uncertainty estimation methods to identify possible misclassifications.
In this work, uncertainty estimation methods applicable to deep learning models are reviewed and those which can be seamlessly integrated to existing deployed deep learning architectures are selected for evaluation. The different uncertainty estimation methods, deep ensembles, test-time data augmentation and Monte Carlo dropout with its variants, are empirically evaluated on two standard datasets (CIFAR-10 and CIFAR-100) and two custom classification datasets (optical inspection and RoboCup@Work dataset). A relative ranking between the methods is provided by evaluating the deep learning classifiers on various aspects such as uncertainty quality, classifier performance and calibration. Standard metrics like entropy, cross-entropy, mutual information, and variance, combined with a rank histogram based method to identify uncertain predictions by thresholding on these metrics, are used to evaluate uncertainty quality.
The results indicate that Monte Carlo dropout combined with test-time data augmentation outperforms all other methods by identifying more than 95% of the misclassifications and representing uncertainty in the highest number of samples in the test set. It also yields a better classifier performance and calibration in terms of higher accuracy and lower Expected Calibration Error (ECE), respectively. A python based uncertainty estimation library for training and real-time uncertainty estimation of deep learning based classification models is also developed.
This paper addresses long-term historical changes in solar irradiance in West Africa (3 to 20° N and 20° W to 16° E) and the implications for photovoltaic systems. Here, we use satellite irradiance (Surface Solar Radiation Data Set – Heliosat, Edition 2.1 – SARAH-2.1) and temperature data from a reanalysis (ERA5) to derive photovoltaic yields. Based on 35 years of data (1983–2017), the temporal and regional variability as well as long-term trends in global and direct horizontal irradiance are analyzed. Furthermore, a detailed time series analysis is undertaken at four locations. According to the high spatial resolution SARAH-2.1 data record (0.05°×0.05°), solar irradiance is largest (up to a 300 W m−2 daily average) in the Sahara and the Sahel zone with a positive trend (up to 5 W m−2 per decade) and a lower temporal variability (<75 W m−2 between 1983 and 2017 for daily averages). In contrast, the solar irradiance is lower in southern West Africa (between 200 W m−2 and 250 W m−2) with a negative trend (up to −5 W m−2 per decade) and a higher temporal variability (up to 150 W m−2). The positive trend in the north is mostly connected to the dry season, whereas the negative trend in the south occurs during the wet season. Both trends show 95 % significance. Photovoltaic (PV) yields show a strong meridional gradient with the lowest values of around 4 kWh kWp−1 in southern West Africa and values of more than 5.5 kWh kWp−1 in the Sahara and Sahel zone.
Failure prognostic builds up on constant data acquisition and processing and fault diagnosis and is an essential part of predictive maintenance of smart manufacturing systems enabling condition based maintenance, optimised use of plant equipment, improved uptime and yield and to prevent safety problems. Given known control inputs into a plant and real sensor outputs or simulated measurements, the model-based part of the proposed hybrid method provides numerical values of unknown parameter degradation functions at sampling time points by the evaluation of equations that have been derived offline from a bicausal diagnostic bond graph. These numerical values are computed concurrently to the constant monitoring of a system and are stored in a buffer of fixed length. The data-driven part of the method provides a sequence of remaining useful life estimates by repeated projection of the parameter degradation into the future based on the use of values in a sliding time window. Existing software can be used to determine the best fitting function and can account for its random parameters. The continuous parameter estimation and their projection into the future can be performed in parallel for multiple isolated simultaneous parametric faults on a multicore, multiprocessor computer.
The proposed hybrid bond graph model-based, data-driven method is verified by an offline simulation case study of a typical power electronic circuit. It can be used to implement embedded systems that enable cooperating machines in smart manufacturing to perform prognostic themselves.
Long-term variability of solar irradiance and its implications for photovoltaic power in West Africa
(2020)
West Africa is one of the least developed regions in the world regarding the energy availability and energy security. Located close to the equator West Africa receives high amounts of global horizontal irradiance (GHI). Thus, solar power and especially photovoltaic (PV) systems seem to be a promising solution to provide electricity with low environmental impact. To plan and to dimension a PV power system climatological data for global horizontal irradiance (GHI) and its variability need to be taken into account. However, ground based measurements of irradiances are not available continuously and cover only a few discrete locations.
Diese Arbeit beschäftigt sich mit der Entwicklung eines, für die kontrollierte Freisetzung hydrophiler Wirkstoffe geeigneten, Verkapselungssystems mit dem Ziel die Freisetzung osteospezifischer P2-Liganden zu verzögern, um bei der Behandlung von Knochendefekten kritischer Größe die Bildung neuen Knochengewebes zu gewährleisten. Hierfür werden, unter Anwendung der immersiven Layer-by-Layer-Beschichtung, mit den Modell-Substanzen Adenosintriphosphat und Suramin versetzte, Alginat sowie κ-Carrageen-Kapseln mit Chitosan und Lignosulfonat beschichtet und auf ihr Freisetzungsverhalten hin untersucht.
In dieser vorliegenden Arbeit wurde der photolytische und photokatalytische Abbau von Lignin untersucht. Eine Charakterisierung des verwendeten Photoreaktors wurde mittels Kalium-Ferrioxalat-Aktinometrie durchgeführt. Zur Analyse der abgebauten Lignine wurde eine Optimierung einer bereits bestehenden Methode zur Bestimmung des Hydroxylgehaltes erarbeitet. Die Bestimmung der Hydroxylgehalte erfolgte demnach bei Raumtemperatur nach einer Acetylierungsdauer von 72 h und zeigte eine Abnahme der Hydroxylgehalte mit andauernder UV-Bestrahlung. Selbige Beobachtung konnte mit Hilfe der ATR-IR-Spektroskopie gemacht werden. Zusätzlich konnte die Bildung von Carbonsäuren und der Abbau von aromatischen Strukturen detektiert werden. Der Abbau aromatischer Strukturen konnte ebenfalls durch UV-VIS-Spektroskopie gezeigt werden. Eine Vermutung, dass es sich bei dem Abbauprozess um einen oxidativen Mechanismus handelt, konnte mit dem Abbau von Hydroxylgruppen über eine Bildung von Carbonsäuren zu Kohlenstoffdioxid bestätigt werden. Eine Freisetzung von Kohlenstoffdioxid konnte durch eine Bestimmung des IC festgestellt werden. Die Ergebnisse der Gel-Permeations-Chromatographie zusammen mit einer TOC-Analyse zeigen einen Abbau der molaren Masse des Lignins auf. Es konnten Fragmente mit einer Molmasse ähnlich der Monomere des Lignins gefunden werden. Der eingesetzte Photokatalysator wurde via Röntgenbeugung untersucht und konnte als das hoch photokatalytisch aktive P25 von Degussa identifiziert werden. Trotz des Einsatzes verschiedener Katalysatorkonzentrationen in einem Bereich von 0-0,5 g L^(-1) konnte kein Einfluss des Photokatalysators auf den Abbauprozess des Lignins beobachtet werden.
The simultaneous operation of multiple different semiconducting metal oxide (MOX) gas sensors is demanding for the readout circuitry. The challenge results from the strongly varying signal intensities of the various sensor types to the target gas. While some sensors change their resistance only slightly, other types can react with a resistive change over a range of several decades. Therefore, a suitable readout circuit has to be able to capture all these resistive variations, requiring it to have a very large dynamic range. This work presents a compact embedded system that provides a full, high range input interface (readout and heater management) for MOX sensor operation. The system is modular and consists of a central mainboard that holds up to eight sensor-modules, each capable of supporting up to two MOX sensors, therefore supporting a total maximum of 16 different sensors. Its wide input range is archived using the resistance-to-time measurement method. The system is solely built with commercial off-the-shelf components and tested over a range spanning from 100Ω to 5 GΩ (9.7 decades) with an average measurement error of 0.27% and a maximum error of 2.11%. The heater management uses a well-tested power-circuit and supports multiple modes of operation, hence enabling the system to be used in highly automated measurement applications. The experimental part of this work presents the results of an exemplary screening of 16 sensors, which was performed to evaluate the system’s performance.
In der vorliegenden Arbeit wurde Kraft-Lignin als Makromonomer für die Synthese von thermoplastischen Polyurethanen mit hoher molarer Masse durch acide Präzipitation aus Schwarzlauge isoliert. Die Charakterisierung des Rohstoffes bezüglich seiner Ausgangsmolmasse erfolgte mittels Gel-Permeations-Chromatographie mit Polystyren-Polymerstandard, welche sich als sehr hilfreiche Analysemethode erwies. Da das Kraft-Lignin die klassische Polyolkomponente bei der Synthese von Polyurethanen ersetzen sollte, war es notwendig, den Hydroxylgehalt des Kraft-Lignins zu bestimmen. Für diesen Zweck wurde eine bereits etablierte Prozedur zur nasschemischen Bestimmung des Hydroxylgehaltes von Polyolen für die Synthese von Polyurethanen einer Adaption unterzogen. Es wurde die Reaktionsdauer bei der Acetylierung des Kraft-Lignins variiert. Das Ergebnis war, dass die Messgenauigkeit durch eine Erhöhung der Reaktionsdauer von 1 h auf 3 h drastisch von 25,5 % auf 3,6 % reduziert werden konnte. Um abschätzen zu können, ob die erzielte Messgenauigkeit im Rahmen einer nasschemischen Prozedur mit manueller Titration liegt, wurden zusätzlich die Hydroxylgehalte von Ethandiol und Saccharose bestimmt. Diese dienten als Referenzsubstanz mit definierten und bekannten Hydroxylgehalten. Die Ermittlung der Hydroxylgehalte mit diesen Substanzen ergab für Ethandiol eine Messgenauigkeit von 2,2 % und für Saccharose eine Messgenauigkeit von 1,4 %. Eine Messgenauigkeit von 3,6 % ist in Anbetracht des Zeitaufwandes akzeptabel.
Für die Synthese von thermoplastischen Polyurethanen wurde Kraft-Lignin mit Methylendiphenyldiisocyanat in Dimethylacetamid mit Zinnoktoat als Katalysator zur Reaktion gebracht. Es wurde das NCO/OH-Verhältnis und die Reaktionsdauer variiert. Die Analyse der synthetisierten Polyurethane erfolgte mittels Ubbelohde-Kapillarviskosimetrie, Fourier-Transformations-Infrarotspektroskopie und Schmelzpunktbestimmung. Die FTIR-Spektren bestätigte eine erfolgreiche Synthese von Polyurethanen aus Kraft-Lignin und Methylendiphenyldiisocyanat und zeigte, dass die Variation des NCO/OH-Verhältnisses und der Reaktionsdauer keinerlei Einflüsse auf die chemische Grundstruktur des Polyurethans hat. Die Ubbelohde-Kapillarviskosimetrie belegte die thermoplastischen Eigenschaften des synthetisierten Polyurethans, die sich in einem thermoplastischen Nassprozess verarbeiten lassen. Sie zeigte auch die Abhängigkeit der Molmasse der synthetisierten Polyurethane von der Reaktionsdauer und vom NCO/OH-Verhältnis. So steigt die Molmasse des Polyurethans mit steigender Reaktionsdauer und sinkendem NCO/OH-Verhältnis. Letztere Beobachtung ist sogar praktisch hinsichtlich der gesundheitsgefährdenden Eigenschaft von Isocyanaten, da so der Einsatz von Isocyanaten reduziert werden kann. Um die schmelzflüssige Verarbeitbarkeit des synthetisierten Polyurethans zu untersuchen, wurden die Schmelzpunkte der Polymere bestimmt. Es konnte in einem Temperaturbereich von 25 °C-410 °C keine Aggregatzustandsänderung, sondern lediglich eine Zersetzungsreaktion beobachtet werden.
Miscanthus crops possess very attractive properties such as high photosynthesis yield and carbon fixation rate. Because of these properties, it is currently considered for use in second-generation biorefineries. Here we analyze the differences in chemical composition between M. x giganteus, a commonly studied Miscanthus genotype, and M. nagara, which is relatively understudied but has useful properties such as increased frost resistance and higher stem stability. Samples of M. x giganteus (Gig35) and M. nagara (NagG10) have been separated by plant portion (leaves and stems) in order to isolate the corresponding lignins. The organosolv process was used for biomass pulping (80% ethanol solution, 170 °C, 15 bar). Biomass composition and lignin structure analysis were performed using composition analysis, Fourier-transform infrared (FTIR), ultraviolet-visible (UV-Vis) and nuclear magnetic resonance (NMR) spectroscopy, thermogravimetric analysis (TGA), size exclusion chromatography (SEC) and pyrolysis gas-chromatography/mass spectrometry (Py-GC/MS) to determine the 3D structure of the isolated lignins, monolignol ratio and most abundant linkages depending on genotype and harvesting season. SEC data showed significant differences in the molecular weight and polydispersity indices for stem versus leaf-derived lignins. Py-GC/MS and hetero-nuclear single quantum correlation (HSQC) NMR revealed different monolignol compositions for the two genotypes (Gig35, NagG10). The monolignol ratio is slightly influenced by the time of harvest: stem-derived lignins of M. nagara showed increasing H and decreasing G unit content over the studied harvesting period (December–April).
Der Beitrag untersucht, wie ein Präsenzlabor durch ein Remote-Labor ergänzt undersetzt werden kann. Dazu wird das Laborpraktikum Digitaltechnik der Hochschule Bonn-Rhein-Sieg betrachtet, bei dem ein Remote-Labor Flexibilität bei der Versuchsdurchführung bietet und Versuche ermöglicht, die allein mit dem Präsenzlabor nicht möglich wären. Neben der Ergänzung der Präsenzversuche können Studie-rende das Praktikum auch komplett im Remote-Labor durchführen. Durch klare Anforderungen an die Erteilung eines Testats ist dies sowohl für sie als auch für Lehrende praktikabel zu handhaben. Rückmeldungen der Studierenden und Nutzungszahlen belegen die Akzeptanz des Remote-Labors. Dabei zeigt sich, dass die Studierenden sehr heterogen mit dem Remote-Labor umgehen: Einige von ihnen nutzen das Remote-Labor als zusätzliche Praktikumszeit für Versuche die auch im Präsenzlabor möglich wären; andere nutzen es als Erweiterung der Praktikumsmöglichkeit für Versuche, die nur im Remote-Labor möglich sind und wieder andere arbeiten intensiv im Remote-Labor und reichen auch das Praktikumsprotokoll elektronisch ein. Für Lehrende besteht über das Protokoll und die Auswertung der Nutzungsdaten ausreichende Sicherheit, um aktive Beteiligung am Praktikum zu testieren.
Im Rahmen der Arbeit wurde Kraft-Lignin mit Natriumsulfit demethyliert, um den Gehalt an aromatischen Hydroxygruppen zu erhöhen und damit die Reaktivität des Lignins in Bezug auf Polyurethan-Synthesen zu erhöhen. Variiert wurden die Demethylierungstemperatur (72°C, 90°C) sowie der pH-Wert zur Isolierung des Kraft-Lignins (pH 2, 3, 4 und 5). Die Analyse der demethylierten Proben erfolgte mittels differentieller UV-Spektroskopie und der OH-Gehaltbestimmung via automatischer Titration (angelehnt an ISO 14900:2001(E)). Weitere Untersuchungen umfassten Löslichkeitstests sowie Strukturanalysen via FTIR- und UV/Vis-Spektroskopie.
The development of sustainable, environmentally friendly insulation materials with a reduced carbon footprint is attracting increased interest. One alternative to conventional insulation materials are foamed geopolymers. Similar to foamed concrete, the mechanical properties of geopolymer foams can also be improved by using fibers for reinforcement. This paper presents an overview of the latest research findings in the field of fiber-reinforced geopolymer foam concrete with special focus on natural fibers reinforcement. Furthermore, some basic and background information of natural fibers and geopolymer foams are reported. In most of the research, foams are produced either through chemical foaming with hydrogen peroxide or aluminum powder, or through mechanical foaming which includes a foaming agent. However, previous reviews have not sufficiently addresses the fabrication of geopolymer foams by syntactic foams. Finally, recent efforts to reduce the fiber degradation in geopolymer concrete are discussed along with challenges for natural fiber reinforced-geopolymer foam concrete.
Development of colored surfaces by formation of nano-structured aggregates is a widely used strategy in nature to color lightweight structures (e.g. butterflies) without the use of dye pigments. The deposition of nanoscale particles mimics nature in it’s approach coloring surfaces. This work presents sol-gel modification of cellulose surfaces used to form a template for growth of Cu/Cu2O core-shell particles with defined size-distributions. Besides improving the adhesion of the deposited particulate material, the sol-gel matrix serves as a template for the control of particle sizes of the Cu/Cu2O structures, and as a consequence of particle size variation the surface color is tunable. As an example, red color was achieved with an average particle size of 35 nm, and shifts gradually to blue appearance when particles have grown to 80 nm on the sol-gel modified fabric. The copper concentration on representative fabrics is kept low to avoid modifying the textile characteristics and were all in the range of 150–170 mg per g of cellulose material. As a result of copper deposition on the surface of the material, the cellulose fabric also became electrically conductive. Remarkably, the electrical conductivity was found to be dependent on the average particle sizes of the deposits and thus related to the change in observed color. The generation of color by growth of nano-sized particles on sol-gel templates provides a highly promising approach to stain surfaces by physical effects without use of synthetic colorants, which opens a new strategy to improve environmental profile of coloration.