Refine
H-BRS Bibliography
- yes (1196) (remove)
Departments, institutes and facilities
- Präsidium (397)
- Fachbereich Angewandte Naturwissenschaften (189)
- Fachbereich Informatik (178)
- Fachbereich Wirtschaftswissenschaften (154)
- Institut für Technik, Ressourcenschonung und Energieeffizienz (TREE) (133)
- Fachbereich Ingenieurwissenschaften und Kommunikation (124)
- Internationales Zentrum für Nachhaltige Entwicklung (IZNE) (100)
- Institut für funktionale Gen-Analytik (IFGA) (72)
- Fachbereich Sozialpolitik und Soziale Sicherung (43)
- Institute of Visual Computing (IVC) (41)
Document Type
- Article (436)
- Part of Periodical (407)
- Conference Object (102)
- Part of a Book (72)
- Report (54)
- Working Paper (42)
- Preprint (19)
- Bachelor Thesis (18)
- Master's Thesis (14)
- Other (10)
Year of publication
Has Fulltext
- yes (1196) (remove)
Keywords
- Hochschule Bonn-Rhein-Sieg (7)
- Machine Learning (7)
- Robotik (7)
- cytokine-induced killer cells (7)
- lignin (7)
- Digitalisierung (6)
- Lignin (6)
- Sustainability (6)
- immunotherapy (6)
- GC/MS (5)
Exposure to microgravity conditions causes cardiovascular deconditioning in astronauts during spaceflight. Until now, no specific drugs are available for countermeasure, since the underlying mechanism is largely unknown. Endothelial cells (ECs) and smooth muscle cells (SMCs) play key roles in various vascular functions, many of which are regulated by purinergic 2 (P2) receptors. However, their function in ECs and SMCs under microgravity conditions is still unclear. In this study, primary ECs and SMCs were isolated from bovine aorta and verified with specific markers. We show for the first time that the P2 receptor expression pattern is altered in ECs and SMCs after 24 h exposure to simulated microgravity using a clinostat. However, conditioned medium compensates this change in specific P2 receptors, for example, P2X7. Notably, P2 receptors such as P2X7 might be the important players during the paracrine interaction. Additionally, ECs and SMCs secreted different cytokines under simulated microgravity, leading into a pathogenic proliferation and migration. In conclusion, our data indicate P2 receptors might be important players responding to gravity changes in ECs and SMCs. Since some artificial P2 receptor ligands are applied as drugs, it is reasonable to assume that they might be promising candidates against cardiovascular deconditioning in the future.
Human mesenchymal stem cells (hMSCs) are considered a promising cell source for regenerative medicine, because they have the potential to differentiate into a variety of lineages among which the mesoderm-derived lineages such adipo- or osteogenesis are investigated best. Human MSCs can be harvested in reasonable to large amounts from several parts of the patient’s body and due to this possible autologous origin, allorecognition can be avoided. In addition, even in allogenic origin-derived donor cells, hMSCs generate a local immunosuppressive microenvironment, causing only a weak immune reaction. There is an increasing need for bone replacement in patients from all ages, due to a variety of reasons such as a new recreational behavior in young adults or age-related diseases. Adipogenic differentiation is another interesting lineage, because fat tissue is considered to be a major factor triggering atherosclerosis that ultimately leads to cardiovascular diseases, the main cause of death in industrialized countries. However, understanding the differentiation process in detail is obligatory to achieve a tight control of the process for future clinical applications to avoid undesired side effects. In this review, the current findings for adipo- and osteo-differentiation are summarized together with a brief statement on first clinical trials.
Background: Human mesenchymal stem cells (hMSCs) have shown their multipotential including differentiating towards endothelial and smooth muscle cell lineages, which triggers a new interest for using hMSCs as a putative source for cardiovascular regenerative medicine. Our recent publication has shown for the first time that purinergic 2 receptors are key players during hMSC differentiation towards adipocytes and osteoblasts. Purinergic 2 receptors play an important role in cardiovascular function when they bind to extracellular nucleotides. In this study, the possible functional role of purinergic 2 receptors during MSC endothelial and smooth muscle differentiation was investigated. Methods and Results: Human MSCs were isolated from liposuction materials. Then, endothelial and smooth muscle-like cells were differentiated and characterized by specific markers via Reverse Transcriptase-PCR (RT-PCR), Western blot and immunochemical stainings. Interestingly, some purinergic 2 receptor subtypes were found to be differently regulated during these specific lineage commitments: P2Y4 and P2Y14 were involved in the early stage commitment while P2Y1 was the key player in controlling MSC differentiation towards either endothelial or smooth muscle cells. The administration of natural and artificial purinergic 2 receptor agonists and antagonists had a direct influence on these differentiations. Moreover, a feedback loop via exogenous extracellular nucleotides on these particular differentiations was shown by apyrase digest. Conclusions: Purinergic 2 receptors play a crucial role during the differentiation towards endothelial and smooth muscle cell lineages. Some highly selective and potent artificial purinergic 2 ligands can control hMSC differentiation, which might improve the use of adult stem cells in cardiovascular tissue engineering in the future.
Cytokine-induced killer cells (CIK) in combination with dendritic cells (DCs) have shown favorable outcomes in renal cell carcinoma (RCC), yet some patients exhibit recurrence or no response to this therapy. In a broader perspective, enhancing the antitumor response of DC-CIK cells may help to address this issue. Considering this, herein, we investigated the effect of anti-CD40 and anti-CTLA-4 antibodies on the antitumor response of DC-CIK cells against RCC cell lines. Our analysis showed that, a) anti-CD40 antibody (G28.5) increased the CD3+CD56+ effector cells of CIK cells by promoting the maturation and activation of DCs, b) G28.5 also increased CTLA-4 expression in CIK cells via DCs, but the increase could be hindered by the CTLA-4 inhibitor (ipilimumab), c) adding ipilimumab was also able to significantly increase the proportion of CD3+CD56+ cells in DC-CIK cells, d) anti-CD40 antibodies predominated over anti-CTLA-4 antibodies for cytotoxicity, apoptotic effect and IFN-g secretion of DC-CIK cells against RCC cells, e) after ipilimumab treatment, the population of Tregs in CIK cells remained unaffected, but ipilimumab combined with G28.5 significantly reduced the expression of CD28 in CIK cells. Taken together, we suggest that the agonistic anti-CD40 antibody rather than CTLA-4 inhibitor may improve the antitumor response of DC-CIK cells, particularly in RCC. In addition, we pointed towards the yet to be known contribution of CD28 in the crosstalk between anti-CTLA-4 and CIK cells.
Cancer is a complex disease where resistance to therapies and relapses often pose a serious clinical challenge. The scenario is even more complicated when the cancer type itself is heterogeneous in nature, e.g., lymphoma, a cancer of the lymphocytes which constitutes more than 70 different subtypes. Indeed, the treatment options continue to expand in lymphomas. Herein, we provide insights into lymphoma-specific clinical trials based on cytokine-induced killer (CIK) cell therapy and other pre-clinical lymphoma models where CIK cells have been used along with other synergetic tumor-targeting immune modules to improve their therapeutic potential. From a broader perspective, we will highlight that CIK cell therapy has potential, and in this rapidly evolving landscape of cancer therapies its optimization (as a personalized therapeutic approach) will be beneficial in lymphomas.
We describe a systematic approach for rendering time-varying simulation data produced by exa-scale simulations, using GPU workstations. The data sets we focus on use adaptive mesh refinement (AMR) to overcome memory bandwidth limitations by representing interesting regions in space with high detail. Particularly, our focus is on data sets where the AMR hierarchy is fixed and does not change over time. Our study is motivated by the NASA Exajet, a large computational fluid dynamics simulation of a civilian cargo aircraft that consists of 423 simulation time steps, each storing 2.5 GB of data per scalar field, amounting to a total of 4 TB. We present strategies for rendering this time series data set with smooth animation and at interactive rates using current generation GPUs. We start with an unoptimized baseline and step by step extend that to support fast streaming updates. Our approach demonstrates how to push current visualization workstations and modern visualization APIs to their limits to achieve interactive visualization of exa-scale time series data sets.
Modern GPUs come with dedicated hardware to perform ray/triangle intersections and bounding volume hierarchy (BVH) traversal. While the primary use case for this hardware is photorealistic 3D computer graphics, with careful algorithm design scientists can also use this special-purpose hardware to accelerate general-purpose computations such as point containment queries. This article explains the principles behind these techniques and their application to vector field visualization of large simulation data using particle tracing.
When the Artemis missions launch, NASA's Orion spacecraft (and crew as of the Artemis II mission) will be exposed to the deep space radiation environment beyond the protection of Earth's magnetosphere. Hence, it is essential to characterize the effects of space radiation, microgravity, and the combination thereof on cells and organisms, i.e., to quantify any correlations between the deep space radiation environment, genetic variation, and induced genetic changes in cells. To address this, the Artemis I mission will include the Peristaltic Laboratory for Automated Science with Multigenerations (PLASM) hardware containing the Deep Space Radiation Genomics (DSRG) experiment. The scientific aims of DSRG are (i) to identify the metabolic and genomic pathways in yeast affected by microgravity, space radiation, and their combination, and (ii) to differentiate between gravity and radiation exposure on single-gene deletion/overexpressing strains' ability to thrive in the spaceflight environment. Yeast is used as a model system because 70% of its essential genes have a human homolog, and over half of these homologs can functionally replace their human counterpart. As part of the experiment preparation towards spaceflight, an Experiment Verification Test (EVT) was performed at the Kennedy Space Center to verify that the experiment design, hardware, and approach to automated operations will enable achieving the scientific aims. For the EVT, fluidic systems were assembled, sterilized, loaded, and acceptance-tested, and subsequently integrated with the engineering parts to produce a flight-like PLASM unit. Each fluidic system consisted of (i) a Media Bag, (ii) four Culture Bags loaded with Saccharomyces cerevisiae (two with deletion series and the remaining two with overexpression series), and (iii) tubing and check valves. The EVT PLASM unit was put under a temperature profile replicating the anticipated different phases of flight, including handover to launch, spaceflight, and splashdown to handover back to the science team, for a 58-day period. At EVT completion, the rate of activation, cellular growth, RNA integrity, and sample contamination were interrogated. All of the experiment's success criteria were satisfied, encouraging our efforts to perform this investigation on Artemis I. This manuscript thus describes the process of spaceflight experiment design maturation with a focus on the EVT, its results, DSRG's preparation for its planned launch on Artemis I in 2022, and how the PLASM hardware can enable other scientific goals on future Artemis missions and/or the Lunar Orbital Platform – Gateway.
Extremophiles are optimal models in experimentally addressing questions about the effects of cosmic radiation on biological systems. The resistance to high charge energy (HZE) particles, and helium (He) ions and iron (Fe) ions (LET at 2.2 and 200 keV/µm, respectively, until 1000 Gy), of spores from two thermophiles, Bacillushorneckiae SBP3 and Bacilluslicheniformis T14, and two psychrotolerants, Bacillus sp. A34 and A43, was investigated. Spores survived He irradiation better, whereas they were more sensitive to Fe irradiation (until 500 Gy), with spores from thermophiles being more resistant to irradiations than psychrotolerants. The survived spores showed different germination kinetics, depending on the type/dose of irradiation and the germinant used. After exposure to He 1000 Gy, D-glucose increased the lag time of thermophilic spores and induced germination of psychrotolerants, whereas L-alanine and L-valine increased the germination efficiency, except alanine for A43. FTIR spectra showed important modifications to the structural components of spores after Fe irradiation at 250 Gy, which could explain the block in spore germination, whereas minor changes were observed after He radiation that could be related to the increased permeability of the inner membranes and alterations of receptor complex structures. Our results give new insights on HZE resistance of extremophiles that are useful in different contexts, including astrobiology.
Intention: Within the research project EnerSHelF (Energy-Self-Sufficiency for Health Facilities in Ghana), i. a. energy-meteorological and load-related measurement data are collected, for which an overview of the availability is to be presented on a poster.
Context: In Ghana, the total electricity consumed has almost doubled between 2008 and 2018 according to the Energy Commission of Ghana. This goes along with an unstable power grid, resulting in power outages whenever electricity consumption peaks. The blackouts called "dumsor" in Ghana, pose a severe burden to the healthcare sector. Innovative solutions are needed to reduce greenhouse gas emissions and improve energy and health access.
Nachhaltige und zukunftsfähige Mobilität in Städten kann langfristig nur durch die aktive Partizipation ihrer Bürger und Institutionen erreicht werden. Betriebliches Mobilitätsmanagement (BMM) kann dabei einen positiven Beitrag im Hinblick auf Umwelt, Gesundheit und Kosten leisten. Die vorliegende Arbeit beschäftigt sich mit der Wahrnehmung gesundheitlicher und finanzieller Wertschöpfungsaspekte des BMM. Im Rahmen des Forschungsprojekts Betriebe lösen Verkehrsprobleme werden Mobilitätsverhalten und Maßnahmen der Betrieblichen Gesundheitsförderung (BFG) in Bonner Betrieben untersucht. Folgenden Aspekten wird besondere Beachtung geschenkt: Bedeutung Betrieblicher Gesundheitsförderung in Bonner Betrieben, Mobilitätsverhalten von Arbeitnehmern auf dem Weg zur Dienststelle, Wahrnehmung eines unmittelbaren Zusammenhangs zwischen körperlicher Aktivität und Gesundheit bzw. krankheitsbedingter Kosten und Umsatzeinbußen durch Bewegungsmangel. Die Analyse resultiert auf der Basis einer schriftlichen Befragung von 178 Unternehmen, einer Online-Umfrage von 1.341 Mitarbeitern aus 14 Unternehmen sowie auf persönlichen Interviews mit 22 Betriebsleitern bzw. Mobilitäts- und Gesundheitsbeauftragten. Die Ergebnisse der Studie machen sowohl Handlungsbedarf als auch Optimierungspotentiale im Bereich BMM auf Betriebsseite deutlich. Kostensimulationen zeigen darüber hinaus auf, dass durch die Implementierung von BGF-Maßnahmen, explizit der Förderung von Bewegung, auf betriebs- und volkswirtschaftlicher Seite beachtliche Kosten im Gesundheitsbereich eingespart sowie höhere Gewinne im Unternehmen erzielt werden können.
Cytokine-induced killer (CIK) cells are an ex vivo expanded heterogeneous cell population with an enriched NK-T phenotype (CD3+CD56+). Due to the convenient and relatively inexpensive expansion capability, together with low incidence of graft versus host disease (GVHD) in allogeneic cancer patients, CIK cells are a promising candidate for immunotherapy. It is well known that natural killer group 2D (NKG2D) plays an important role in CIK cell-mediated antitumor activity; however, it remains unclear whether its engagement alone is sufficient or if it requires additional co-stimulatory signals to activate the CIK cells. Likewise, the role of 2B4 has not yet been identified in CIK cells. Herein, we investigated the individual and cumulative contribution of NKG2D and 2B4 in the activation of CIK cells. Our analysis suggests that (a) NKG2D (not 2B4) is implicated in CIK cell (especially CD3+CD56+ subset)-mediated cytotoxicity, IFN-γ secretion, E/T conjugate formation, and degranulation; (b) NKG2D alone is adequate enough to induce degranulation, IFN-γ secretion, and LFA-1 activation in CIK cells, while 2B4 only provides limited synergy with NKG2D (e.g., in LFA-1 activation); and (c) NKG2D was unable to costimulate CD3. Collectively, we conclude that NKG2D engagement alone suffices to activate CIK cells, thereby strengthening the idea that targeting the NKG2D axis is a promising approach to improve CIK cell therapy for cancer patients. Furthermore, CIK cells exhibit similarities to classical invariant natural killer (iNKT) cells with deficiencies in 2B4 stimulation and in the costimulation of CD3 with NKG2D. In addition, based on the current data, the divergence in receptor function between CIK cells and NK (or T) cells can be assumed, pointing to the possibility that molecular modifications (e.g., using chimeric antigen receptor technology) on CIK cells may need to be customized and optimized to maximize their functional potential.
Healing of large bone defects requires implants or scaffolds that provide structural guidance for cell growth, differentiation, and vascularization. In the present work, an agarose-hydroxyapatite composite scaffold was developed that acts not only as a 3D matrix, but also as a release system. Hydroxyapatite (HA) was incorporated into the agarose gels in situ in various ratios by a simple procedure consisting of precipitation, cooling, washing, and drying. The resulting gels were characterized regarding composition, porosity, mechanical properties, and biocompatibility. A pure phase of carbonated HA was identified in the scaffolds, which had pore sizes of up to several hundred micrometers. Mechanical testing revealed elastic moduli of up to 2.8 MPa for lyophilized composites. MTT testing on Lw35human mesenchymal stem cells (hMSCs) and osteosarcoma MG-63 cells proved the biocompatibility of the scaffolds. Furthermore, scaffolds were loaded with model drug compounds for guided hMSC differentiation. Different release kinetic models were evaluated for adenosine 5′-triphosphate (ATP) and suramin, and data showed a sustained release behavior over four days.
Bone tissue engineering is an ever-changing, rapidly evolving, and highly interdisciplinary field of study, where scientists try to mimic natural bone structure as closely as possible in order to facilitate bone healing. New insights from cell biology, specifically from mesenchymal stem cell differentiation and signaling, lead to new approaches in bone regeneration. Novel scaffold and drug release materials based on polysaccharides gain increasing attention due to their wide availability and good biocompatibility to be used as hydrogels and/or hybrid components for drug release and tissue engineering. This article reviews the current state of the art, recent developments, and future perspectives in polysaccharide-based systems used for bone regeneration.
Renewable resources are gaining increasing interest as a source for environmentally benign biomaterials, such as drug encapsulation/release compounds, and scaffolds for tissue engineering in regenerative medicine. Being the second largest naturally abundant polymer, the interest in lignin valorization for biomedical utilization is rapidly growing. Depending on its resource and isolation procedure, lignin shows specific antioxidant and antimicrobial activity. Today, efforts in research and industry are directed toward lignin utilization as a renewable macromolecular building block for the preparation of polymeric drug encapsulation and scaffold materials. Within the last five years, remarkable progress has been made in isolation, functionalization and modification of lignin and lignin-derived compounds. However, the literature so far mainly focuses lignin-derived fuels, lubricants and resins. The purpose of this review is to summarize the current state of the art and to highlight the most important results in the field of lignin-based materials for potential use in biomedicine (reported in 2014⁻2018). Special focus is placed on lignin-derived nanomaterials for drug encapsulation and release as well as lignin hybrid materials used as scaffolds for guided bone regeneration in stem cell-based therapies.
Polyether and polyether/ester based TPU (thermoplastic polyurethanes) were investigated with wide-angle XRD (X-ray diffraction) and SAXS (small angle X-ray scattering). Furthermore, SAXS measurements were performed in the temperature range of 30 °C to 130 °C. Polyether based polymers exhibit only one broad diffraction signal in a region of 2 θ 15° to 25°. In case of polyurethanes with ether/ester modification, the broad diffraction signal arises with small sharp diffraction signals. SAXS measurements of polymers reveal the size and shape of the crystalline zones of the polymer. Between 30 °C and 130 °C the size of the crystalline zone changes significantly. The size decreases in most of investigated TPU. In the case of Desmopan 9365D an increase of the particle size was observed.
Approximately 45% of global greenhouse gas emissions are caused by the construction and use of buildings. Thermal insulation of buildings in the current context of climate change is a well-known strategy to improve the energy efficiency of buildings. The development of renewable insulation material can overcome the drawbacks of widely used insulation systems based on polystyrene or mineral wool. This study analyzes the sustainability and thermal conductivity of new insulation materials made of Miscanthus x giganteus fibers, foaming agents, and alkali-activated fly ash binder. Life cycle assessments (LCA) are necessary to perform benchmarking of environmental impacts of new formulations of geopolymer-based insulation materials. The global warming potential (GWP) of the product is primarily determined by the main binder component sodium silicate. Sodium silicate's CO2 emissions depend on local production, transportation, and energy consumption. The results, which have been published during recent years, vary in a wide range from 0.3 kg to 3.3 kg CO2-eq. kg-1. The overall GWP of the insulation system based on Miscanthus fibers, with properties according to current thermal insulation regulations, reaches up to 95% savings of CO2 emissions compared to conventional systems. Carbon neutrality can be achieved through formulations containing raw materials with carbon dioxide emissions and renewable materials with negative GWP, thus balancing CO2 emissions.
The clear-sky radiative effect of aerosol-radiation interactions is of relevance for our understanding of the climate system. The influence of aerosol on the surface energy budget is of high interest for the renewable energy sector. In this study, the radiative effect is investigated in particular with respect to seasonal and regional variations for the region of Germany and the year 2015 at the surface and top of atmosphere using two complementary approaches.
First, an ensemble of clear-sky models which explicitly consider aerosols is utilized to retrieve the aerosol optical depth and the surface direct radiative effect of aerosols by means of a clear sky fitting technique. For this, short-wave broadband irradiance measurements in the absence of clouds are used as a basis. A clear sky detection algorithm is used to identify cloud free observations. Considered are measurements of the shortwave broadband global and diffuse horizontal irradiance with shaded and unshaded pyranometers at 25 stations across Germany within the observational network of the German Weather Service (DWD). Clear sky models used are MMAC, MRMv6.1, METSTAT, ESRA, Heliosat-1, CEM and the simplified Solis model. The definition of aerosol and atmospheric characteristics of the models are examined in detail for their suitability for this approach.
Second, the radiative effect is estimated using explicit radiative transfer simulations with inputs on the meteorological state of the atmosphere, trace-gases and aerosol from CAMS reanalysis. The aerosol optical properties (aerosol optical depth, Ångström exponent, single scattering albedo and assymetrie parameter) are first evaluated with AERONET direct sun and inversion products. The largest inconsistency is found for the aerosol absorption, which is overestimated by about 0.03 or about 30 % by the CAMS reanalysis. Compared to the DWD observational network, the simulated global, direct and diffuse irradiances show reasonable agreement within the measurement uncertainty. The radiative kernel method is used to estimate the resulting uncertainty and bias of the simulated direct radiative effect. The uncertainty is estimated to −1.5 ± 7.7 and 0.6 ± 3.5 W m−2 at the surface and top of atmosphere, respectively, while the annual-mean biases at the surface, top of atmosphere and total atmosphere are −10.6, −6.5 and 4.1 W m−2, respectively.
The retrieval of the aerosol radiative effect with the clear sky models shows a high level of agreement with the radiative transfer simulations, with an RMSE of 5.8 W m−2 and a correlation of 0.75. The annual mean of the REari at the surface for the 25 DWD stations shows a value of −12.8 ± 5 W m−2 as average over the clear sky models, compared to −11 W m−2 from the radiative transfer simulations. Since all models assume a fixed aerosol characterisation, the annual cycle of the aerosol radiation effect cannot be reproduced. Out of this set of clear sky models, the largest level of agreement is shown by the ESRA and MRMv6.1 models.
The clear-sky radiative effect of aerosol–radiation interactions is of relevance for our understanding of the climate system. The influence of aerosol on the surface energy budget is of high interest for the renewable energy sector. In this study, the radiative effect is investigated in particular with respect to seasonal and regional variations for the region of Germany and the year 2015 at the surface and top of atmosphere using two complementary approaches.
First, an ensemble of clear-sky models which explicitly consider aerosols is utilized to retrieve the aerosol optical depth and the surface direct radiative effect of aerosols by means of a clear-sky fitting technique. For this, short-wave broadband irradiance measurements in the absence of clouds are used as a basis. A clear-sky detection algorithm is used to identify cloud-free observations. Considered are measurements of the short-wave broadband global and diffuse horizontal irradiance with shaded and unshaded pyranometers at 25 stations across Germany within the observational network of the German Weather Service (DWD). The clear-sky models used are the Modified MAC model (MMAC), the Meteorological Radiation Model (MRM) v6.1, the Meteorological–Statistical solar radiation model (METSTAT), the European Solar Radiation Atlas (ESRA), Heliosat-1, the Center for Environment and Man solar radiation model (CEM), and the simplified Solis model. The definition of aerosol and atmospheric characteristics of the models are examined in detail for their suitability for this approach.
Second, the radiative effect is estimated using explicit radiative transfer simulations with inputs on the meteorological state of the atmosphere, trace gases and aerosol from the Copernicus Atmosphere Monitoring Service (CAMS) reanalysis. The aerosol optical properties (aerosol optical depth, Ångström exponent, single scattering albedo and asymmetry parameter) are first evaluated with AERONET direct sun and inversion products. The largest inconsistency is found for the aerosol absorption, which is overestimated by about 0.03 or about 30 % by the CAMS reanalysis. Compared to the DWD observational network, the simulated global, direct and diffuse irradiances show reasonable agreement within the measurement uncertainty. The radiative kernel method is used to estimate the resulting uncertainty and bias of the simulated direct radiative effect. The uncertainty is estimated to −1.5 ± 7.7 and 0.6 ± 3.5 W m−2 at the surface and top of atmosphere, respectively, while the annual-mean biases at the surface, top of atmosphere and total atmosphere are −10.6, −6.5 and 4.1 W m−2, respectively.
The retrieval of the aerosol radiative effect with the clear-sky models shows a high level of agreement with the radiative transfer simulations, with an RMSE of 5.8 W m−2 and a correlation of 0.75. The annual mean of the REari at the surface for the 25 DWD stations shows a value of −12.8 ± 5 W m−2 as the average over the clear-sky models, compared to −11 W m−2 from the radiative transfer simulations. Since all models assume a fixed aerosol characterization, the annual cycle of the aerosol radiation effect cannot be reproduced. Out of this set of clear-sky models, the largest level of agreement is shown by the ESRA and MRM v6.1 models.
Bei der Übertragung und Speicherung von Daten ist es eine wesentliche Frage, inwieweit die Daten komprimiert werden können, ohne dass deren Informationsgehalt verloren geht.
Ein Maß für den Informationsgehalt von Daten ist also von grundlegender Bedeutung. Vor etwa siebzig Jahren hat C. E. Shannon ein solches Maß eingeführt und damit das Lehr- und Forschungsgebiet der Informationstheorie begründet, welches seit dem bis heute hin wesentlich zur Konzeption und Realisierung von Informationsund Kommunikationstechnologien beigetragen hat. Etwa zwanzig Jahre später hat A. N. Kolmogorov ein anderes Maß für den Informationsgehalt von Daten eingeführt. Während die Shannonsche Informationstheorie zum Curriculum von mathematischen, informatischen und elektrotechnischen Studiengängen gehört, ist die Algorithmische Informationstheorie von Kolmogorov weit weniger bekannt und eher Gegenstand von speziellen Lehrveranstaltungen.
Seit einigen Jahren nimmt allerdings die Beschäftigung mit dieser Theorie zu, zumal in der einschlägigen Literatur von erfolgreichen praktischen Anwendungen der Theorie berichtet wird. Die vorliegende Arbeit gibt eine Einführung in grundlegende Ideen dieser Theorie und beschreibt deren Anwendungsmöglichkeiten bei einigen ausgewählten Problemstellungen der Theoretischen Informatik.
Die Ausarbeitung kann als Skript für einführende Lehrveranstaltungen in die Algorithmische Informationstheorie sowie als Lektüre zur Einarbeitung in die Thematik als Ausgangspunkt für Forschungs- und Entwicklungsarbeiten verwendet werden.
The development of mobile robotic systems is a demanding task regarding its complexity, required resources and skills in multiple fields such as software development, artificial intelligence, mechanical design, electrical engineering, signal processing, sensor technology or control theory. This holds true particularly for soccer playing robots, where additional aspects like high dynamics, cooperation and high physical stress have to be dealt with. In robot competitions such as RoboCup, additional skills in the domains of team, project and knowledge management are of importance.
Der Beitrag untersucht, wie ein Präsenzlabor durch ein Remote-Labor ergänzt undersetzt werden kann. Dazu wird das Laborpraktikum Digitaltechnik der Hochschule Bonn-Rhein-Sieg betrachtet, bei dem ein Remote-Labor Flexibilität bei der Versuchsdurchführung bietet und Versuche ermöglicht, die allein mit dem Präsenzlabor nicht möglich wären. Neben der Ergänzung der Präsenzversuche können Studie-rende das Praktikum auch komplett im Remote-Labor durchführen. Durch klare Anforderungen an die Erteilung eines Testats ist dies sowohl für sie als auch für Lehrende praktikabel zu handhaben. Rückmeldungen der Studierenden und Nutzungszahlen belegen die Akzeptanz des Remote-Labors. Dabei zeigt sich, dass die Studierenden sehr heterogen mit dem Remote-Labor umgehen: Einige von ihnen nutzen das Remote-Labor als zusätzliche Praktikumszeit für Versuche die auch im Präsenzlabor möglich wären; andere nutzen es als Erweiterung der Praktikumsmöglichkeit für Versuche, die nur im Remote-Labor möglich sind und wieder andere arbeiten intensiv im Remote-Labor und reichen auch das Praktikumsprotokoll elektronisch ein. Für Lehrende besteht über das Protokoll und die Auswertung der Nutzungsdaten ausreichende Sicherheit, um aktive Beteiligung am Praktikum zu testieren.
Seit 2012 wird an der Hochschule Bonn-Rhein-Sieg die Studieneingangsphase im Qualitätspakt Lehre gefördert. Ein wesentliches Anliegen im Projekt „Pro-MINT-us“ ist die Einbeziehung der gesamten Hochschule, um keine isolierten Maßnahmen anzubieten, sondern die im Projekt entwickelten Lehrideen nachhaltig zu verankern.
Improving the study entry supports students in a decisive phase of their university education. Implementing improvements is a change process and can only be successful if the relevant stakeholders are addressed and convinced. In the described Teaching Quality Pact project evaluation data is used as a mean to discuss in the university the situation of the study programs. As these discussions were based on empirical data rather than on opinion, it was possible to achieve an open discussion about measures that are implemented. The open discussion is maintained during the project when results of the measures taken are analyzed.
Low power dissipation is a current topic in digital design, and therefore, it should be covered in a state-of-the-art electrical engineering curriculum. This paper describes how low-power design can be addressed within a digital design course. Doing so would be beneficial for both topics because low-power design is not detached from the systems perspective, and the digital design course would be enriched by references to current challenges and applications. Thus, the presented course should serve as an example of how a course can be developed to also teach students about sustainable engineering.
Background: the potency of drugs that interfere with glucose metabolism, i.e., glucose transporters (GLUT) and nicotinamide phosphoribosyltransferase (NAMPT) was analyzed in neuroendocrine tumor (NET, BON-1, and QPG-1 cells) and small cell lung cancer (SCLC, GLC-2, and GLC-36 cells) tumor cell lines. (2) Methods: the proliferation and survival rate of tumor cells was significantly affected by the GLUT-inhibitors fasentin and WZB1127, as well as by the NAMPT inhibitors GMX1778 and STF-31. (3) Results: none of the NET cell lines that were treated with NAMPT inhibitors could be rescued with nicotinic acid (usage of the Preiss–Handler salvage pathway), although NAPRT expression could be detected in two NET cell lines. We finally analyzed the specificity of GMX1778 and STF-31 in NET cells in glucose uptake experiments. As previously shown for STF-31 in a panel NET-excluding tumor cell lines, both drugs specifically inhibited glucose uptake at higher (50 μM), but not at lower (5 μM) concentrations. (4) Conclusions: our data suggest that GLUT and especially NAMPT inhibitors are potential candidates for the treatment of NET tumors.
In thyroid carcinoma cells, the soluble βgalactosidespecific lectin, galectin3, is extra and intracellularly expressed and plays a significant role in thyroid cancer diagnosis. The functional relevance of this molecule, particularly in its extracellular environment however, warrants further elucidation. To gain insight into this topic, the present study characterized principal functional properties of galectin3 in 3 commonly used thyroid carcinoma cell lines (BCPAP, Cal62 and FTC133) that express the molecule intra and extracellulary. Cellintrinsic galectin3 harbors a functional carbohydrate recognition domain as determined by affinity purification. Moreover, cell surface expressed galectin3 can be partially removed by treatment with lactose or asialofetuin, but not with sucrose. Thyroid carcinoma cells adhere to substratebound galectin3 in a βgalactosidespecific manner, whereby only cell adhesion, but not cell migration is promoted. Thus, thyroid tumor cells harbor functional active galectin3 that, inter alia, specifically interacts with cell surfaceexpressed molecular ligands in a βgalactosidedependent manner, whereby the molecule can at least interfere with cell adhesion. The modulation of galectin3 expression level or its ligands in such tumor cells could be of therapeutic interest and needs further experimental clarification.
Stably stratified Taylor–Green vortex simulations are performed by lattice Boltzmann methods (LBM) and compared to other recent works using Navier–Stokes solvers. The density variation is modeled with a separate distribution function in addition to the particle distribution function modeling the flow physics. Different stencils, forcing schemes, and collision models are tested and assessed. The overall agreement of the lattice Boltzmann solutions with reference solutions from other works is very good, even when no explicit subgrid model is used, but the quality depends on the LBM setup. Although the LBM forcing scheme is not decisive for the quality of the solution, the choice of the collision model and of the stencil are crucial for adequate solutions in underresolved conditions. The LBM simulations confirm the suppression of vertical flow motion for decreasing initial Froude numbers. To gain further insight into buoyancy effects, energy decay, dissipation rates, and flux coefficients are evaluated using the LBM model for various Froude numbers.
Turbulent compressible flows are traditionally simulated using explicit time integrators applied to discretized versions of the Navier-Stokes equations. However, the associated Courant-Friedrichs-Lewy condition severely restricts the maximum time-step size. Exploiting the Lagrangian nature of the Boltzmann equation’s material derivative, we now introduce a feasible three-dimensional semi-Lagrangian lattice Boltzmann method (SLLBM), which circumvents this restriction. While many lattice Boltzmann methods for compressible flows were restricted to two dimensions due to the enormous number of discrete velocities in three dimensions, the SLLBM uses only 45 discrete velocities. Based on compressible Taylor-Green vortex simulations we show that the new method accurately captures shocks or shocklets as well as turbulence in 3D without utilizing additional filtering or stabilizing techniques other than the filtering introduced by the interpolation, even when the time-step sizes are up to two orders of magnitude larger compared to simulations in the literature. Our new method therefore enables researchers to study compressible turbulent flows by a fully explicit scheme, whose range of admissible time-step sizes is dictated by physics rather than spatial discretization.
This work thoroughly investigates a semi-Lagrangian lattice Boltzmann (SLLBM) solver for compressible flows. In contrast to other LBM for compressible flows, the vertices are organized in cells, and interpolation polynomials up to fourth order are used to attain the off-vertex distribution function values. Differing from the recently introduced Particles on Demand (PoD) method , the method operates in a static, non-moving reference frame. Yet the SLLBM in the present formulation grants supersonic flows and exhibits a high degree of Galilean invariance. The SLLBM solver allows for an independent time step size due to the integration along characteristics and for the use of unusual velocity sets, like the D2Q25, which is constructed by the roots of the fifth-order Hermite polynomial. The properties of the present model are shown in diverse example simulations of a two-dimensional Taylor-Green vortex, a Sod shock tube, a two-dimensional Riemann problem and a shock-vortex interaction. It is shown that the cell-based interpolation and the use of Gauss-Lobatto-Chebyshev support points allow for spatially high-order solutions and minimize the mass loss caused by the interpolation. Transformed grids in the shock-vortex interaction show the general applicability to non-uniform grids.
Turbulent compressible flows are traditionally simulated using explicit Eulerian time integration applied to the Navier-Stokes equations. However, the associated Courant-Friedrichs-Lewy condition severely restricts the maximum time step size. Exploiting the Lagrangian nature of the Boltzmann equation's material derivative, we now introduce a feasible three-dimensional semi-Lagrangian lattice Boltzmann method (SLLBM), which elegantly circumvents this restriction. Previous lattice Boltzmann methods for compressible flows were mostly restricted to two dimensions due to the enormous number of discrete velocities needed in three dimensions. In contrast, this Rapid Communication demonstrates how cubature rules enhance the SLLBM to yield a three-dimensional velocity set with only 45 discrete velocities. Based on simulations of a compressible Taylor-Green vortex we show that the new method accurately captures shocks or shocklets as well as turbulence in 3D without utilizing additional filtering or stabilizing techniques, even when the time step sizes are up to two orders of magnitude larger compared to simulations in the literature. Our new method therefore enables researchers for the first time to study compressible turbulent flows by a fully explicit scheme, whose range of admissible time step sizes is only dictated by physics, while being decoupled from the spatial discretization.
Off-lattice Boltzmann methods increase the flexibility and applicability of lattice Boltzmann methods by decoupling the discretizations of time, space, and particle velocities. However, the velocity sets that are mostly used in off-lattice Boltzmann simulations were originally tailored to on-lattice Boltzmann methods. In this contribution, we show how the accuracy and efficiency of weakly and fully compressible semi-Lagrangian off-lattice Boltzmann simulations is increased by velocity sets derived from cubature rules, i.e. multivariate quadratures, which have not been produced by the Gauß-product rule. In particular, simulations of 2D shock-vortex interactions indicate that the cubature-derived degree-nine D2Q19 velocity set is capable to replace the Gauß-product rule-derived D2Q25. Likewise, the degree-five velocity sets D3Q13 and D3Q21, as well as a degree-seven D3V27 velocity set were successfully tested for 3D Taylor–Green vortex flows to challenge and surpass the quality of the customary D3Q27 velocity set. In compressible 3D Taylor–Green vortex flows with Mach numbers on-lattice simulations with velocity sets D3Q103 and D3V107 showed only limited stability, while the off-lattice degree-nine D3Q45 velocity set accurately reproduced the kinetic energy provided by literature.
The identification of energetic materials in containments is an important challenge for analytical methods in the field of safety and security. Opening a package without knowledge of its contents and the resulting hazards is highly involved with risks and should be avoided whenever possible. Therefore, preferable methods work non-destructive with minimal interaction and are capable of identifying target substances in a containment quickly and reliably. Most spectroscopic methods find their limits, if the target substance is shielded by a covering material. To solve this problem, a combined laser drilling method with subsequent identification of the target substance by means of Raman spectroscopic measurements through microscopic bore holes of the covering material is presented. A pulsed laser beam is used for both the drilling process and as an excitation source for Raman measurements in the same optical setup. Results show the ability of this new method to gain high-quality spectra even when performed through microscopic small bore channels. With the laser parameters chosen right, the method can even be performed on highly sensitive explosives like triacetone triperoxide (TATP). Another advantageous effect arises in an observed reduction in unwanted fluorescence signal in the spectral data, resulting from the confocal-like measurement setup with the bore hole acting as aperture.
Die vorliegende Forschungsarbeit setzt sich mit nachhaltigem Verhalten in Bezug auf die Nutzung von Kaffeebehältern an der HBRS auseinander. Anlass dafür ist, dass Pappbecher aufgrund einer Plastikbeschichtung nur schwer recycelbar sind und somit die Umwelt erheblich beeinträchtigen. In diesem Zusammenhang nahmen 204 Studierende an einer Online-Befragung teil. Den Ergebnissen zufolge kommen derzeit vor allem Einweg-Pappbecher zum Einsatz. Zur Modifizierung dieses umweltschädlichen Verhaltens bedarf es an geeigneten Interventionsstrategien. Basierend auf den Ergebnissen sind Maßnahmen zu implementieren, die dem Defizit an Handlungswissen und dem hohen Aufwand entgegenwirken, welcher mit der Verwendung eigens mitgebrachter Becher und den vorhandenen Porzellantassen assoziiert wird. Nach Sicherstellung der ökologischen Vorteile und finanziellen Umsetzbarkeit sollte das bestehende Pfandsystem um praktischere Becher sowie flexible Rückgabemöglichkeiten erweitert werden. Unterstützend ist eine Belohnung in Form von Freigetränken oder einem geringen finanziellen Rabatt sinnvoll, um den automatischen Verbrauch von Pappbechern zu unterbinden.
Digital ecosystems are driving the digital transformation of business models. Meanwhile, the associated processing of personal data within these complex systems poses challenges to the protection of individual privacy. In this paper, we explore these challenges from the perspective of digital ecosystems' platform providers. To this end, we present the results of an interview study with seven data protection officers representing a total of 12 digital ecosystems in Germany. We identified current and future challenges for the implementation of data protection requirements, covering issues on legal obligations and data subject rights. Our results support stakeholders involved in the implementation of privacy protection measures in digital ecosystems, and form the foundation for future privacy-related studies tailored to the specifics of digital ecosystems.
Risk-based authentication (RBA) extends authentication mechanisms to make them more robust against account takeover attacks, such as those using stolen passwords. RBA is recommended by NIST and NCSC to strengthen password-based authentication, and is already used by major online services. Also, users consider RBA to be more usable than two-factor authentication and just as secure. However, users currently obtain RBA's high security and usability benefits at the cost of exposing potentially sensitive personal data (e.g., IP address or browser information). This conflicts with user privacy and requires to consider user rights regarding the processing of personal data. We outline potential privacy challenges regarding different attacker models and propose improvements to balance privacy in RBA systems. To estimate the properties of the privacy-preserving RBA enhancements in practical environments, we evaluated a subset of them with long-term data from 780 users of a real-world online service. Our results show the potential to increase privacy in RBA solutions. However, it is limited to certain parameters that should guide RBA design to protect privacy. We outline research directions that need to be considered to achieve a widespread adoption of privacy preserving RBA with high user acceptance.
Risk-based Authentication (RBA) is an adaptive security measure that improves the security of password-based authentication by protecting against credential stuffing, password guessing, or phishing attacks. RBA monitors extra features during login and requests for an additional authentication step if the observed feature values deviate from the usual ones in the login history. In state-of-the-art RBA re-authentication deployments, users receive an email with a numerical code in its body, which must be entered on the online service. Although this procedure has a major impact on RBA's time exposure and usability, these aspects were not studied so far.
We introduce two RBA re-authentication variants supplementing the de facto standard with a link-based and another code-based approach. Then, we present the results of a between-group study (N=592) to evaluate these three approaches. Our observations show with significant results that there is potential to speed up the RBA re-authentication process without reducing neither its security properties nor its security perception. The link-based re-authentication via "magic links", however, makes users significantly more anxious than the code-based approaches when perceived for the first time. Our evaluations underline the fact that RBA re-authentication is not a uniform procedure. We summarize our findings and provide recommendations.
Risk-based authentication (RBA) aims to protect users against attacks involving stolen passwords. RBA monitors features during login, and requests re-authentication when feature values widely differ from those previously observed. It is recommended by various national security organizations, and users perceive it more usable than and equally secure to equivalent two-factor authentication. Despite that, RBA is still used by very few online services. Reasons for this include a lack of validated open resources on RBA properties, implementation, and configuration. This effectively hinders the RBA research, development, and adoption progress.
To close this gap, we provide the first long-term RBA analysis on a real-world large-scale online service. We collected feature data of 3.3 million users and 31.3 million login attempts over more than 1 year. Based on the data, we provide (i) studies on RBA’s real-world characteristics plus its configurations and enhancements to balance usability, security, and privacy; (ii) a machine learning–based RBA parameter optimization method to support administrators finding an optimal configuration for their own use case scenario; (iii) an evaluation of the round-trip time feature’s potential to replace the IP address for enhanced user privacy; and (iv) a synthesized RBA dataset to reproduce this research and to foster future RBA research. Our results provide insights on selecting an optimized RBA configuration so that users profit from RBA after just a few logins. The open dataset enables researchers to study, test, and improve RBA for widespread deployment in the wild.
Risk-based authentication (RBA) aims to strengthen password-based authentication rather than replacing it. RBA does this by monitoring and recording additional features during the login process. If feature values at login time differ significantly from those observed before, RBA requests an additional proof of identification. Although RBA is recommended in the NIST digital identity guidelines, it has so far been used almost exclusively by major online services. This is partly due to a lack of open knowledge and implementations that would allow any service provider to roll out RBA protection to its users. To close this gap, we provide a first in-depth analysis of RBA characteristics in a practical deployment. We observed N=780 users with 247 unique features on a real-world online service for over 1.8 years. Based on our collected data set, we provide (i) a behavior analysis of two RBA implementations that were apparently used by major online services in the wild, (ii) a benchmark of the features to extract a subset that is most suitable for RBA use, (iii) a new feature that has not been used in RBA before, and (iv) factors which have a significant effect on RBA performance. Our results show that RBA needs to be carefully tailored to each online service, as even small configuration adjustments can greatly impact RBA's security and usability properties. We provide insights on the selection of features, their weightings, and the risk classification in order to benefit from RBA after a minimum number of login attempts.
Risk-based Authentication (RBA) is an adaptive security measure to strengthen password-based authentication. RBA monitors additional features during login, and when observed feature values differ significantly from previously seen ones, users have to provide additional authentication factors such as a verification code. RBA has the potential to offer more usable authentication, but the usability and the security perceptions of RBA are not studied well.
We present the results of a between-group lab study (n=65) to evaluate usability and security perceptions of two RBA variants, one 2FA variant, and password-only authentication. Our study shows with significant results that RBA is considered to be more usable than the studied 2FA variants, while it is perceived as more secure than password-only authentication in general and comparably secure to 2FA in a variety of application types. We also observed RBA usability problems and provide recommendations for mitigation. Our contribution provides a first deeper understanding of the users' perception of RBA and helps to improve RBA implementations for a broader user acceptance.
Risk-based authentication (RBA) is an adaptive security measure to strengthen password-based authentication against account takeover attacks. Our study on 65 participants shows that users find RBA more usable than two-factor authentication equivalents and more secure than password-only authentication. We identify pitfalls and provide guidelines for putting RBA into practice.
Risk-based authentication (RBA) aims to strengthen password-based authentication rather than replacing it. RBA does this by monitoring and recording additional features during the login process. If feature values at login time differ significantly from those observed before, RBA requests an additional proof of identification. Although RBA is recommended in the NIST digital identity guidelines, it has so far been used almost exclusively by major online services. This is partly due to a lack of open knowledge and implementations that would allow any service provider to roll out RBA protection to its users.
To close this gap, we provide a first in-depth analysis of RBA characteristics in a practical deployment. We observed N=780 users with 247 unique features on a real-world online service for over 1.8 years. Based on our collected data set, we provide (i) a behavior analysis of two RBA implementations that were apparently used by major online services in the wild, (ii) a benchmark of the features to extract a subset that is most suitable for RBA use, (iii) a new feature that has not been used in RBA before, and (iv) factors which have a significant effect on RBA performance. Our results show that RBA needs to be carefully tailored to each online service, as even small configuration adjustments can greatly impact RBA's security and usability properties. We provide insights on the selection of features, their weightings, and the risk classification in order to benefit from RBA after a minimum number of login attempts.
Studi ini bertujuan untuk memvalidasi perangkat penilaian efikasi diri yang berkaitan dengan kesehatan kerja yang dikembangkan pada tahap studi pendahuluan. Skala Efikasi Diri untuk Kesehatan Kerja (SEDKK) berlandaskan konsep efikasi diri pada teori kognitif sosial yang mengukur empat faktor yang berpengaruh pada kesehatan setiap individu yang bekerja, seperti: perilaku makan dan minum, tidur, keamanan dan kesehatan kerja, serta kegiatan pemulihan dari stres bekerja. Hasil analisis faktor eksploratori menunjukan bahwa ada empat faktor yang terefleksikan dari butir-butir SEDKK. Validitas konstruk SEDKK dapat dibuktikan dengan korelasi positif antara SEDKK dan skala Efikasi Diri Umum yang sangat signifikan. Pengujian validitas kriteria dapat ditelusuri melalui efek SEDKK terhadap kondisi kesehatan umum, kepuasan akan kesehatan pribadi, keseimbangan kehidupan kerja/KKK (work life balance), perilaku sehat, dan perilaku berisiko. Namun demikian, asumsi mengenai reliabilitas tes berulang (test-retest) pada penelitian ini ditolak. Implikasi dan saran-saran untuk penelitian selanjutnya didiskusikan pada artikel ini.
Many workers experience their jobs as effortful or even stressful, which can result in strain. Although recovery from work would be an adaptive strategy to prevent the adverse effects of work-related strain, many workers face problems finding enough time to rest and to mentally disconnect from work during nonwork time. What goes on in workers’ minds after a stressful workday? What is it about their jobs that makes them think about their work? This special issue aims to bridge the gap between research on recovery processes mainly examined in Occupational Health Psychology, and research on work stress and working hours, often investigated in the field of Human Resource Management. We first summarize conceptual and theoretical streams from both fields of research. In the following, we discuss the contributions of the five special issue papers and conclude with key messages and directions for further research.
Recessive mutations in the MPV17 gene cause mitochondrial DNA depletion syndrome, a fatal infantile genetic liver disease in humans. Loss of function in mice leads to glomerulosclerosis and sensineural deafness accompanied with mitochondrial DNA depletion. Mutations in the yeast homolog Sym1, and in the zebra fish homolog tra cause interesting, but not obviously related phenotypes, although the human gene can complement the yeast Sym1 mutation. The MPV17 protein is a hydrophobic membrane protein of 176 amino acids and unknown function. Initially localised in murine peroxisomes, it was later reported to be a mitochondrial inner membrane protein in humans and in yeast. To resolve this contradiction we tested two new mouse monoclonal antibodies directed against the human MPV17 protein in Western blots and immunohistochemistry on human U2OS cells. One of these monoclonal antibodies showed specific reactivity to a protein of 20 kD absent in MPV17 negative mouse cells. Immunofluorescence studies revealed colocalisation with peroxisomal, endosomal and lysosomal markers, but not with mitochondria. This data reveal a novel connection between a possible peroxisomal/endosomal/lysosomal function and mitochondrial DNA depletion.
Work-related thoughts during off-job time have been studied extensively in occupational health psychology and related fields. We provide a focused review of the research on overcommitment—a component within the effort–reward imbalance model—and aim to connect this line of research to the most commonly studied aspects of work-related rumination. Drawing on this integrative review, we analyze survey data on ten facets of work-related rumination, namely (1) overcommitment, (2) psychological detachment, (3) affective rumination, (4) problem-solving pondering, (5) positive work reflection, (6) negative work reflection, (7) distraction, (8) cognitive irritation, (9) emotional irritation, and (10) inability to recover. First, we apply exploratory factor analysis to self-reported survey data from 357 employees to calibrate overcommitment items and to position overcommitment within the nomological net of work-related rumination constructs. Second, we leverage apply confirmatory factor analysis to self-reported survey data from 388 employees to provide a more specific test of uniqueness vs. overlap among these constructs. Third, we apply relative weight analysis to assess the unique criterion-related validity of each work-related rumination facet regarding (1) physical fatigue, (2) cognitive fatigue, (3) emotional fatigue, (4) burnout, (5) psychosomatic complaints, and (6) satisfaction with life. Our results suggest that several measures of work-related rumination (e.g., overcommitment and cognitive irritation) can be used interchangeably. Emotional irritation and affective rumination emerge as the strongest unique predictors of fatigue, burnout, psychosomatic complaints, and satisfaction with life. Our study is intended to assist researchers in making informed decisions on selecting scales for their research and paves the way for integrating research on the effort–reward imbalance and work-related rumination.
Although work events can be regarded as pivotal elements of organizational life, only a few studies have examined how positive and negative events relate to and combine to affect work engagement over time. Theory suggests that to better understand how current events affect work engagement (WE), we have to account for recent events that have preceded these current events. We present competing theoretical views on how recent and current work events may affect employees (e.g., getting used to a high frequency of negative events or becoming more sensitive to negative events). Although the occurrence of events implies discrete changes in the experience of work, prior research has not considered whether work events actually accumulate to sustained mid-term changes in WE. To address these gaps in the literature, we conducted a week-level longitudinal study across a period of 15 consecutive weeks among 135 employees, which yielded 849 weekly observations. While positive events were associated with higher levels of WE within the same week, negative events were not. Our results support neither satiation nor sensitization processes. However, high frequencies of negative events in the preceding week amplified the beneficial effects of positive events on WE in the current week. Growth curve analyses show that the benefits of positive events accumulate to sustain high levels of WE. WE dissipates in the absence of continuous experience of positive events. Our study adds a temporal component and informs research that has taken a feature-oriented perspective on the dynamic interplay of job demands and resources.
Although work events can be regarded as pivotal elements of organizational life, only a few studies have examined how positive and negative events relate to and combine to affect work engagement over time. Theory suggests that, to better understand how current events affect work engagement (WE), we have to account for recent events that have preceded these current events. We present competing theoretical views on how recent and current work events may affect employees (e.g., getting used to a high frequency of negative events or becoming more sensitive to negative events). Although the occurrence of events implies discrete changes in the experience of work, prior research has not considered whether work events actually accumulate to sustained mid-term changes in WE. To address these gaps in the literature, we conducted a week-level longitudinal study across a period of 15 consecutive weeks among 135 employees, which yielded 849 weekly observations. While positive events were associated with higher levels of WE within the same week, negative events were not. Our results support neither satiation nor sensitization processes. However, a high frequency of negative events in the preceding week amplified the beneficial effects of positive events on WE in the current week. Growth curve analyses show that the benefits of positive events accumulate to sustain high levels of WE. WE dissipates in the absence of a continuous experience of positive events. Our study adds a temporal component by highlighting that positive events affect work engagement, particularly in light of recent negative events. Our study informs research that has taken a feature-oriented perspective on the dynamic interplay of job demands and resources.
In the literature on occupational stress and recovery from work several facets of thinking about work in off-job time have been conceptualized. However, research on the focal concepts is currently rather disintegrated. In this study we take a closer look at the five most established concepts, namely (1) psychological detachment, (2) affective rumination, (3) problem-solving pondering, (4) positive work reflection, and (5) negative work reflection. More specifically, we scrutinized (1) whether the five facets of work-related rumination are empirically distinct, (2) whether they yield differential associations with different facets of employee well-being (burnout, work engagement, thriving, satisfaction with life, and flourishing), and (3) to what extent the five facets can be distinguished from and relate to conceptually similar constructs, such as irritation, worry, and neuroticism. We applied structural equation modeling techniques to cross-sectional survey data from 474 employees. Our results provide evidence that (1) the five facets of work-related rumination are highly related, yet empirically distinct, (2) that each facet contributes uniquely to explain variance in certain aspects of employee well-being, and (3) that they are distinct from related concepts, albeit there is a high overlap between (lower levels of) psychological detachment and cognitive irritation. Our study contributes to clarify the structure of work-related rumination and extends the nomological network around different types of thinking about work in off-job time and employee well-being.
In the literature on occupational stress and recovery from work, several facets of thinking about work during off-job time have been conceptualized. However, research on the focal concepts is currently rather diffuse. In this study we take a closer look at the five most well-established concepts: (1) psychological detachment, (2) affective rumination, (3) problem-solving pondering, (4) positive work reflection, and (5) negative work reflection. More specifically, we scrutinized (1) whether the five facets of work-related rumination are empirically distinct, (2) whether they yield differential associations with different facets of employee well-being (burnout, work engagement, thriving, satisfaction with life, and flourishing), and (3) to what extent the five facets can be distinguished from and relate to conceptually similar constructs, such as irritation, worry, and neuroticism. We applied structural equation modeling techniques to cross-sectional survey data from 474 employees. Our results provide evidence for (1) five correlated, yet empirically distinct facets of work-related rumination. (2) Each facet yields a unique pattern of association with the eight aspects of employee well-being. For instance, detachment is strongly linked to satisfaction with life and flourishing. Affective rumination is linked particularly to burnout. Problem-solving pondering and positive work reflection yield the strongest links to work engagement. (3) The five facets of work-related rumination are distinct from related concepts, although there is a high overlap between (lower levels of) psychological detachment and cognitive irritation. Our study contributes to clarifying the structure of work-related rumination and extends the nomological network around different types of thinking about work during off-job time and employee well-being.