Refine
H-BRS Bibliography
- yes (3723) (remove)
Departments, institutes and facilities
- Fachbereich Wirtschaftswissenschaften (1089)
- Fachbereich Informatik (970)
- Fachbereich Angewandte Naturwissenschaften (577)
- Fachbereich Ingenieurwissenschaften und Kommunikation (512)
- Fachbereich Sozialpolitik und Soziale Sicherung (359)
- Institut für Technik, Ressourcenschonung und Energieeffizienz (TREE) (347)
- Institute of Visual Computing (IVC) (272)
- Institut für funktionale Gen-Analytik (IFGA) (169)
- Institut für Verbraucherinformatik (IVI) (141)
- Institut für Cyber Security & Privacy (ICSP) (106)
Document Type
- Article (1168)
- Conference Object (1017)
- Part of a Book (618)
- Book (monograph, edited volume) (365)
- Report (91)
- Contribution to a Periodical (81)
- Doctoral Thesis (70)
- Preprint (69)
- Working Paper (45)
- Master's Thesis (43)
Year of publication
Has Fulltext
- no (3723) (remove)
Keywords
- Lehrbuch (85)
- Controlling (23)
- Deutschland (23)
- Nachhaltigkeit (23)
- Unternehmen (21)
- Management (17)
- Betriebswirtschaftslehre (16)
- Informationstechnik (13)
- Robotics (13)
- FPGA (12)
An essential measure of autonomy in assistive service robots is adaptivity to the various contexts of human-oriented tasks, which are subject to subtle variations in task parameters that determine optimal behaviour. In this work, we propose an apprenticeship learning approach to achieving context-aware action generalization on the task of robot-to-human object hand-over. The procedure combines learning from demonstration and reinforcement learning: a robot first imitates a demonstrator’s execution of the task and then learns contextualized variants of the demonstrated action through experience. We use dynamic movement primitives as compact motion representations, and a model-based C-REPS algorithm for learning policies that can specify hand-over position, conditioned on context variables. Policies are learned using simulated task executions, before transferring them to the robot and evaluating emergent behaviours. We additionally conduct a user study involving participants assuming different postures and receiving an object from a robot, which executes hand-overs by either imitating a demonstrated motion, or adapting its motion to hand-over positions suggested by the learned policy. The results confirm the hypothesized improvements in the robot’s perceived behaviour when it is context-aware and adaptive, and provide useful insights that can inform future developments.
Machine learning-based solutions are frequently adapted in several applications that require big data in operations. The performance of a model that is deployed into operations is subject to degradation due to unanticipated changes in the flow of input data. Hence, monitoring data drift becomes essential to maintain the model’s desired performance. Based on the conducted review of the literature on drift detection, statistical hypothesis testing enables to investigate whether incoming data is drifting from training data. Because Maximum Mean Discrepancy (MMD) and Kolmogorov-Smirnov (KS) have shown to be reliable distance measures between multivariate distributions in the literature review, both were selected from several existing techniques for experimentation. For the scope of this work, the image classification use case was experimented with using the Stream-51 dataset. Based on the results from different drift experiments, both MMD and KS showed high Area Under Curve values. However, KS exhibited faster performance than MMD with fewer false positives. Furthermore, the results showed that using the pre-trained ResNet-18 for feature extraction maintained the high performance of the experimented drift detectors. Furthermore, the results showed that the performance of the drift detectors highly depends on the sample sizes of the reference (training) data and the test data that flow into the pipeline’s monitor. Finally, the results also showed that if the test data is a mixture of drifting and non-drifting data, the performance of the drift detectors does not depend on how the drifting data are scattered with the non-drifting ones, but rather their amount in the test set
The design of an efficient digital circuit in term of low-power has become a very challenging issue. For this reason, low-power digital circuit design is a topic addressed in electrical and computer engineering curricula, but it also requires practical experiments in a laboratory. This PhD research investigates a novel approach, the low-power design laboratory system by developing a new technical and pedagogical system. The low-power design laboratory system is composed of two types of laboratories: the on-site (hands-on) laboratory and the remote laboratory. It has been developed at the Bonn-Rhine-Sieg University of Applied Sciences to teach low-power techniques in the laboratory. Additionally, this thesis contributes a suggestion on how the learning objectives can be complemented by developing a remote system in order to improve the teaching process of the low-power digital circuit design. This laboratory system enables online experiments that can be performed using physical instruments and obtaining real data via the internet. The laboratory experiments use a Field Programmable Gate Array (FPGA) as a design platform for circuit implementation by students and use image processing as an application for teaching low-power techniques.
This thesis presents the instructions for the low-power design experiments which use a top-down hierarchical design methodology. The engineering student designs his/her algorithm with a high level of abstraction and the experimental results are obtained and measured at a low level (hardware) so that more information is available to correctly estimate the power dissipation such as specification, latency, thermal effect, and technology used. Power dissipation of the digital system is influenced by specification, design, technology used, as well as operating temperature. Digital circuit designers can observe the most influential factors in power dissipation during the laboratory exercises in the on-site system and then use the remote system to supplement investigating the other factors. Furthermore, the remote system has obvious benefits such as developing learning outcomes, facilitating new teaching methods, reducing costs and maintenance, cost-saving by reducing the numbers of instructors, saving instructor time and simplifying their tasks, facilitating equipment sharing, improving reliability, and finally providing flexibility of usage the laboratories.
For many different applications, current information about the bandwidth-related metrics of the utilized connection is very useful as they directly impact the performance of throughput sensitive applications such as streaming servers, IPTV and VoIP applications. In literature, several tools have been proposed to estimate major bandwidth-related metrics such as capacity, available bandwidth and achievable throughput. The vast majority of these tools fall into one of Packet Pair (PP), Variable Packet Size (VPS), Self-Loading of Periodic Streams (SLoPS) or Throughput approaches. In this study, seven popular bandwidth estimation tools including nettimer, pathrate, pathchar, pchar, clink, pathload and iperf belonging to these four well-known estimation techniques are presented and experimentally evaluated in a controlled testbed environment. Differently from the rest of studies in literature, all tools have been uniformly classified and evaluated according to an objective and sophisticated classification and evaluation scheme. The performance comparison of the tools incorporates not only the estimation accuracy but also the probing time and overhead caused.
Queueing Theory
(2024)
Entrepreneurship education serves a conduit for new venture creation as it provides the knowledge and skills needed to increase the self-efficacy of individuals to start and run new businesses and to grow existing ones. This study, therefore, sought to assess the relationship between the approaches to the teaching of entrepreneur-ship and entrepreneurial intention on a cohort of 292 respondents consisting of students who have studied entrepreneurship in three selected Universities. A structured questionnaire was used to obtain data randomly from students. The canonical correlation results indicate that education for and through entrepreneurship is the best approach to promoting entrepreneurial intensity among University students, if the aim of teaching entrepreneur-ship is to promote start-up activities. The findings provide valuable insights for institutions of higher learning and policy makers in Ghana with respect to the appropriate methodologies to be adopted in the teaching of entrepreneurship in our universities.
As competition for tourists becomes more global, understanding and accommodating the needs of international tourists, with their different cultural backgrounds, has become increasingly important. This study highlights the variations in tourist industry service--particularly as they relate to different cultures. Specifically, service failures experienced by Japanese and German tourists in the U.S. were categorized using the Critical Incident Technique (CIT). The results were compared with earlier studies of service failures experienced by American consumers in the tourist industry. The sample consists of 128 Japanese and 94 “Germanic” (German, Austrian, Swiss-German) respondents. The Japanese and German sample rated “Inappropriate employee behavior” most significant category of service failure. More than half of these respondents said that, because of the failure, they would avoid the offending U.S. business. This is a much stronger response than an American sample had reported in an earlier study. The implications for managers and researchers are discussed.
Robust Indoor Localization Using Optimal Fusion Filter For Sensors And Map Layout Information
(2014)
This paper presents the b-it-bots RoboCup@Work team and its current hardware and functional architecture for the KUKA youBot robot.We describe the underlying software framework and the developed capabilities required for operating in industrial environments including features such as reliable and precise navigation, flexible manipulation and robust object recognition.
Estimation of Prediction Uncertainty for Semantic Scene Labeling Using Bayesian Approximation
(2018)
With the advancement in technology, autonomous and assisted driving are close to being reality. A key component of such systems is the understanding of the surrounding environment. This understanding about the environment can be attained by performing semantic labeling of the driving scenes. Existing deep learning based models have been developed over the years that outperform classical image processing algorithms for the task of semantic labeling. However, the existing models only produce semantic predictions and do not provide a measure of uncertainty about the predictions. Hence, this work focuses on developing a deep learning based semantic labeling model that can produce semantic predictions and their corresponding uncertainties. Autonomous driving needs a real-time operating model, however the Full Resolution Residual Network (FRRN) [4] architecture, which is found as the best performing architecture during literature search, is not able to satisfy this condition. Hence, a small network, similar to FRRN, has been developed and used in this work. Based on the work of [13], the developed network is then extended by adding dropout layers and the dropouts are used during testing to perform approximate Bayesian inference. The existing works on uncertainties, do not have quantitative metrics to evaluate the quality of uncertainties estimated by a model. Hence, the area under curve (AUC) of the receiver operating characteristic (ROC) curves is proposed and used as an evaluation metric in this work. Further, a comparative analysis about the influence of dropout layer position, drop probability and the number of samples, on the quality of uncertainty estimation is performed. Finally, based on the insights gained from the analysis, a model with optimal configuration of dropout is developed. It is then evaluated on the Cityscape dataset and shown to be outperforming the baseline model with an AUC-ROC of about 90%, while the latter having AUC-ROC of about 80%.
Despite perfect functioning of its internal components, a robot can be unsuccessful in performing its tasks because of unforeseen situations. These situations occur when the behavior of the objects in the robot’s environment deviates from its expected values. For robots, such deviations are exhibited in the form of unknown external faults which prohibit them from performing their tasks successfully. In this work we propose to use naive physics knowledge to reason about such faults in the robotics domain. We propose an approach that uses naive physics concepts to find information about the situations which result in a detected unknown fault. The naive physics knowledge is represented by the physical properties of objects which are formalized in a logical framework. The proposed approach applies a qualitative version of physical laws to these properties for reasoning about the detected fault. By interpreting the reasoning results the robot finds the information about the situations which can cause the fault. We apply the proposed approach to the scenarios in which a robot performs manipulation tasks of picking and placing objects. Results of this application show that naive physics holds great promise for reasoning about unknown ex- ternal faults in robotics.
Due to the use of fossil fuel resources, many environmental problems have been increasingly growing. Thus, the recent research focuses on the use of environment friendly materials from sustainable feedstocks for future fuels, chemicals, fibers and polymers. Lignocellulosic biomass has become the raw material of choice for these new materials. Recently, the research has focused on using lignin as a substitute material in many industrial applications. The antiradical and antimicrobial activity of lignin and lignin-based films are both of great interest for applications such as food packaging additives. DPPH assay was used to determine the antioxidant activity of Kraft lignin compared to Organosolv lignins from different biomasses. The purification procedure of Kraft lignin showed that double-fold selective extraction is the most efficient confirmed by UV-Vis, FTIR, HSQC, 31PNMR, SEC, and XRD. The antioxidant capacity was discussed regarding the biomass source, pulping process, and degree of purification. Lignin obtained from industrial black liquor are compared with beech wood samples: Biomass source influences the DPPH inhibition (softwood > grass) and the TPC (softwood < grass). DPPH inhibition affected by the polarity of the extraction solvent. Following the trend: ethanol > diethylether > acetone. Reduced polydispersity has positive influence on the DPPH inhibition. Storage decreased the DPPH inhibition but increased the TPC values. The DPPH assay was also used to discuss the antiradical activity of HPMC/lignin and HPMC/lignin/chitosan films. In both binary (HPMC/lignin) and ternary (HPMC/lignin/chitosan) systems the 5% addition showed the highest activity and the highest addition had the lowest. Both scavenging activity and antimicrobial activity are dependent on the biomass source; Organosolv of softwood > Kraft of softwood > Organosolv of grass. Lignins and lignin-containing films showed high antimicrobial activities against Gram-positive and Gram-negative bacteria at 35 °C and at low temperatures (0-7 °C). Purification of Kraft lignin has a negative effect on the antimicrobial activity while storage has positive effect. The lignin leaching in the produced films affected the activity positively and the chitosan addition enhances the activity for both Gram-positive and Gram-negative bacteria. Testing the films against food spoilage bacteria that grow at low temperatures revealed the activity of the 30% addition on HPMC/L1 film against both B. thermosphacta and P. fluorescens while L5 was active only against B. thermosphacta. In HPMC/lignin/chitosan films, the 5% addition exhibited activity against both food spoilage bacteria.
Sind kleinere und mittlere Unternehmen (KMU) bereits auf die Digitale Transformation vorbereitet?
(2018)
Eine von den Autoren durchgeführte Untersuchung konnte deutliche Indizien dafür ausmachen, dass viele kleinere und mittlere Unternehmen (KMU) aktuell noch keine ausreichende Reife zur Digitalen Transformation haben. Zur Lösung des Problems wird vorgeschlagen, ein agiles IT-Management-Konzept zu entwickeln, um den IT-Bereich dynamisch und ohne formalen Ballast des klassischen IT-Managements zu steuern.
Multi-Merger-Szenarien als Herausforderung für das IT-Controlling - Checklisten zur IT-Integration
(2006)
Digitalisierung für kleinere und mittlere Unternehmen (KMU): Anforderungen an das IT-Management
(2018)
Mergers and acquisitions take place all over the world and in many industries, typically motivated by corporate politics. While IT management is often not involved in the decision-making, it has to solve a wide range of problems in the post-merger phase. Indeed, merging two or more companies implies not only merging their core businesses, but also creating a new and efficiently integrated IT organisation from the individual ones, since persistence of the current IT organisations usually does not make sense. In addition, corporate management frequently imposes constraints, e.g., cost reductions, on the IT infrastructure. The principal critical success factor when merging IT organisations is the uninterrupted operation of the IT business, because a service gap is neither acceptable for in-house functional departments nor for external customers. Therefore, the IT rebuilding phase has to focus on IT services that facilitate the processes of functional departments, support processes, and processes of customers and suppliers, so that any transformation work is transparent to internal and external customers. In this article we describe a real-world but anonymous case study. Our goals are to highlight the points important for merging IT organisations, and to help decision-makers, particularly in the areas of IT organisation and IT personnel. We focus on the arising organisational and non-technical issues from a management perspective, i.e., the CIO's view, and provide checklists intended to help IT managers to address the most pressing issues. To assist CIOs surviving in the post merger phase, we give check lists for merging IT organisations, check lists for merging IT human resources, check lists for IT budgets and reporting, and assess activities in a merger scenario. IT hardware, software and IT infrastructure as well as running IT projects are not considered in this paper.
IT performance measurement is often associated by chief executive officers with IT cost cutting although IT protects business processes from increasing IT costs. IT cost cutting only endangers the company’s efficiency. This opinion discriminates those who do IT performance measurement in companies as a bean-counter. The present paper describes an integrated reference model for IT performance measurement based on a life cycle model and a performance oriented framework. The presented model was created from a practical point of view. It is designed lank compared with other known concepts and is very appropriate for small and medium enterprises (SME).
A plethora of architectural patterns and elements for developing service-oriented applications can be gathered from the state-of-the-art. Most of these approaches are merely applicable for single-tenant applications. However, less methodical support is provided for scenarios, in which multiple different tenants with varying requirements access the same application stack concurrently. In order to fill this gap, both novel and existing architectural patterns, architectural elements, as well as fundamental design decisions must be considered and integrated into a framework that leverages the devel- opment of multi-tenant application. This paper addresses this demand and presents the SOAdapt framework. It promotes the development of adaptable multi-tenant applications based on a service-oriented architecture that is capable of incorporating specific requirements of new tenants in a flexible manner.
Trueness and precision of milled and 3D printed root-analogue implants: A comparative in vitro study
(2023)
The need for innovation around the control functions of inverters is great. PV inverters were initially expected to be passive followers of the grid and to disconnect as soon as abnormal conditions happened. Since future power systems will be dominated by generation and storage resources interfaced through inverters these converters must move from following to forming and sustaining the grid. As “digital natives” PV inverters can also play an important role in the digitalisation of distribution networks. In this short review we identified a large potential to make the PV inverter the smart local hub in a distributed energy system. At the micro level, costs and coordination can be improved with bidirectional inverters between the AC grid and PV production, stationary storage, car chargers and DC loads. At the macro level the distributed nature of PV generation means that the same devices will support both to the local distribution network and to the global stability of the grid. Much success has been obtained in the former. The later remains a challenge, in particular in terms of scaling. Yet there is some urgency in researching and demonstrating such solutions. And while digitalisation offers promise in all control aspects it also raises significant cybersecurity concerns.
In 1991 the researchers at the center for the Learning Sciences of Carnegie Mellon University were confronted with the confusing question of “where is AI” from the users, who were interacting with AI but did not realize it. Three decades of research and we are still facing the same issue with the AItechnology users. In the lack of users’ awareness and mutual understanding of AI-enabled systems between designers and users, informal theories of the users about how a system works (“Folk theories”) become inevitable but can lead to misconceptions and ineffective interactions. To shape appropriate mental models of AI-based systems, explainable AI has been suggested by AI practitioners. However, a profound understanding of the current users’ perception of AI is still missing. In this study, we introduce the term “Perceived AI” as “AI defined from the perspective of its users”. We then present our preliminary results from deep-interviews with 50 AItechnology users, which provide a framework for our future research approach towards a better understanding of PAI and users’ folk theories.
For most people, using their body to authenticate their identity is an integral part of daily life. From our fingerprints to our facial features, our physical characteristics store the information that identifies us as "us." This biometric information is becoming increasingly vital to the way we access and use technology. As more and more platform operators struggle with traffic from malicious bots on their servers, the burden of proof is on users, only this time they have to prove their very humanity and there is no court or jury to judge, but an invisible algorithmic system. In this paper, we critique the invisibilization of artificial intelligence policing. We argue that this practice obfuscates the underlying process of biometric verification. As a result, the new "invisible" tests leave no room for the user to question whether the process of questioning is even fair or ethical. We challenge this thesis by offering a juxtaposition with the science fiction imagining of the Turing test in Blade Runner to reevaluate the ethical grounds for reverse Turing tests, and we urge the research community to pursue alternative routes of bot identification that are more transparent and responsive.
Technological objects present themselves as necessary, only to become obsolete faster than ever before. This phenomenon has led to a population that experiences a plethora of technological objects and interfaces as they age, which become associated with certain stages of life and disappear thereafter. Noting the expanding body of literature within HCI about appropriation, our work pinpoints an area that needs more attention, “outdated technologies.” In other words, we assert that design practices can profit as much from imaginaries of the future as they can from reassessing artefacts from the past in a critical way. In a two-week fieldwork with 37 HCI students, we gathered an international collection of nostalgic devices from 14 different countries to investigate what memories people still have of older technologies and the ways in which these memories reveal normative and accidental use of technological objects. We found that participants primarily remembered older technologies with positive connotations and shared memories of how they had adapted and appropriated these technologies, rather than normative uses. We refer to this phenomenon as nostalgic reminiscence. In the future, we would like to develop this concept further by discussing how nostalgic reminiscence can be operationalized to stimulate speculative design in the present.
When dialogues with voice assistants (VAs) fall apart, users often become confused or even frustrated. To address these issues and related privacy concerns, Amazon recently introduced a feature allowing Alexa users to inquire about why it behaved in a certain way. But how do users perceive this new feature? In this paper, we present preliminary results from research conducted as part of a three-year project involving 33 German households. This project utilized interviews, fieldwork, and co-design workshops to identify common unexpected behaviors of VAs, as well as users’ needs and expectations for explanations. Our findings show that, contrary to its intended purpose, the new feature actually exacerbates user confusion and frustration instead of clarifying Alexa's behavior. We argue that such voice interactions should be characterized as explanatory dialogs that account for VA’s unexpected behavior by providing interpretable information and prompting users to take action to improve their current and future interactions.
AI (artificial intelligence) systems are increasingly being used in all aspects of our lives, from mundane routines to sensitive decision-making and even creative tasks. Therefore, an appropriate level of trust is required so that users know when to rely on the system and when to override it. While research has looked extensively at fostering trust in human-AI interactions, the lack of standardized procedures for human-AI trust makes it difficult to interpret results and compare across studies. As a result, the fundamental understanding of trust between humans and AI remains fragmented. This workshop invites researchers to revisit existing approaches and work toward a standardized framework for studying AI trust to answer the open questions: (1) What does trust mean between humans and AI in different contexts? (2) How can we create and convey the calibrated level of trust in interactions with AI? And (3) How can we develop a standardized framework to address new challenges?
In the fermentation process sugars are transformed into lactic acid. pH meters have traditionally been used for fermentation process monitoring based on acidity. More recently, near infrared (NIR) spectroscopy has proven to provide an accurate and non-invasive method to detect when the transformation of sugars into lactic acid is finished. The fermentation process when sugars are transformed into lactic acid. This research proposes the use of simplified NIR spectroscopy using multispectral optical sensors as a simpler and less expensive measure to end the fermentation process. The NIR spectrum of milk and yogurt is compared to find and extract features that can be used to design a simple sensor to monitor the yogurt fermentation process. Multispectral images in four selected wavebands within the NIR spectrum are captured and show different spectral remission characteristics for milk, yogurt and water, which support the selection of these wavebands for milk and yogurt classification.
Information reliability and automatic computation are two important aspects that are continuously pushing the Web to be more semantic. Information uploaded to the Web should be reusable and extractable automatically to other applications, platforms, etc. Several tools exist to explicitly markup Web content. The Web services may also have a positive role on the automatic processing of Web contents, especially when they act as flexible and agile agents. However, Web services themselves should be developed with semantics in mind. They should include and provide structured information to facilitate their use, reuse, composition, query, etc. In this chapter, the authors focus on evaluating state-of-the-art semantic aspects and approaches in Web services. Ultimately, this contributes to the goal of Web knowledge management, execution, and transfer.
Software testing in web services environment faces different challenges in comparison with testing in traditional software environments. Regression testing activities are triggered based on software changes or evolutions. In web services, evolution is not a choice for service clients. They have always to use the current updated version of the software. In addition test execution or invocation is expensive in web services and hence providing algorithms to optimize test case generation and execution is vital. In this environment, we proposed several approach for test cases' selection in web services' regression testing. Testing in this new environment should evolve to be included part of the service contract. Service providers should provide data or usage sessions that can help service clients reduce testing expenses through optimizing the selected and executed test cases.
ENaC channels
(2023)
Introduction: After cellulose, lignin represents the most abundant biopolymer on earth that accounts for up to 18-35 % by weight of lignocellulose biomass. Today, it is a by-product of the paper and pulping industry. Although lignin is available in huge amounts, mainly in form of so called black liquor produced via Kraft-pulping, processes for the valorization of lignin are still limited [1]. Due to its hyperbranched polyphenol-like structure, lignin gained increasing interest as biobased building block for polymer synthesis [2]. The present work is focused on extraction and purification of lignin from industrial black liquor and synthesis of lignin-based polyurethanes.
Once aberrantly activated, the Wnt/βcatenin pathway may result in uncontrolled proliferation and eventually cancer. Efforts to counter and inhibit this pathway are mainly directed against βcatenin, as it serves a role on the cytoplasm and the nucleus. In addition, speciallygenerated lymphocytes are recruited for the purpose of treating liver cancer. Peripheral blood mononuclear lymphocytes are expanded by the timely addition of interferon γ, interleukin (IL)1β, IL2 and anticluster of differentiation 3 antibody. The resulting cells are called cytokineinduced killer (CIK) cells. The present study utilised these cells and combine them with drugs inhibiting the Wnt pathway in order to examine whether this resulted in an improvement in the killing ability of CIK cells against liver cancer cells. Drugs including ethacrynic acid (EA) and ciclopirox olamine (CPX) were determined to be suitable candidates, as determined by previous studies. Drugs were administered on their own and combined with CIK cells and then a cell viability assay was performed. These results suggest that EAtreated cells demonstrated apoptosis and were significantly affected compared with untreated cells. Unlike EA, CPX killed normal and cancerous cells even at low concentrations. Subsequent to combining EA with CIK cells, the potency of killing was increased and a greater number of cells died, which proves a synergistic action. In summary, EA may be used as an antihepatocellular carcinoma drug, while CPX possesses a high toxicity to cancerous as well as to normal cells. It was proposed that EA should be integrated into present therapeutic methods for cancer.
In today’s business, culture plays a vital role or to a high degree influences the attitude, perception and decision making process of an individual. Culture is an unavoidable state of rules and regulations that defines people’s daily life in a particular environment or society. There are plenty examples of business failures or stagnation or failure of joint ventures, on account of the management's inability to recognize cross-cultural challenges and tackle them appropriately.
BWL kompakt für Dummies
(2018)
"BWL kompakt für Dummies" bietet Ihnen eine verständliche Einführung in die Betriebswirtschaftslehre, egal ob Sie sie für die Aus- oder Weiterbildung brauchen oder sich einfach schlau machen wollen. Tobias Amely stellt Ihnen die wesentlichen Elemente und Grundbegriffe der Betriebswirtschaftslehre vor und zeigt die Bezüge zur Unternehmenspraxis auf: Materialwirtschaft, Leistungsbereitstellung und Produktion, Marketing, Investition und Finanzierung, Unternehmensorganisation und -führung, Rechnungswesen und Controlling.
BWL-Formeln für Dummies
(2012)
BWL-Formeln für Dummies
(2018)
Können Sie sich auch Formeln so schlecht merken? Das ist mit diesem handlichen Nachschlagewerk auch gar nicht mehr nötig! Der Autor des Bestsellers "BWL für Dummies" Tobias Amely hat für Sie die wichtigsten BWL-Formeln zusammengestellt. Zu jeder Formel finden Sie auch gleich ein anschauliches Beispiel, wie sie eingesetzt wird und eine Erklärung, wofür man sie eigentlich braucht. "BWL-Formeln für Dummies" ist also viel mehr als eine reine Auflistung von Formeln.
BWL kompakt für Dummies
(2016)
BWL-Klausuren für Dummies
(2019)
Investition und Finanzierung sind wichtige Themen in der Unternehmenspraxis und im Studium der Betriebswirtschaftslehre. Dieses Buch führt Sie anhand anschaulicher Beispiele in die Grundlagen des Themas ein und zeigt die Ziele finanzwirtschaftlichen Handelns auf. Tobias Amely und Christine Immenkötter zeigen Ihnen die Grundzüge der Finanzwirtschaft und stellen Ihnen die wichtigsten Instrumente sowohl der Außen- und Innenfinanzierung als auch des Finanzmanagements vor. Lernen Sie die statische und dynamische Investitionsrechnung kennen und erfahren Sie, was man über Investitionen in Wertpapiere wissen muss. So liefert Ihnen dieses Buch im bewährten ... für Dummies-Stil einen guten und leicht verständlichen Überblick über alle wichtigen Themen der Investition und Finanzierung. (Verlagsangaben)
BWL für Dummies
(2021)
BWL für Dummies
(2009)
BWL für Dummies
(2016)
"BWL für Dummies" führt kompetent, prägnant und umfassend in die Betriebswirtschaftslehre ein. Dabei werden die wesentlichen Elemente der Betriebswirtschaftslehre praxisorientiert vorgestellt und in ihrem Zusammenhang dargestellt. Folgende Themen werden behandelt: Materialwirtschaft, Leistungsbereitstellung und Produktion, Marketing, Investition und Finanzierung, Unternehmensorganisation und -führung, Rechnungswesen, Controlling.
BWL für Dummies
(2013)
Competitions for Benchmarking: Task and Functionality Scoring Complete Performance Assessment
(2015)
LiDAR-based Indoor Localization with Optimal Particle Filters using Surface Normal Constraints
(2023)
Die Darlegungsform von Nachhaltigkeitsleistungen sind üblicherweise gesondert ausgegebene Nachhaltigkeitsberichte auf die DAX-30-Unternehmen in ihren Geschäftsberichten hinweisen. Insbesondere die für kapitalmarktorientierte Unternehmen bedeutende Stakeholdergruppe der Analysten und Investoren fordert jedoch zunehmend eine integrative Darstellung aller Dimensionen der Triple-Bottom-Line auch im Lagebericht. Die gesetzlichen Offenlegungspflichten nach § 289 Abs. 3 bzw. § 315 Abs. 1 S. 4 HGB und DRS 15.32 erhöhen den Druck auf die DAX-30-Unternehmen zusätzlich. Diese Arbeit thematisiert im Kern das aktuelle Spannungsfeld zwischen Ökonomie, Ökologie und sozialem Engagement von Unternehmen. Auf Basis einer umfassenden theoretischen Analyse werden konkrete Kennzahlen zur Erfassung von Indikatoren der Nachhaltigkeit gebildet und deren Ausprägung bei den DAX 30 Unternehmen erarbeitet.
Die Debatte um das menschliche Erkenntnisvermögen, also die Frage nach der Art und Weise, wie Menschen Wissen und Erkenntnis erlangen, ist nicht neu, sondern sie stellt sich seitdem philosophische Fragen gestellt werden – ohne dass allerdings über die Jahrhunderte hinweg eine definitive Antwort auf diese Frage gefunden werden konnte.
Ausgangspunkt unserer Überlegungen ist die Feststellung, dass die Legitimierung moderner Formen des Wissens mit dem Verlust von legitimierenden Metaerzählungen einhergeht. Diese Feststellung bezieht sich nicht nur ganz allgemein auf die klassischen Geistes- und Sozialwissenschaften, sondern auch konkret auf die angewandte Management- und Organisationsforschung. Traditionell werden diese untergeordneten Diskursarten durch den übergeordneten Diskurs der Aufklärung legitimiert und unterwerfen sich dem Diktat der Rationalität des Modernismus (Ant 2004).
„Ein Wort gibt das andere“ – zwischenmenschliche Kommunikation folgt bestimmten Regeln. Wer diese Mechanismen durchschaut, kann nicht nur eigene Gesprächsziele besser erreichen, sondern auch andere Menschen leichter verstehen und erfolgreicher mit ihnen interagieren.
Nicht von ungefähr zählt Kommunikationskompetenz zu den gefragtesten Soft Skills in Beruf und Alltag. Diese Einführung in die Theorie und Praxis der Kommunikation erläutert die Prinzipien effizienter Kommunikation nach wie vor ein Klassiker zur Veranschaulichung von gruppendynamischen Prozessen und Rollenverhalten. Das Lehrbuch erklärt das Phänomen der Kommunikation anhand verschiedener sozialpsychologischer Untersuchungen, Theorien, Beispiele und Sichtweisen, regt zu einer erweiterten Reflexion darüber an und liefert konkrete Hinweise und Übungen, welche die eigene Kommunikationspraxis effektiv verbessern.
(Verlagsangaben)
Nowadays, we input text not only on stationary devices, but also on handheld devices while walking, driving, or commuting. Text entry on the move, which we term as nomadic text entry, is generally slower. This is partially due to the need for users to move their visual focus from the device to their surroundings for navigational purposes and back. To investigate if better feedback about users' surroundings on the device can improve performance, we present a number of new and existing feedback systems: textual, visual, textual & visual, and textual & visual via translucent keyboard. Experimental comparisons between the conventional and these techniques established that increased ambient awareness for mobile users enhances nomadic text entry performance. Results showed that the textual and the textual & visual via translucent keyboard conditions increased text entry speed by 14% and 11%, respectively, and reduced the error rate by 13% compared to the regular technique. The two methods also significantly reduced the number of collisions with obstacles.
In den letzten Jahren haben sich elektronische Zahlungssysteme als populäre Alternative zur klassischen Bargeldzahlung etabliert. Diese Zahlungssysteme bestehen in der Regel aus zwei elementaren Komponenten: einem Terminal und einer Kasse. Damit ist der Käufer eines Produktes in der Lage, seine Schuld gegenüber dem Verkäufer bargeldlos und elektronisch zu begleichen. Die dabei am Häufigsten anfallenden Geschäftsprozesse, das Buchen und das Stornieren von Zahlungsbelegen, werden hierbei als Transaktionen bezeichnet, da diese entweder vollständig gelingen oder im Fehlerfall ohne Auswirkungen bleiben müssen. In diesem Buch wird daher die Implementierung eines zuverlässigen Zahlungssystems mit einem TeleCash-Terminal dargestellt. Dabei werden in den geforderten Geschäftsprozessen die wichtigen Transaktionseigenschaften sichergestellt. Es werden dazu zunächst die Grundlagen von Transaktionen erarbeitet und ein geeignetes Transaktionskonzept entwickelt. Anschließend wird die konkrete Realisierung des Systems mit Hilfe der Java Transaction Services durchgeführt. Abschließend wird das entstandene System hinsichtlich seiner Transaktionseigenschaften untersucht.
Current robot platforms are being employed to collaborate with humans in a wide range of domestic and industrial tasks. These environments require autonomous systems that are able to classify and communicate anomalous situations such as fires, injured persons, car accidents; or generally, any potentially dangerous situation for humans. In this paper we introduce an anomaly detection dataset for the purpose of robot applications as well as the design and implementation of a deep learning architecture that classifies and describes dangerous situations using only a single image as input. We report a classification accuracy of 97 % and METEOR score of 16.2. We will make the dataset publicly available after this paper is accepted.