Refine
H-BRS Bibliography
- yes (56) (remove)
Departments, institutes and facilities
- Fachbereich Informatik (42)
- Institute of Visual Computing (IVC) (17)
- Fachbereich Ingenieurwissenschaften und Kommunikation (12)
- Institut für Technik, Ressourcenschonung und Energieeffizienz (TREE) (11)
- Institut für Cyber Security & Privacy (ICSP) (7)
- Institut für funktionale Gen-Analytik (IFGA) (4)
- Fachbereich Angewandte Naturwissenschaften (2)
- Fachbereich Wirtschaftswissenschaften (2)
- Institut für Sicherheitsforschung (ISF) (1)
Document Type
- Conference Object (56) (remove)
Year of publication
- 2014 (56) (remove)
Language
- English (56) (remove)
Keywords
- education (2)
- path planning (2)
- Application Software (1)
- Applied and Practice-Oriented Research (1)
- Augmented reality (1)
- BFS (1)
- Business Sector (1)
- Control (1)
- Domain Expert (1)
- Eclipse Modeling Framework (1)
Improving data acquisition techniques and rising computational power keep producing more and larger data sets that need to be analyzed. These data sets usually do not fit into a GPU's memory. To interactively visualize such data with direct volume rendering, sophisticated techniques for problem domain decomposition, memory management and rendering have to be used. The volume renderer Volt is used to show how CUDA is efficiently utilised to manage the volume data and a GPU's memory with the aim of low opacity volume renderings of large volumes at interactive frame rates.
The ability to track moving people is a key aspect of autonomous robot systems in real-world environments. Whilst for many tasks knowing the approximate positions of people may be sufficient, the ability to identify unique people is needed to accurately count people in the real world. To accomplish the people counting task, a robust system for people detection, tracking and identification is needed.
Realism and plausibility of computer controlled entities in entertainment software have been enhanced by adding both static personalities and dynamic emotions. Here a generic model is introduced that allows findings from real-life personality studies to be transferred to a computational model. Adaptive behavior patterns are enabled by introducing dynamic event-based emotions. The advantages of this model have been validated using a four-way crossroad in a traffic simulation. Driving agents using the introduced model enhanced by dynamics were compared to agents based on static personality profiles and simple rule-based behavior. The results show that adding a dynamic factor to agents improves perceivable plausibility and realism.
In the field of domestic service robots, recovery from faults is crucial to promote user acceptance. In this context we focus in particular on some specific faults, which arise from the interaction of a robot with its real world environment. Even a well-modelled robot may fail to perform its tasks successfully due to unexpected situations, which occur while interacting. These situations occur as deviations of properties of the objects (manipulated by the robot) from their expected values. Hence, they are experienced by the robot as external faults.
Robots, which are able to carry out their tasks robustly in real world environments, are not only desirable but necessary if we want them to be more welcome for a wider audience. But very often they may fail to execute their actions successfully because of insufficient information about behaviour of objects used in the actions.
Unexpected Situations in Service Robot Environment: Classification and Reasoning Using Naive Physics
(2014)
Software repository data, for example in issue tracking systems, include natural language text and technical information, which includes anything from log files via code snippets to stack traces. However, data mining is often only interested in one of the two types e.g. in natural language text when looking at text mining. Regardless of which type is being investigated, any techniques used have to deal with noise caused by fragments of the other type i.e. methods interested in natural language have to deal with technical fragments and vice versa. This paper proposes an approach to classify unstructured data, e.g. development documents, into natural language text and technical information using a mixture of text heuristics and agglomerative hierarchical clustering. The approach was evaluated using 225 manually annotated text passages from developer emails and issue tracker data. Using white space tokenization as a basis, the overall precision of the approach is 0.84 and the recall is 0.85.
This article describes an approach to rapidly prototype the parameters of a Java application run on the IBM J9 Virtual Machine in order to improve its performance. It works by analyzing VM output and searching for behavioral patterns. These patterns are matched against a list of known patterns for which rules exist that specify how to adapt the VM to a given application. Adapting the application is done by adding parameters and changing existing ones. The process is fully automated and carried out by a toolkit. The toolkit iteratively cycles through multiple possible parameter sets, benchmarks them and proposes the best alternative to the user. The user can, without any prior knowledge about the Java application or the VM improve the performance of the deployed application and quickly cycle through a multitude of different settings to benchmark them. When tested with the representative benchmarks, improvements of up to 150% were achieved.
Application systems are often advertised with features, and features are used heavily for requirements man- agement. However, often software manufacturers only have incomplete information about the features of their software. The information is distributed over different sources, such as requirements documents, issue trackers, user manuals, and code. In this paper, we research the occurrence of feature information in open source software engineering data. We report on a case study with three open source systems. We analyze what information about features can be found in issue trackers and user documentation. Furthermore, we study the abstraction levels on which the features are described, how feature information is related, and we discuss the possibility to discover such information semi-automatically. To mirror the diversity of software development contexts, we choose open source systems, which are quite different, e.g., in the rigor of issue tracker usage. The results differ accordingly. One main result is that the user documentation did not provide more accurate information than the issue tracker compared to a provided feature list. The results also give hints on how the management of feature relevant information can be supported.
This work describes extensions to the well-known Distributed Coordination Function (DCF) model to account for IEEE802.11n point-to-point links. The developed extensions cover adaptions to the throughput and delay estimation for this type of link as well peculiarities of hardware and implementations within the Linux Kernel. Instead of using simulations, the approach was extensively verified on real-world deployments at various link distances. Additionally, trials were conducted to optimize the CWmin values and the number of retries to maximize throughput and minimize delay. The results of this work can be used to estimate the properties of long-distance 802.11 links beforehand, allowing the network to be planned more accurately.
A cost-efficient alternative to outside-in tracking systems for pointing interaction with large displays is to equip the pointing device with a camera, whose images are matched to display content. This work presents the Dynamic Marker Camera Tracking (DMCT) framework for display-based camera tracking. It accounts for typical display characteristics and uses dynamic on-screen markers overlaid to the display content that follow the camera. An example marker implementation and a tracking recovery method are presented. DMCT can measure pointing locations with sub-millimeter precision in large tracking volumes and computes 6-DoF camera poses for 3D interaction. 60 Hz update rate and 24 ms latency were achieved. DMCT's main limitation is the visible marker interfering with display content. In pointing effciency, the prototype is comparable to an OptiTrack system.
The RoCKIn@Home Challenge
(2014)
Sustainability is a key issue in current research activities and programs. In this conjunction three major functions of research have been identified: Basic research, knowledge reservoirs, and knowledge transfer. With regard to a transmission to the private sector, knowledge transfer is the most important factor. In this process, universities of applied sciences can play an important part as they typically have a long-standing experience in linking science and business in their teaching and research. Another important agent in the process of knowledge transfer are networks and clusters. Their strength lies integrating the different competencies of its partners and using them to a mutual benefit.
The International Centre for Sustainable Development (IZNE) – with a major focus on responsible business and sustainable food – takes the advantage of being part of a University of Applied Sciences (Bonn-Rhein-Sieg, BRSU), and being a member of several regional and international clusters and networks. These co-operations aim to establish and strengthen linkages between science and business, in particular by investigating research needs for business and business relevant research activities. Moreover, IZNE established and expanded regional and international co-operations of its own to get more transparency about regional and international value-added chains in the food sector and the issue of responsible business.