006 Spezielle Computerverfahren
Refine
Department, Institute
- Fachbereich Informatik (21)
- Fachbereich Wirtschaftswissenschaften (11)
- Institut für Verbraucherinformatik (IVI) (11)
- Institute of Visual Computing (IVC) (10)
- Institut für Technik, Ressourcenschonung und Energieeffizienz (TREE) (5)
- Institut für Sicherheitsforschung (ISF) (4)
- Fachbereich Elektrotechnik, Maschinenbau, Technikjournalismus (3)
- Fachbereich Angewandte Naturwissenschaften (1)
- Graduierteninstitut (1)
- Institut für funktionale Gen-Analytik (IFGA) (1)
Document Type
- Conference Object (14)
- Article (9)
- Part of a Book (3)
- Contribution to a Periodical (3)
- Doctoral Thesis (2)
- Preprint (2)
- Book (1)
- Report (1)
Year of publication
Is part of the Bibliography
- yes (35) (remove)
Keywords
- Augmented Reality (2)
- AR (1)
- AR design (1)
- AR development (1)
- AR/VR (1)
- Agile software development (1)
- Algorithmik (1)
- Altenhilfe (1)
- Aneignungsstudie (1)
- Ball Tracking (1)
- Bayesian Deep Learning (1)
- Blasendiagramm (1)
- Business Process Intelligence (1)
- Case study (1)
- Community of Practice (1)
- Computergrafik (1)
- Curriculum (1)
- Datenanalyse (1)
- Dementia (1)
- Demenz (1)
- Design (1)
- Disco (1)
- Drosophila (1)
- Educational Data Mining (1)
- Educational Process Mining (1)
- Emotion (1)
- Exergame (1)
- Experten (1)
- Facial Emotion Recognition (1)
- Fallbeschreibung (1)
- Fuzzy Mining (1)
- Geschäftsprozess (1)
- HCI (1)
- Hyperspectral image (1)
- ICT (1)
- Inductive Visual Mining (1)
- Kinect (1)
- Kollektiventscheidung (1)
- Komplexitätstheorie (1)
- Langzeitbehandlung (1)
- Living Lab (1)
- MR (1)
- Navigation (1)
- Optical Flow (1)
- Pflegepersonal (1)
- ProM (1)
- Process Mining (1)
- Qualitative study (1)
- Raman microscopy (1)
- RapidMiner (1)
- Ray tracing (1)
- Real-Time Image Processing (1)
- Smart Home (1)
- Social-Choice-Theorie (1)
- Spherical Treadmill (1)
- Spieltheorie (1)
- Studenten (1)
- Studienverlauf (1)
- Technologie (1)
- UAV (1)
- Uncertainty Quantification (1)
- Unterstützung (1)
- User Experience (1)
- User centered design (1)
- User feedback (1)
- VR (1)
- Videogame (1)
- Virtual Reality (1)
- Virtuelle Realität (1)
- Visuelle Wahrnehmung (1)
- Wissensaustausch (1)
- XR (1)
- aerodynamics (1)
- audio-tactile feedback (1)
- authoring tools (1)
- brightfield microscopy (1)
- computer vision (1)
- depth perception (1)
- dynamic vector fields (1)
- fitness-fatigue model (1)
- flight zone (1)
- geofence (1)
- guidance (1)
- image fusion (1)
- interactive computer graphics (1)
- mathematical modeling (1)
- mixed reality (1)
- pansharpening (1)
- performance modeling (1)
- performance prediction (1)
- practitioners (1)
- training performance relationship (1)
Background: Virtual reality combined with spherical treadmills is used across species for studying neural circuits underlying navigation.
New Method: We developed an optical flow-based method for tracking treadmil ball motion in real-time using a single high-resolution camera.
Results: Tracking accuracy and timing were determined using calibration data. Ball tracking was performed at 500 Hz and integrated with an open source game engine for virtual reality projection. The projection was updated at 120 Hz with a latency with respect to ball motion of 30 ± 8 ms.
Comparison: with Existing Method(s) Optical flow based tracking of treadmill motion is typically achieved using optical mice. The camera-based optical flow tracking system developed here is based on off-the-shelf components and offers control over the image acquisition and processing parameters. This results in flexibility with respect to tracking conditions – such as ball surface texture, lighting conditions, or ball size – as well as camera alignment and calibration.
Conclusions: A fast system for rotational ball motion tracking suitable for virtual reality animal behavior across different scales was developed and characterized.
HCI and CSCW research as well as practice has strongly indicated the value of integrating (end) users in software development processes. Such integration can help address actual needs and wants, to avoid undesirable developments and to strengthen the User Experience of a product. A user-focused approach to software development has some conceptual overlap with agile software development practices, such as quick and iterative (user) testing. However, out in the wild, organisations seem to have difficulties actually mapping user-centered development with agile processes for a variety of reasons ranging from organisational or hierarchical aspects up to financial issues. This problem seems specially prevalent in Small and Medium sized Enterprises (SMEs) where such constraints can be even tighter than in larger organisations. To help understand those problems and to identify possible solutions, we turned to three quite different German software SMEs, varying in size, market focus and organisational structure. By way of qualitative field studies, we were able to identify key roles and tools as well as methodological, organisational and analytical practices and challenges in integrating (end) users into agile software development.
„Industrie 4.0“ und weitere Schlagwörter wie „Big Data“, „Internet der Dinge“ oder „Cyber-physical Systems“ werden gegenwärtig in der Wirtschaft häufig aufgegriffen. Ausgangspunkt hierfür ist die Vernetzung von IT-Technologien sowie die durchgängige Digitalisierung. Nicht nur die Geschäftsfelder und Business-Modelle der Unternehmen selbst unterliegen dabei ei-nem entsprechend radikalen Wandel, dieser bezieht sich auch auf die Arbeitsumgebungen der Mitarbeiter, sowie den privaten und den öffentlichen Raum (Botthof, 2015; Hartmann, 2015).
Die Entwicklung intelligenter Technologien zur Unterstützung im Alltag und in den eigenen vier Wänden begleitet unsere Gesellschaft schon seit dem Zeitalter des Personal Computers. Mit dem Aufkommen des Internet der Dinge und begünstigt durch immer kleiner und günstiger werdende Hardware ergeben sich neue Potenziale, die das Thema Smart Home attraktiver als je zuvor werden lassen. Eine Vielzahl der aktuell im Markt verfügbaren Lösungen adressiert die Bedürfnisse Komfort, Sicherheit und effiziente Energienutzung. Die versprochene Intelligenz – smartness, wie sie der Begriff selbst suggeriert – wird vor allem bei Lösungen im privaten Nachrüstbereich überwiegend durch die Interaktion der Nutzer selbst und entsprechende regelbasierte Konfigurationen erzeugt. Diese notwendige Art der Interaktion und die damit verbundenen Aufwände sind jedoch von starker Bedeutung für das gesamte Nutzungserlebnis Smart Home und führen nicht selten zu Frustration oder gar Resignation in der Nutzung.
In this paper, we introduce an optical sensor system, which is integrated into an industrial push-button. The sensor allows to classify the type of material that is in contact with the button when pressed into different material categories on the basis of the material's so called "spectral signature". An approach for a safety sensor system at circular table saws on the same base has been introduced previously on SIAS-2007. This contactless working sensor is able to distinguish reliably between skin, textiles, leather and various other kinds of materials. A typical application for this intelligent push-button is the use at possibly dangerous machines, whose operating instructions include either the prohibition or the obligation to wear gloves during the work at the machine. An exemple of machines at which no gloves are allowed are pillar drilling machines, because of the risk of getting caught in the drill chuck and being turned in by the machine. In many cases this causes very serious hand injuries. Depending on the application needs, the sensor system integrated into the push-button can be configured flexibly by software to prevent the operator from accidentally starting a machine with or without gloves, which can decrease the risk of severe accidents significantly. Especially two-hand controls are incentive to manipulation for easier handling. By equipping both push-buttons of a two-hand control with material classification properties, the user is forced to operate the controls with his bare fingers. That limitation disallows the manipulation of a two-hand control by a simple rodding device.
Computer graphics research strives to synthesize images of a high visual realism that are indistinguishable from real visual experiences. While modern image synthesis approaches enable to create digital images of astonishing complexity and beauty, processing resources remain a limiting factor. Here, rendering efficiency is a central challenge involving a trade-off between visual fidelity and interactivity. For that reason, there is still a fundamental difference between the perception of the physical world and computer-generated imagery. At the same time, advances in display technologies drive the development of novel display devices. The dynamic range, the pixel densities, and refresh rates are constantly increasing. Display systems enable a larger visual field to be addressed by covering a wider field-of-view, due to either their size or in the form of head-mounted devices. Currently, research prototypes are ranging from stereo and multi-view systems, head-mounted devices with adaptable lenses, up to retinal projection, and lightfield/holographic displays. Computer graphics has to keep step with, as driving these devices presents us with immense challenges, most of which are currently unsolved. Fortunately, the human visual system has certain limitations, which means that providing the highest possible visual quality is not always necessary. Visual input passes through the eye’s optics, is filtered, and is processed at higher level structures in the brain. Knowledge of these processes helps to design novel rendering approaches that allow the creation of images at a higher quality and within a reduced time-frame. This thesis presents the state-of-the-art research and models that exploit the limitations of perception in order to increase visual quality but also to reduce workload alike - a concept we call perception-driven rendering. This research results in several practical rendering approaches that allow some of the fundamental challenges of computer graphics to be tackled. By using different tracking hardware, display systems, and head-mounted devices, we show the potential of each of the presented systems. The capturing of specific processes of the human visual system can be improved by combining multiple measurements using machine learning techniques. Different sampling, filtering, and reconstruction techniques aid the visual quality of the synthesized images. An in-depth evaluation of the presented systems including benchmarks, comparative examination with image metrics as well as user studies and experiments demonstrated that the methods introduced are visually superior or on the same qualitative level as ground truth, whilst having a significantly reduced computational complexity.
Advances in computer graphics enable us to create digital images of astonishing complexity and realism. However, processing resources are still a limiting factor. Hence, many costly but desirable aspects of realism are often not accounted for, including global illumination, accurate depth of field and motion blur, spectral effects, etc. especially in real‐time rendering. At the same time, there is a strong trend towards more pixels per display due to larger displays, higher pixel densities or larger fields of view. Further observable trends in current display technology include more bits per pixel (high dynamic range, wider color gamut/fidelity), increasing refresh rates (better motion depiction), and an increasing number of displayed views per pixel (stereo, multi‐view, all the way to holographic or lightfield displays). These developments cause significant unsolved technical challenges due to aspects such as limited compute power and bandwidth. Fortunately, the human visual system has certain limitations, which mean that providing the highest possible visual quality is not always necessary. In this report, we present the key research and models that exploit the limitations of perception to tackle visual quality and workload alike. Moreover, we present the open problems and promising future research targeting the question of how we can minimize the effort to compute and display only the necessary pixels while still offering a user full visual experience.
Noncooperative Game Theory
(2016)
Playing is something profoundly human, and the ability to play is tightly tied to the intelligence of human beings, to their capability of thinking foresightedly and strategically, of choosing a particularly profitable move among all possible moves, of anticipating possible response moves by their adversaries, and thus to their capability of maximizing their own profit. By playing a game we here mean, in general, an interaction under preassigned rules, amongst several players each interested in maximizing their gains and acting strategically to this end. Games are encountered everywhere, be it as a party game, a card game, a computer game, or a game of hazard, be it as an individual or team sport such as chess, foil fencing, soccer, or ice hockey, be it companies organizing their strategies in a market economy, or states and other global players deciding on their geopolitical strategies.
In this course, we will take a detailed look at various breeds of spatial navigation interfaces that allow for locomotion in digital 3D environments such as games, virtual environments or even the exploration of abstract data sets. We will closely look into the basics of navigation, unraveling the psychophysics (including wayfinding) and actual navigation (travel) aspects. The theoretical foundations form the basis for the practical skillset we will develop, by providing an in-depth discussion of navigation devices and techniques, and a step-by-step discussion of multiple real-world case studies. Doing so, we will cover the full range of navigation techniques from handheld to full-body, highly engaging and partly unconventional methods and tackle spatial navigation with hands-on-experience and tips for design and validation of novel interfaces. In particular, we will be looking at affordable setups and ways to “trick” out users to enable a realistic feeling of self-motion in the explored environments. As such, the course unites the theory and practice of spatial navigation, serving as entry point to understand and improve upon currently existing methods for the application domain at hand.
In this paper, we provide a participatory design study of a mobile health platform for older adults that provides an integrative perspective on health data collected from different devices and apps. We illustrate the diversity and complexity of older adults’ perspectives in the context of health and technology use, the challenges which follow on for the design of mobile health platforms that support active and healthy ageing (AHA) and our approach to addressing these challenges through a participatory design (PD) process. Interviews were conducted with older adults aged 65+ in a two-month study with the goal of understanding perspectives on health and technologies for AHA support. We identified challenges and derived design ideas for a mobile health platform called “My-AHA”. For researchers in this field, the structured documentation of our procedures and results, as well as the implications derived provide valuable insights for the design of mobile health platforms for older adults.
Informations- und Kommunikationstechnologie (IKT) in den Bereichen Smart Home und Smart Living ist durch die zunehmende Vernetzung des häuslichen Anwendungsfelds mit der Digitalisierung des Stromnetzes, alternativen Möglichkeiten der Energiegewinnung und -speicherung und neuer Mobilitätskonzepte geprägt und zu einem unverzichtbaren Bestandteil privaten wie unternehmerischen Handelns geworden.
In mathematical modeling by means of performance models, the Fitness-Fatigue Model (FF-Model) is a common approach in sport and exercise science to study the training performance relationship. The FF-Model uses an initial basic level of performance and two antagonistic terms (for fitness and fatigue). By model calibration, parameters are adapted to the subject’s individual physical response to training load. Although the simulation of the recorded training data in most cases shows useful results when the model is calibrated and all parameters are adjusted, this method has two major difficulties. First, a fitted value as basic performance will usually be too high. Second, without modification, the model cannot be simply used for prediction. By rewriting the FF-Model such that effects of former training history can be analyzed separately – we call those terms preload – it is possible to close the gap between a more realistic initial performance level and an athlete's actual performance level without distorting other model parameters and increase model accuracy substantially. Fitting error of the preload-extended FF-Model is less than 32% compared to the error of the FF-Model without preloads. Prediction error of the preload-extended FF-Model is around 54% of the error of the FF-Model without preloads.
Facial emotion recognition is the task to classify human emotions in face images. It is a difficult task due to high aleatoric uncertainty and visual ambiguity. A large part of the literature aims to show progress by increasing accuracy on this task, but this ignores the inherent uncertainty and ambiguity in the task. In this paper we show that Bayesian Neural Networks, as approximated using MC-Dropout, MC-DropConnect, or an Ensemble, are able to model the aleatoric uncertainty in facial emotion recognition, and produce output probabilities that are closer to what a human expects. We also show that calibration metrics show strange behaviors for this task, due to the multiple classes that can be considered correct, which motivates future work. We believe our work will motivate other researchers to move away from Classical and into Bayesian Neural Networks.