Refine
H-BRS Bibliography
- yes (35) (remove)
Departments, institutes and facilities
- Fachbereich Informatik (30)
- Institute of Visual Computing (IVC) (15)
- Fachbereich Ingenieurwissenschaften und Kommunikation (4)
- Institut für Sicherheitsforschung (ISF) (2)
- Fachbereich Angewandte Naturwissenschaften (1)
- Fachbereich Wirtschaftswissenschaften (1)
- Institut für Cyber Security & Privacy (ICSP) (1)
- Institut für Detektionstechnologien (IDT) (1)
- Institut für Verbraucherinformatik (IVI) (1)
- Institut für funktionale Gen-Analytik (IFGA) (1)
Document Type
- Conference Object (35) (remove)
Year of publication
- 2012 (35) (remove)
Has Fulltext
- no (35)
Keywords
- Bag of Features (2)
- classifier combination (2)
- clustering (2)
- feature extraction (2)
- machine learning (2)
- object categorization (2)
- All-Swap Algorithm (1)
- CUDA (1)
- Design automation (1)
- EEG (1)
Für die Entwicklung steuerungstechnischer Sicherheitsfunktionen muss ab 2012 die Normen EN ISO 13849-1 oder EN 62061 befolgt werden, die sowohl Anforderungen an die Hardware als auch Anforderungen an die Software beschreibt. Die Anforderungen an die Software spielten bis vor einigen Jahren kaum eine Rolle, da Sicherheitsfunktionen vorzugsweise in Hardware realisiert wurden. Heutzutage ist es jedoch sehr häufig üblich, Sicherheitsfunktionen mit einer dafür geeigneten programmierbaren SPS zu realisieren. Die neuen Normen bzgl. der sicheren Steuerung von Maschinen verlangen neben der Quantifizierung der Hardware-Ausfallraten von Sicherheitsfunktionen noch ein Management der Sicherheitsfunktionen. Dazu gehört auch ein Management der Softwareentwicklung für Sicherheitsfunktionen, um systematische Fehler zu minimieren. Dieses Management der Softwareentwicklung wird im Wesentlichen durch das V-Modell repräsentiert. Für die Maschinebauindustrie darf dieser Managementprozess nicht zu aufwendig sein, ansonsten wird dieser in der Praxis nur schwer angenommen. Eine Möglichkeit der Abarbeitung des V-Modells wird vorgestellt. Wahrscheinlich ist diese aufgezeigte Möglichkeit für die Industrie noch zu aufwendig.
Traditionally traffic simulations are used to predict traffic jams, plan new roads or highways, and estimate road safety. They are also used in computer games and virtual environments. There are two general concepts of modeling traffic: macroscopic and microscopic modeling. Macroscopic traffic models take vehicle collectives into account and do not consider individual vehicles. Parameters like average velocity and density are used to model the flow of traffic. In contrast, microscopic traffic models consider each vehicle individually. Therefore, vehicle specific parameters are of importance, e.g. current velocity, desired velocity, velocity difference to the lead vehicle, individual time gap.
Approximate clone detection is the process of identifying similar process fragments in business process model collections. The tool presented in this paper can efficiently cluster approximate clones in large process model repositories. Once a repository is clustered, users can filter and browse the clusters using different filtering parameters. Our tool can also visualize clusters in the 2D space, allowing a better understanding of clusters and their member fragments. This demonstration will be useful for researchers and practitioners working on large process model repositories, where process standardization is a critical task for increasing the consistency and reducing the complexity of the repository.
This paper compares the memory allocation of two Java virtual machines, namely Oracle Java HotSpot VM 32-bit (OJVM) and Jamaica JamaicaVM (JJVM). The basic difference of the architectures in both machines is that the JamaicaVM uses fixed-size blocks for allocating objects on the heap. The basic difference of the architectures is that the JJVM uses fixed size block allocation on the heap. This means that objects have to be split into several connected blocks if they are bigger than the specified block-size. On the other hand, for small objects a full block must be allocated. The paper contains both theoretical and experimental analysis on the memory-overhead. The theoretical analysis is based on specifications of the two virtual machines. The experimental analysis is done with a modified JVMTI Agent together with the SPECjvm2008 Benchmark.
This paper describes adaptive time frequency analysis of EEG signals, both in theory as well as in practice. A momentary frequency estimation algorithm is discussed and applied to EEG time series of test persons performing a concentration experiment. The motivation for deriving and implementing a time frequency estimator is the assumption that an emotional change implies a transient in the measured EEG time series, which again are superimposed by biological white noise as well as artifacts. It will be shown how accurately and robustly the estimator detects the transient even under such complicated conditions.
The work presented in this paper focuses on the comparison of well-known and new techniques for designing robust fault diagnosis schemes in the robot domain. The main challenge for fault diagnosis is to allow the robot to effectively cope not only with internal hardware and software faults but with external disturbances and errors from dynamic and complex environments as well.
In the realm of service robots recovery from faults is indispensable to foster user acceptance. Here fault is to be understood not in the sense of robot internal, rather as interaction faults while situated in and interacting with an environment (aka ex-ternal faults). We reason along the most frequent failures in typical scenarios which we observed during real-world demonstrations and competitions using our Care-O-bot III 1 robot. They take place in an apartment-like environments which is known as closed world. We suggest four different -for now adhoc -fault categories caused by disturbances, imperfect per-ception, inadequate planning or chaining of action sequences. The fault are categorized and then mapped to a handful of partly known, partly extended fault handling techniques. Among them we applied qualitative reasoning, use of simu-lation as oracle, learning for planning (aka en-hancement of plan operators) or -in future -case-based reasoning. Having laid out this frame we mainly ask open questions related to the applicability of the pre-sented approach. Amongst them: how to find new categories, how to extend them, how to as-sure disjointness, how to identify old and label new faults on the fly.
Along with the success of the digitally revived stereoscopic cinema, other events beyond 3D movies become attractive for movie theater operators, i.e. interactive 3D games. In this paper, we present a case that explores possible challenges and solutions for interactive 3D games to be played by a movie theater audience. We analyze the setting and showcase current issues related to lighting and interaction. Our second focus is to provide gameplay mechanics that make special use of stereoscopy, especially depth-based game design. Based on these results, we present YouDash3D, a game prototype that explores public stereoscopic gameplay in a reduced kiosk setup. It features live 3D HD video stream of a professional stereo camera rig rendered in a real-time game scene. We use the effect to place the stereoscopic effigies of players into the digital game. The game showcases how stereoscopic vision can provide for a novel depth-based game mechanic. Projected trigger zones and distributed clusters of the audience video allow for easy adaptation to larger audiences and 3D movie theater gaming.
Traffic simulations for virtual environments are concerned with the behavior of individual traffic participants. The complexity of behavior in these simulations is often rather simple to abide by the constraints of processing resources. In sophisticated traffic simulations, the behavior of individual traffic participants is also modeled, but the focus lies on the overall behavior of the entire system, e.g. to identify possible bottle necks of traffic flow [8].