Intelligent Coaching Space
In this project, we explore how to most effectively support humans during the performance and the learning of motor actions.
Deep Familiarization and Learning Grounded in Cooperative Manual Action and Language
The Large scale project FAMULA focuses on a robot's autonomous familiarization with novel objects and their affordances.
Brain machine interfaces to improve human interactions using resource efficieny and adaptivity
How can multimodal brain-machine interfaces be applied to improve human machine interactions qualitatively?
This project introduces a novel physical, visuo-haptic and bi-manual, maze task to investigate the question of how humans acquire a newmanual skill.
Cognitive control and neurophysiological bases of planning and executing manual actions
This project's main purpose is to provide a better understanding of the (neuro)-functional relationship of cognitive functions and action.
A cognitive and mobile assistance system to provide individualized audiovisual action support in the modern working world
Intelligent glasses initially developed in the ADAMAAS project will be enhanced by an acoustic system and adapted for industrial application to reduce the burden on employees and increase productivity.
Adaptive and Mobile Action Assistance in Daily Living Activities - adaptive technical systems - for an intuitive interaction between humans and complex technologies
This project focuses on the development and testing of intelligent glasses. It combines techniques from memory research ...
Networked Living - The Smart Apartment (KogniHome)
The research goal of KogniHome is to develop and implement technology that improves the everyday live quality of inhabitants of a smart home.
KogniHome Subproject III
The supportive personal coach project strives to develop an individualized, adaptive health-training utilizing modern sensor technology and psychological methods.
Digital Motor Rehabilitation
The Vecury project investigates how virtual reality in combination with 3D motion tracking can support patients and therapists during physical rehabilitation.
Adaptiv virtuelle Rehabilitation bei Verletzungen der oberen Extremität
This project examines the experiences of patients and therapists during the introduction of an individualized and adaptive virtual-reality rehabilitation protocol for the upper extremities.
Adaptive and Modulare Architecture of Rich Motor Skills
Design and Evaluation of a Modular Adaptive Cognitive Mobile Assistive System for Flexible Use in Professional and Private Environments
DFG-Grant to Support the Initiation of International Collaboration.
Chess Expertise from Eye Gaze and Emotions
The aim of this project is to experimentally evaluate and compare current theories for mental modeling for problem solving and attention.
Cognitive mechanisms of motor planning - how cognitive and mechanical costs of a motor task affect the fractions of motor plan reuse and novel planning
This project from Bielefeld University and the v. Bodelschwinghsch Bethel Foundation aims to help handicapped individuals easily master the steps of on-the-job tasks
In this project interactive media is used in particular sporting situations to promote health and prevention.
Psychological support and consultancy of the Germany's female Youth Volleyball National Team on its way to the European and World Championship.
The main objective of the present project was the psychological support and supervision of the German female volleyball national team (under 18) on their way to the European and World championship.
This Art and Science project focuses on entrainment in dance, investigating complex processes of interaction, coordination and synchronization from a multidisciplinary perspective.
The overarching goal of this Thematic Network is to establish permanent international research cooperation at the educational and scientific levels.
Thematic Network Interactive Intelligent SystemsThis project aims to understand the influence of physical activity on the progress and outcome of standard depression therapy and to understand the underlying changes in (brain) neuroplasticity.
This project investigates the impact of movement expertise on visual perception of the environment when we observe objects, interactions or events. Our focus is on the specific interplay between eye movement parameters and the quality of mental representations during the selective perception of action-relevant information. By measuring visual attention patterns of experts and novices, we analyze how different levels of movement expertise affect the relative influence of top-down and bottom-up mechanisms during the perception of complex actions. We hope that the resulting insights can contribute to better informed designs of human-machine interfaces, humanoid robots and intelligent systems.
Instructions about sequences of actions are better memorized when they are offered with appropriate gestures. In this project, the virtual human MAX will accompany instructions with self-generated gestures, which leads to memory representations in the human listener. The quality of these representations will then be assessed by MAX (Split-Method), which provides for a measure of listener's comprehension, and which can in turn be used by MAX as a basis for the adjustment of its future use of particular instructions and gestures in a closed loop scenario.
The goal of this project is to create an incrementally growing database of manual interactions to help put manual intelligence research on a firmer empirical basis. In order to populate such a database necessitates the study of manual interactions in humans. The database should be populated with multimodal information: geometry information, tactile sensor information, vision information and sound information. Using these multimodal information sources will allow models to be built that represent manual interaction. The database can then be used to aid robots carry out complex tasks of the type that humans perform with ease. In order to decide on the structure of the database involves the answering of several important scientific questions: How should manual interactions be represented for storage, comparison and retrieval? What are suitable similarity measures for manual interactions? What are the elementary building blocks of a manual interaction? How do manual interactions motivated on the perceptual, control and task levels differ? Solving these questions will involve using skills in both psychology and computer science.
This project develops and establishes AcouMotion as a method for research in cognitive science and as novel tool for interaction technology. AcouMotion is a hard-/software system for combining human body motion, tangible interfaces and sonification to a closed-loop human computer interface that allows non-visual motor control by using sonification (nonspeech auditory displays) as major feedback channel. AcouMotion's main components are sensors to measure motion parameters, a computer simulation to represent the dynamical evolution of a model world, and sonification to render a real-time auditory representation of objects and any interactions in the model world. The applications of AcouMotion range from new vision-free sport games over physiotherapy to cognitive research.
The development of appropriate representations for manual action episodes and their coordination to more complex action sequences still poses a major challenge for robotics. It is a natural approach to gain insight to these topics from a cognitive and experimental perspective revealing the development of sensorimotor representations in humans accomplishing those complex tasks. Especially, it is interesting to learn about the relationship between the structure of representations and the performance in the execution of sensorimotor tasks, like sequential object manipulations. Hence, one of the most interesting questions is what kind of representation helps people to control actions. This project focuses on the question of how structures of sensorimotor representations are established and gradually changed, while considering the physical properties of objects, task constraints (affordances) and perceptual discordances in sensorimotor adaptation tasks.
A remarkable ability humans have, is to organize and discretize perceived sensory input and knowledge about the world into categories and concepts. This way of knowledge representation is an essential part of efficient communication as well as choosing appropriate actions in novel situations. The goal of this project is to investigate ways how the formation of concepts and categories could be artificially created and probably implemented on real robots. Robots endowed with such a mechanism should be able to incrementally increase their knowledge about the world thus enlarging their opportunities to act in the world.
What insights can we gain from psychological measurements of biomechanical parameters and subjective judgments of manual actions (like object grasping) about the structures of the underlying cognitive representations? In this project, we will bring together statistical methods (like structure dimensional and principal analysis) with connectionist approaches employing artificial neural networks to test different hypotheses about the cognitive structure of manual actions. A major goal will be to emulate and control grasping behavior for a broad range of objects in kinematic simulations and - as a longer term objective - in real physics on a robot platform.
Within this project we investigate the development of motor skills by explorative learning based on the control basis framework of Grupen et al, which allows a flexible combination of simple control units to achieve complex control tasks. One major aspect of the proposed project will be the basic implementation of the control basis framework within the 2-Arm Manipulation-Setup and for the Barthoc torsos. This includes to choose the right set of elementary controller units, and to investigate the resulting behavior if they are combined in various ways. Another aspect will be to develop coordination mechanisms required to execute and monitor several primitives in sequence or in parallel. Here the question arises how to prioritize several conflicting behaviors that operate in parallel. Which mechanisms are needed to dynamically weight primitives depending on situational context, perceptual confidence of different sensor modalities, and user-specified task priority? If elementary controllers are combined in a hierarchical manner, how can their resource requirements be automatically propagated through this hierarchy to provide a foundation for scheduling? Finally, the most challenging aspect is the development of learning methods, which allow the robot to acquire basic motor behavior from scratch by exploration, eventually primed from visually or otherwise observed trajectories. Especially interesting in this context are of course manipulative movements, e.g. pushing an object with a single finger or re-grasping an object in order to improve the grasp quality or to realize a desired object orientation w.r.t. the hand. Research Areas
Representations of spatial relations play an important role for joint attention, joint action, cooperation and competition between humans as well as between humans and robotic systems. Spatial representations in human-human or human-robot interaction are in many cases coordinated via language. In this project we want to apply psycholinguistic and actionrelated behavioural paradigms to corroborate the position that the type of viewing point used might be highly relevant for the characterization of a reference frame because different viewing points may be based on different (partly modality-dependent) types of perceptual information.
One central issue for the cognitive control of movement is the compensation of errors and learning processes that enhance error compensation mechanisms. This is especially true for very precise movements such as many manual actions. The present project combines methods and conventional experimental settings (first order reality) with approaches from Virtual Reality and Augmented Reality to embed subjects in interaction loops in which the occurrence and perception of errors can be manipulated and studied in novel ways. In this way we hope to gain new clues about error correction mechanisms, error compensation learning and their replication in technical systems such as robots.
Speed Stacking is a bimanual scenario that consists of upstacking and downstacking pyramids of plastic cups in pre-determined sequences as quickly as possible. Since learners usually improve significantly during short periods of time, speed stacking is particularly well suited to investigate motor skill learning and automatization. This rapid improvement and additional constant feedback about errors and stacking duration renders the task very motivating. We will analyze how attentional guidance and motion paths change during the learning process and how two individuals align their action timing. Research will be based on the processbased measures motion path and gaze pattern, accompanied by product-based measures like task duration and error rate.
The aim of this project is to investigate complex movements of the human body in sports, dance or every-day life activities, and to understand how these movements are controlled, learned and reproduced under changing conditions. By means of biomechanical and psychological methods, we will investigate physical, physiological and cognitive aspects of movement performance in order to understand how the different systems play together to control full body movements in an adaptive way. Based on our experimental results, we want to develop a simulation framework that integrates the different control systems and their interaction, including a high level control network that generates and adapts sequences of motor primitives representing control on the level of mental representation.