Projektleiter

 Öffnet einen externen Link in einem neuen FensterDr. Isabelle Bülthoff
Phone: +49 7071 601-611
Fax: +49 7071 601-616
isabelle.buelthoff[at]tuebingen.mpg.de
 
 

RecCat-Poster


Aktuellste Veröffentlichungen

Danyeli L, Alizadeh S, Surova G, Jamalabadi H, Schultz M und Walter M (Juni-2017): Effects of Neurexan® on brain responses to deviant stimuli during an auditory oddball task, ISAD LONDON 2017: Perspectives on Mood and Anxiety Disorders: Looking to the future, London, UK, Frontiers in Psychiatry, Conference Abstracts: ISAD LONDON 2017.
CiteID: DanyeliASJSW2017
Chuang LL, Gehring S, Kay J und Schmidt A: Ambient Notification Environments, Dagstuhl Seminar 17161, -, Leibniz-Zentrum für Informatik, Schloss Dagstuhl, Germany, (April-2017).im Druck
-, Series: Dagstuhl Reports
CiteID: ChuangGKS2017
Chuang LL (November-5-2015) Invited Lecture: Beyond Steering in Human-Centered Closed-Loop Control, Institute for Neural Computation: INC Chalk Talk Series, San Diego, CA, USA.
CiteID: Chuang2015_3
Stangl M, Meilinger T, Pape A-A, Schultz J, Bülthoff HH und Wolbers T (Oktober-19-2015): Triggers of entorhinal grid cell and hippocampal place cell remapping in humans, 45th Annual Meeting of the Society for Neuroscience (Neuroscience 2015), Chicago, IL, USA.
CiteID: StanglMPSBW2015
Fademrecht L, Bülthoff I, Barraclough NE und de la Rosa S (Oktober-18-2015): The spatial extent of action sensitive perceptual channels decrease with visual eccentricity, 45th Annual Meeting of the Society for Neuroscience (Neuroscience 2015), Chicago, IL, USA.
CiteID: FademrechtBBd2015_2

Export als:
BibTeX, XML, pubman, Edoc, RTF

Alle RecCat Veröffentlichungen

For all publications by RecCat members, click here

 

Dynamic stimuli: Emotional expressions in general are rather easily recognized in static displays, but they represent only 10 percent of all human facial expressions. In contrast, we have shown that dynamic information conveyed in conversational expressions (eg., “I don’t understand” ) is essential for correct interpretations of those expressions [Kaulard]. Another aspect of facial movements is how they help us to recognize the identity of a person. Facial movements of several persons were recorded and these movements were retargeted to avatar faces to investigate the role of motion idiosyncrasy for the recognition of face identity [Dobs]. Using functional brain imaging, we found that the increased response to moving compared to static faces is the result of both the fluid facial motion itself as well as the increase in static information [Schultz]. In the domain of dynamic objects, we could show that brain regions showed neural activity varying in parallel with the degree to which a single moving dot was perceived as inanimate or animate, suggesting that biological motion processing and detection of animacy are implemented in different brain areas [Schultz].
 
Active observers: View generalization for faces was tested in a virtual setup that allowed participants to move freely and collect actively numerous views of sitting and standing avatars that they encountered. Despite that, view generalization along the pitch axis did not occur for avatar faces, indicating that view dependency is not the consequence of passive viewing [I. Bülthoff]. Furthermore, data collected from active observers manipulating virtual objects revealed that observers rely on non-accidental properties of the objects (i.e., symmetry and elongation) to select views that they used for subsequent recognition [Chuang].
 
Cross-modal: Using haptic objects (plastic 3D face masks and sea shells) we have shown that training in one sensory modality can transfer to another modality, resulting in improved discrimination performance [Dopjans, Gaißert]. This transfer indicates that vision and touch share similar representations.
 
Other populations: In accordance with previous studies, we have shown that Korean observers fixate faces differently than their German counterparts, but interestingly, despite those differences in gaze movements, we found no effect of own-face expertise in terms of accuracy performance when the tasks did not involve memorizations of facial identity [I. Bülthoff]. We have started to investigate a large population of congenital Prosopagnosics. Preliminary results indicate that they are less sensitive to configural than to featural changes in faces that were modified parametrically. The use of a parametric stimulus space will allow a sensitive and precise quantitative description of their deficits [Esins]. Using parametric motion stimuli varying in complexity from simple translational motion to interacting shapes, we found that people with Autism Spectrum Disorder show deficits in assessing social interactions between moving objects [Schultz]. Faces were also used in a project to give a definite answer to the controversy about whether faces are perceived categorically in terms of their sex or not. Using our face database and the morphable model to manipulate the face stimuli to obtain very similar faces in terms of perceived masculinity and feminity allowed us to show that sex is not perceived categorically [Armann].
 


Last updated: Freitag, 13.10.2017