Academia.eduAcademia.edu

THE VISUAL ENCODING OF TOOL–OBJECT AFFORDANCES

Abstract

The perception of tool-object pairs involves understanding their action-relationships (affordances). Here, we sought to evaluate how an observer visually encodes tool-object affordances. To this end, eye-movements were recorded as right-handed participants freely viewed static, right-handed, egocentric tool-object images across 3 contexts: correct (hammer-nail), incorrect (hammer-paper), spatial/ambiguous (hammer-wood), and 3 grasp-types: no hand, functional grasp-posture (grasp hammer-handle), non-functional/manipulative grasp-posture (grasp hammer-head). There were three Areas of Interests (AOI): the object (nail), the operant tool-end (hammer-head), the graspable tool-end (hammer-handle). Participants passively evaluated whether tool-object pairs were functionally correct/incorrect. Clustering of gaze scanpaths and AOI weightings grouped conditions into three distinct grasp-specific clusters, especially across correct and spatial tool-object contexts and to a lesser extent within the incorrect tool-object context. Permutation tests revealed that the grasp-specific gaze scanpath clusters were reasonably robust to the temporal order of gaze scanpaths. Gaze was therefore automatically primed to grasp-affordances though the task required evaluating tool-object context. Participants also primarily focused on the object and the operant tool-end and sparsely attended to the graspable tool-end, even in images with functional grasp-postures. In fact, in the absence of a grasp, the object was foveally weighted the most, indicative of a possible object-oriented action priming effect wherein the observer may be evaluating how the tool engages on the object. Unlike the functional grasp-posture, the manipulative grasp-posture was a gaze attractor and caused the greatest disruption in the object-oriented priming effect, ostensibly as it does not afford tool-object action due to its non-functional interaction with the operant tool-end that actually engages with the object (e.g. hammer-head to nail). The enhanced attention may therefore serve to encode the intent of the manipulative grasp-posture. Results here show how contextual and grasp-specific affordances directly modulate how an observer gathers action-information when evaluating static tool-object scenes.