MIT CogNet, The Brain Sciences ConnectionFrom the MIT Press, Link to Online Catalog
SPARC Communities
Subscriber : Stanford University Libraries » LOG IN

space

Powered By Google 
Advanced Search

 

Predictive Eye-movements in Incremental Processing of Head-final Structures

 Yuki Kamide, Gerry T. M. Altmann and Sarah L. Haywood
  
 

Abstract:
In recent years, it has been demonstrated that lexical information is rapidly integrated with contextual information in order to predict and narrow down the domain of reference. More specifically, and based on the experimental technique developed by Tanenhaus et al. (1995), several eye-movements studies have shown that people fixate on a certain object in a visual scene as soon as they encounter auditory lexical items which distinguish that object from other potential referents in the visual scene (such distinguishing information includes adjectives in NPs (Sedivy et al., 1999), selectional restrictions of verbs in VPs (Altmann & Kamide, 1999), and the semantic relationship between a preposition and a subsequent NP in a PP (Chambers et al., 1998 )). For instance, Altmann & Kamide (1999) demonstrated that eye-movements were driven to a picture of a cake more frequently after subjects heard "The boy will eat _(II_(J" than after they heard "The boy will move_(II_(J", when the cake was the only edible object amongst several moveable objects in the same scene (a ball, a toy car, a toy train set, as well as the cake). However, it is crucial to note that the majority of those studies looked at the effect of heads in a head-initial language (English _(IQ_(J e.g., V in a VP; Altmann & Kamide, 1999; P in a PP; Chambers et al., 1998). An important question to be addressed here is whether similar patterns of predictive eye-movements will be found during the processing of pre-head constituents in head-final languages.

Strict variants of head-driven parsing (e.g., Pritchett, 1991, 1992) argue that structural analyses of the sentence occur only after the head has been encountered. Within this framework, processing of head-final structures should result in no predictive eye-movements of the kind revealed in the previous English studies.

We are currently conducting an eye-tracking experiment in which native speakers of Japanese are presented with pairs of a visual picture and an auditory sentence. There are two sentences (e.g., 1a & 1b) for each picture (e.g., 2):

(1a) Josee-ga fooku-de yukkurito supagettii-o taberu.
woman-nom fork-with slowly spaghetti-acc eat.
"The woman will eat the spaghetti with the fork slowly."

(1b) Josee-ga fooku-no-yoko-ni yukkurito naifu-o idoosuru.
woman-nom fork-next to slowly knife-acc move.
"The woman will move the knife next to the fork slowly."

(2) A visual scene consisting of "a woman", "a fork", "a plate of spaghetti", "a knife" and "a napkin"

The pre-head-driven parsing account (e.g., Inoue, 1991; Bader & Lasser, 1994; Koh, 1997; Kim, 1999; Kamide & Mitchell, 1999) predicts that people will fixate on the spaghetti after hearing "fork-with" even before hearing "spaghetti-acc" in (1a), since, in the visual scene, the spaghetti is the most plausible object for an action involving a fork as instrument. However, such eye-movements (to the spaghetti before "spaghetti" is heard) could be due to a simple lexical association between "fork" and "spaghetti". To rule this out, (1b) serves as the control condition for (1a). Anticipatory eye-movements to the spaghetti will be interpreted as support for the pre-head-driven parsing account if hearing "fork-next to" does not guide eye-movements to "spaghetti" in (1b) as frequently as "fork-with" does in (1a). In our presentation, we shall discuss the results of the experiment in the light of (pre-) head-driven parsing theories as well as the recent empirical studies in the same experimental paradigm using a typologically different language.

Altmann, G.T.M., & Kamide, Y. (1999). Incremental interpretation at verbs: Restricting the domain of subsequent reference. Cognition, 73(3), 247_(IP_(J264.

 
 


© 2010 The MIT Press
MIT Logo