Skip to main content

METHODS article

Front. Psychol., 17 November 2015
Sec. Cognition
This article is part of the Research Topic Reaching to grasp cognition: analyzing motor behaviour to investigate social interactions View all 13 articles

The Multilingual CID-5: A New Tool to Study the Perception of Communicative Interactions in Different Languages

  • 1CoBTeK Laboratory, Faculty of Medicine, University of Nice Sophia Antipolis, Nice, France
  • 2Department of Psychology, University of Turin, Turin, Italy
  • 3Faculty of Psychology, University of Warsaw, Warsaw, Poland
  • 4Department of Psychology, Center for Mind and Brain & The MIND Institute, University of California, Davis, Davis, CA, USA
  • 5Centre Mémoire de Ressources et de Recherche, CHU de Nice, Nice, France
  • 6Department of Psychiatry, University Hospital Cologne, Cologne, Germany
  • 7Max Planck Institute of Psychiatry, Munich, Germany
  • 8Laboratory of Experimental Psychology, KU Leuven, Leuven, Belgium
  • 9Cognitive Neuroscience – Institute of Neuroscience and Medicine (INM3), Research Center Jülich, Jülich, Germany
  • 10Department of Psychiatry and Psychotherapy, Heinrich-Heine-University of Düsseldorf, Rhineland State Clinics Düsseldorf, Düsseldorf, Germany
  • 11Department of Robotics, Brain and Cognitive Sciences, Fondazione Istituto Italiano di Tecnologia, Genova, Italy

The investigation of the ability to perceive, recognize, and judge upon social intentions, such as communicative intentions, on the basis of body motion is a growing research area. Cross-cultural differences in ability to perceive and interpret biological motion, however, have been poorly investigated so far. Progress in this domain strongly depends on the availability of suitable stimulus material. In the present method paper, we describe the multilingual CID-5, an extension of the CID-5 database, allowing for the investigation of how non-conventional communicative gestures are classified and identified by speakers of different languages. The CID-5 database contains 14 communicative interactions and 7 non-communicative actions performed by couples of agents and presented as point-light displays. For each action, the database provides movie files with the point-light animation, text files with the 3-D spatial coordinates of the point-lights, and five different response alternatives. In the multilingual CID-5 the alternatives were translated into seven languages (Chinese, Dutch, English, French, German, Italian, and Polish). Preliminary data collected to assess the recognizability of the actions in the different languages suggest that, for most of the action stimuli, information presented in point-light displays is sufficient for the distinctive classification of the action as communicative vs. individual, as well as for identification of the specific communicative gesture performed by the actor in all the available languages.

Introduction

Successful gestural communication depends on the recipient understanding and recognizing the intention of the communicative act (Sperber and Wilson, 1986). To do so, the recipient needs to be able to (a) discriminate between communicative gestures and individual actions (intention classification), and (b) identify the specific communicative content conveyed by the gesture (intention identification). Conventional emblematic communicative gestures, such as the ‘okay’ sign or the ‘thumbs-up’ gesture, have a certain format and an explicit meaning established by the conventions of specific communities. It is thus unsurprising that they may have radically different meanings from one society to another, or even within a single communicative tradition. ‘The horns,’ made by extending the pinkie and index finger while making a fist, for example, is used to ward off the evil eye in traditional Mediterranean cultures. Variants of this gesture were used in Elizabethan England to accuse a man of having an unfaithful wife, in modern England and the US to express a passion for heavy metal music (Casasanto, 2013). Non-conventional gestures, on the contrary, may be more easily understood across cultures. Pointing when giving directions, reaching up to show how tall someone is, gesturing towards an empty seat, are all examples of communicative gestures that can serve as a ‘quasi-universal’ language (Marsh et al., 2007). Comparison of results obtained in different linguistic contexts and cultures, however, have so far been hindered by lack of evaluation instruments adapted and validated in different languages.

In the present work, we describe the multilingual CID-5, a new tool being made available in seven languages for the investigation of how non-conventional communicative gestures are classified and identified in different linguistic and cultural contexts. In the following, we present first a brief background on the point-light technique used to create stimuli included in the multilingual CID-5, and in the original CID-5 database. Next, we provide a detailed description of multilingual CID-5 database, including all the materials available for download. Finally, we present normative data collected to assess the stimulus classification (communicative vs. individual) and intention identification by speakers of seven different languages, namely Chinese, Dutch, English, French, German, Italian, and Polish.

Gestural Communication through Point-light Displays

The point-light technique is a method for representing biological motion through limited visual information (Johansson, 1973). With this method, the movements of a body of a living being are represented by a small number of point lights indicating the major joints of a person performing a given action. Despite the absence of other cues such as contour, color, or texture, observers can quite easily identify what an actor is doing (e.g., Vanrie and Verfaillie, 2004), as well as many features of the actor themselves, including identity (e.g., Loula et al., 2005), gender (e.g., Kozlowski and Cutting, 1977; Pollick et al., 2005; Brooks et al., 2008), age (Montpare and Zebrowitz-McArthur, 1988), emotional state (e.g., Pollick et al., 2001; Atkinson et al., 2004; Clarke et al., 2005), and personality traits (Heberlein et al., 2004).

Given this keen sensitivity to action motion signatures, it is reasonable to expect that people are also able to discern communicative gestures from point-light displays. Along these lines, recent evidence suggests that biological motion information is sufficient for clear classification of a non-conventional action as communicative, as well as for the identification of the specific communicative intent (Manera et al., 2010, 2011a,b, 2013; Centelles et al., 2013). Furthermore, Manera et al. (2011a, 2013) demonstrated that in the context of a communicative interaction between two point-light agents, observing the communicative gesture of one agent enhances the visual discrimination of a second agent who responds appropriately.

The generalizability of these findings across different cultures and linguistic communities, however, is until now poorly documented. There is evidence that biological motion perception is not necessarily influenced by culture, and that point-light stimuli reproducing simple and putatively universal human actions, such as walking, can be recognized even by indigene populations of Amazonian territories (Pica et al., 2011; see also Barrett et al., 2005), as well as by newborns (Simion et al., 2008). It remains possible, however, that cultural tendencies to display particular non-conventional gestures in certain styles influence intention-from-movement judgments, and that speakers of different languages may classify and describe the same actions differently.

The CID-5 Database

The CID-5 database (Communicative Interaction Database, Five Alternative Forced Choice format, 5AFC) contains 21 full-body point-light stimuli depicting two agents (A and B) engaged either in communicative interactions (N = 14) or non-communicative individual actions (N = 7) as seen from four different viewpoints. Following Dekeyser et al. (2002), stimuli were constructed by combining motion capture techniques and 3-D animation software to provide precise control over the computer-generated actions and allow the actions of the two agents to be independently manipulated. For each action stimulus, the CID-5 provides (i) coordinate files for each actor; (ii) movie files depicting the action of the two agents as seen from four different perspectives; (iii) five action alternatives describing the action performed by the two agents. The CID-5 database can be freely downloaded from http://bsb-lab.org/research/.

Results collected on a sample of 113 Italian speaking participants using these stimuli confirmed that naive observers are able to distinguish communicative and individual gestures, and to identify the correct action description among the five alternatives (Manera et al., 2015). The multilingual CID-5 extends the CID-5 by providing a translation of the response alternatives into seven different languages: Chinese, Dutch, English, French, German, Italian, and Polish. Furthermore, it provides some normative data to validate the alternative action descriptions in the different languages.

The Multilingual CID-5 Database

Building on the CID-5 database, the multilingual CID-5 database provides a new tool to investigate classification and identification of non-conventional communicative gestures by speakers of different languages. The database is available as Supplementary Material to this article, or from the website of the Biology of Social Behavior Lab, University of Torino (http://bsb-lab.org/research/).

Actions

A brief description of each action stimulus is reported in Table 1. Stimuli consist of the 21 point-light actions depicting two point-light agents, each consisting of 13 markers indicating head, shoulders, elbows, wrists, hips, knees, and feet. For each stimulus, we report the stimulus classification (communicative vs. individual), a brief description of the actions of agent A and agent B, and the actors’ gender.

TABLE 1
www.frontiersin.org

TABLE 1. Description of the actions included in the CID-5 database.

Stimuli were originally constructed by capturing the movements of four actors, two Italian females and two Dutch males, each wearing 30 reflective spherical markers (Qualisys MacReflex motion capture system; Qualisys; Gothenburg, Sweden, consisting of six 30-Hz position units). For the communicative interactions, the two female and the two male actors worked in pair (a couple of male actors and a couple of female actresses) and were assigned to a ‘communicator’ and ‘responder’ role. The communicator (agent A) always initiated the interaction by performing a communicative gesture; the responder (agent B) perceived the communicative gesture and acted in response, based on a predefined interaction plot. To ensure that the responder’s action matched the communicator’s gesture in all respects (e.g., timing, position, kinematics), interactions were captured in real time, with the actors facing each other, at a distance of approximately 2 m. Individual actions were performed by agent A acting in isolation. Objects (e.g., table, chair, coins, fruits) were present during the production of actions to aid the actors in producing natural movements.

After the capture session, the 2-D data from all the position units were processed offline to calculate the 3-D coordinates of the markers. Missing data points (less than 5%) were filled in manually using customized functions of the Fluey 2 motion toolkit (MTK, Televirtual). The data from the markers were then imported into Character Studio (Autodesk Inc, 1998). This allowed to animate a biped for each actor, consisting of a transparent skeleton and 13 bright dots attached to the center of the major joints (shoulders, elbows, wrists, hips, knees, and ankles) and the head. To create the actual movie files, the smoothed data were imported into 3-D Studio as moving bright spheres, and all the frames of the action were rendered as avi-files from four different viewpoints. Some manual smoothing was performed to avoid any remaining “jumpy” dot movements. An orthographic projection was used, and there was no occlusion, so no explicit depth cues were available. To create the communicative action stimuli avi-files, data from the two actors of each couple were imported into the same 3D studio environment, making sure that the actors were exactly at the same distance as in the original recording session. To create the individual action stimuli.avi files, the communicator’s gesture was substituted with an individual action performed by the same actor, making sure to match stimulus duration. Objects present in the scene during motion capturing were never visible in any of the point-light displays.

Response Alternatives

The ‘Response Alternatives’ folder contains seven.doc files (Supplementary Data Sheet S1) reporting the list of the five response alternatives for each action stimulus in seven different languages (Chinese, Dutch, English, French, German, Italian, and Polish).

The five alternatives included the correct action description and four incorrect response alternatives. The incorrect response alternatives were generated according to the following criteria. For each action stimulus (e.g., A asks B to walk away), two incorrect communicative alternatives (e.g., A opens the door for B; A asks B to move something) and two incorrect non-communicative alternatives (A stretches; A draws a line) were generated by modifying the description of the action of agent A. All alternative action descriptions were constructed to be physically compatible with the action performed by agent A. For instance, if agent A performed an arm movement, then reference to arm movement was included in all incorrect response alternatives describing the action stimulus. Finally, to avoid that for communicative stimuli the correct alternative was selected simply based on the congruence between the actions of the two agents (i.e., agent A asks B to perform an action, and agent B responds accordingly), for each action stimulus, one of the incorrect communicative alternatives always described a congruent interaction between the two agents (see Supplementary Table S1). The description of the action of agent B was the same for all response alternatives.

Translation of the Alternatives

Translations in each language were performed by two independent native speakers. Translators were provided with the English version of the alternatives, and the original point-light movie files. The two translations were then compared, and in case of discrepancies, the translators were asked to decide together which description matched better the English version of the alternative and the corresponding point-light video.

Collection of Preliminary Data

Participants

One hundred and forty healthy volunteers (61 male, 79 female; age, M = 24.9, SD = 4.6, years of education, M = 15.8, SD = 2.2) took part in this study, 20 for each of the following languages: Chinese, Dutch, English, French, German, Italian, and Polish. Participants were recruited at the University and Polytechnic of Torino, in Italy (Chinese and Italian speakers), at the Katholieke Universiteit Leuven, in Belgium (Dutch speakers), at the University of California at Davis, in the US (English speakers), at the University Hospital Cologne, in Germany (German speakers), at the University of Nice Sophia Antipolis and the Nice University Hospitals, in France (French speakers), and at the University of Warsaw, in Poland (Polish speakers). They received course credits or payment for their participation. Demographic characteristics of the participants of each country are reported in Table 2 and Figure 1. All participants had normal or corrected to normal vision, and were naive as to the purpose of the study. The study was approved by the local ethical committees.

TABLE 2
www.frontiersin.org

TABLE 2. Participant’s demographics.

FIGURE 1
www.frontiersin.org

FIGURE 1. Mean age and education for the different language groups.

Stimuli and Procedure

Twenty-one point-light actions taken from CID-5 database were employed (Manera et al., 2015), including 14 communicative interactions in which the two agents (A and B) were engaged in a communicative interaction (e.g., agent A points out at the ceiling, agent B looks at the ceiling) and 7 non-communicative individual actions, in which A and B were acting independently of each other (e.g., A drinks, B sits down). Agent A was positioned on the right side of the screen, and agent B was positioned on the left side of the screen (corresponding to the 125° perspective in the CID-5 database; see the description of the video perspectives reported in Manera et al., 2010) in all the action stimuli. The two agents were displayed simultaneously, with action of agent B (the responder in the communicative stimuli) always following in time the action of agent A (the communicator in the communicative stimuli). Stimuli were presented in a randomized order. Following the procedure used in previous reports (e.g., Manera et al., 2015), each video with the two agents was shown twice consecutively, with the two videos separated by a 500 ms fixation cross. After the second presentation of each video, participants were, firstly, asked to decide whether the two agents were communicating vs. acting independently of each other (intention classification). The question was displayed on the screen until a response was provided. Secondly, participants were asked to select the correct action description among five numbered response alternatives displayed simultaneously (intention identification). The order of the response alternatives was randomized across stimuli. The question was presented on the screen until response, with no time restriction. No feedback concerning response correctness was given to the participants. Depending on the sample, instructions, questions, and response alternatives were presented in Chinese, Dutch, English, French, German, Italian, or Polish.

The Chinese and Polish versions of the procedure were created with E-prime software (Psychology Software Tool, Inc), while the Dutch, English, French, German, and Italian versions of the procedure were created with Presentation software (Neurobehavioral Systems, Inc.) In all language samples, stimuli were displayed on a 14″ to 17″ LCD screen. The task took approximately 15–20 min to complete.

Results

Demographics

Chi Square analysis revealed no gender differences among the seven samples corresponding to the different languages (χ2 = 4.76, p = 0.574). A between-subject ANOVA with age as dependent variable and language (Chinese, Dutch, English, French, German, Italian, and Polish) as between subject factor revealed a significant difference in age among the different language groups [F(6,133) = 11.52, p < 0.001]. Post hoc comparisons (Bonferroni corrected) revealed that French-speaking participants were significantly older than Chinese-speaking participants (p < 0.001), Dutch-speaking participants (p = 0.003), English-speaking participants (p < 0.001), Italian-speaking participants (p = 0.010) and Polish-speaking participants (p < 0.001). Dutch-speaking participants were significantly older than English-speaking participants (p = 0.027). English-speaking participants were younger than German-speaking participants (p < 0.001), and Italian-speaking participants (p = 0.008). Finally, German-speaking participants were significantly older than Polish-speaking participants (p = 0.004).

A between-subject ANOVA with education as dependent variable and language (Chinese, Dutch, English, French, German, Italian, and Polish) as between subject factor revealed a significant difference in education between the different language groups [F(6,133)= 11.00, p < 0.001]. Post hoc comparisons (Bonferroni corrected) revealed that French-speaking participants had more education years compared to Chinese-speaking participants (p < 0.001), Dutch-speaking participants (p = 0.022), English-speaking participants (p < 0.001), and Polish-speaking participants (p < 0.001). Chinese-speaking participants had fewer education years compared to German-speaking participants (p < 0.001). English-speaking participants had fewer education years compared to German-speaking participants (p = 0.001) and Italian-speaking participants (p = 0.022). Finally, Polish speaking participants had fewer education years compared to German-speaking participants (p < 0.001) and Italian-speaking participants (p = 0.001).

As age and education differed among the seven language-samples, they were added as covariates in all the between-subject analyses.

Multilingual CID-5 Task

Separate analyses were conducted to evaluate global performance and recognizability of single stimuli.

Data analysis

Global performance. To evaluate global performance, for each language we calculated the percentage of participants who correctly responded to the intention classification and the intention identification questions, and we extracted Signal Detection Theory (SDT) parameters.

To evaluate how participants distinguished between communicative and individual action stimuli (intention classification), we calculated sensitivity (d′) and criterion (c′). For each participant, we calculated the proportion of hits (arbitrarily defined as “communicative” responses when the action stimulus was communicative) and false alarms (“communicative” responses when the action stimulus was individual). Proportions of 0 were replaced with 0.5/N, and proportions of 1 were replaced with (N–0.5)/N (where N is the number of communicative and individual stimuli). d′ and c were then submitted to single sample t-tests (test value = 0) to ascertain whether discrimination performance was above chance level, and to verify the presence of any systematic response bias. Furthermore, to ascertain whether d′ and c varied across languages, they were submitted to separate ANCOVAs with Language (Chinese, Dutch, English, French, German, Italian, and Polish) and Gender (Male vs. Female) as between-subject factors, and Age and Education as covariates. Finally, in order to verify the presence of interactions between participants’ gender and the gender of the actors in the ability to classify the actions as communicative vs. individual, the d′ was submitted to a repeated measures ANOVA with Actor gender as within-subject factor, and Gender as between-subject factor.

To evaluate global performance on the intention identification question, we first recodified responses as communicative vs. individual to calculate sensitivity (d′) and criterion (c). d′ and c were submitted to single sample t-tests (test value = 0) to ascertain whether discrimination performance was above chance level, and to verify the presence of any systematic response bias. Second, to evaluate the ability to select the correct response alternative, following the standard SDT approach to mAFC (e.g., Macmillan and Creelman, 2005), we used the proportion correct responses as a measure of sensitivity. To compare performance across different languages, we submitted the mean proportion of correct responses to a repeated measures ANCOVA with Intention (Communicative vs. Individual) as within-subject factor, Language (Chinese, Dutch, English, French, German, Italian, and Polish) and Gender (Male vs. Female) as between-subject factors, and Age and Education as covariates. Finally, in order to verify the presence of interactions between participants’ gender and the gender of the actors in the intention identification ability, the proportion of correct responses was submitted to a repeated measures ANOVA with Actor gender as within-subject factor and Gender as between-subject factor.

Stimulus recognizability. To provide researchers with detailed data on the classification and identification of single stimuli across languages, for each action stimulus, we first calculated whether the proportion of correct responses differed from chance level – that is, from 0.5 for question 1 (corresponding to 50% of correct responses) and 0.2 for question 2 (corresponding to 20% of correct responses) – by employing binomial tests. Bonferroni corrections were applied to adjust for multiple comparisons (α = 0.05/21, = 0.0023). Second, we verified whether the distribution of correct responses (0 for incorrect response, 1 for correct response) and the distribution of errors (communicative alternative 1, communicative alternative 2, individual alternative 1, and individual alternative 2) varied depending on the factor Language (Chinese, Dutch, English, French, German, Italian, and Polish) by means of Chi-square analyses (see Supplementary Table S1). Bonferroni corrections were applied (p < 0.0023).

Global performance: results

Descriptive statistics (mean and SD) for the percentage of correct responses for the intention classification and the intention identification questions for each language are reported in Table 3.

TABLE 3
www.frontiersin.org

TABLE 3. Percentage of correct responses for the intention classification and identification questions.

Intention classification. Sensitivity (d′) for the full sample (N = 140) ranged from 0.27 to 3.23 (M = 2.46, SD = 0.57; see Figure 2A), and was significantly higher than zero [t(139) = 51.15, p < 0.001], thus suggesting that participants, as a group, were able to discriminate communicative from individual action stimuli well above the chance level. The ANCOVA on d′ with Language and Gender as between subject factors, and Age and Education as covariates revealed no statistically significant effects [corrected model, F(15,124)= 0.81, p = 0.664]. Specifically, no significant effect of Language [F(6,124)= 0.83, p = 0.546], Gender [F(1,124)= 0.19, p = 0.662], Age [F(1,124)= 3.48, p = 0.064] or Education [F(1,124)= 0.23, p = 0.633] was found. Criterion c for the full sample (N = 140) ranged from –1.61 to 0.65 (M = –0.15, SD = 0.36; see Figure 2B), and was significantly lower than zero [t(139) = –4.84, p < 0.001] thus suggesting that participants, as a group, had a tendency to rate stimuli as communicative. The ANCOVA with Language and Gender as between subject factors, and Age and Education as covariates was statistically significant [corrected model, F(15,124)= 2.95, p < 0.001]. Specifically, a main effect of Language was found [F(6,124)= 4.75, p < 0.001]. Post hoc comparisons revealed that Chinese-speaking participants had a stronger tendency to rate stimuli as communicative compared to French-speaking (p = 0.014) and Italian-speaking participants (p = 0.007). Similarly, Polish-speaking participants had a stronger tendency to rate stimuli as communicative compared to French-speaking (p = 0.004) and Italian-speaking participants (p = 0.001). No significant effect of Gender [F(1,124)= 3.08, p = 0.082], Age [F(1,124)= 1.54, p = 0.218] or Education [F(1,124)= 0.03, p = 0.864] on c was found.

FIGURE 2
www.frontiersin.org

FIGURE 2. Sensitivity (A) and criterion (B) for the intention classification question.

The repeated measures ANOVA on d′ with Actor gender as within-subject factor and Gender as between-subject factor revealed a significant main effect of Actor Gender [F(1,138) = 6.70, p = 0.011], with the proportion of correct responses being significantly higher for the female actresses (M = 2.56) compared to the male actors (M = 2.33). No effect of Gender [F(1,138) = 0.92, p = 0.340] and no significant interaction between Actor Gender and Gender [F(1,138) = 0.06, p = 0.815] was found.

Intention identification. The d′ calculated on the full sample (N = 140) after re-codifying the responses as communicative vs. individual ranged from 0.40 to 3.27 (M = 2.60, SD = 0.53), and was significantly higher than zero [t(139) = 58.18, p < 0.001], thus suggesting that, also for the action identification question, participants were able to discriminate communicative from individual action stimuli well above the chance level. Criterion c calculated on the full sample ranged from –1.27 to 0.65 (M = –0.04, SD = 0.28), and was not significantly different from zero [t(139) = –1.56, p = 0.122], thus suggesting that participants, contrary to what happened in the intention classification question, when asked to select the correct action description among several action alternatives, showed no response bias toward a communicative response.

The proportion of correct response alternatives for each language is reported in Figure 3. The repeated measures ANCOVA with Intention as within-subject factor, Language and Gender as between subject factors and Age and Education as covariates revealed no significant effect of Intention [F(1,124)= 0.76, p = 0.384], Gender [F(1,24)= 0.27, p = 0.606], Age [F(1,124)= 2.21, p = 0.139], or Education [F(1,124)= 2.13, p = 0.147] on the proportion of correct responses. However, a significant effect of Language was found [F(6,124)= 2.71, p = 0.017]. Post hoc comparisons revealed that French-speaking participants performed significantly better compared to the Dutch-speaking participants (p = 0.026). No two-way or three-way interaction reached statistical significance (all ps > 0.056).

FIGURE 3
www.frontiersin.org

FIGURE 3. Proportion of correct responses for the intention identification question.

The repeated measures ANOVA with Actor gender as within-subject factor and Gender as between-subject factor revealed a significant main effect of Actor Gender [F(1,138) = 70.40, p < 0001], with the proportion of correct responses being significantly higher for the female actresses (M = 0.83) compared to the male actors (M = 0.71). No effect of Gender [F(1,138) = 0.19, p = 0.664] and no significant interaction between Actor Gender and Gender [F(1,138) = 1.07, p = 0.302] was found.

Stimulus recognizability: results

For each stimulus, the percentage of participants who correctly responded to the classification question, and the percentage of participants who reported each of the alternatives in the identification question are reported in Supplementary Table S1.

Intention classification. On average, participants correctly classified 90% of the action stimuli as communicative vs. individual (range = 72–99%; SD = 8%; Communicative stimuli, M = 91%, SD = 9%; Individual stimuli, M = 87%, SD = 8%). The actions that were less consistently recognized were “Look at the ceiling” for the communicative condition (correctly classified as communicative by 74% of the participants) and “Sneeze” for the individual condition (correctly classified as individual by 72% of the participants). Bonferroni corrected binomial tests conducted on the full sample revealed that action classification was above chance level (proportion of correct responses of 0.50) for all the action stimuli (all ps < 0.001). Chi-square tests performed on the single action stimuli (intention classification × Language) revealed that intention classification did not differ between languages for any of the 21 actions (see Supplementary Table S1).

Intention identification. On average, participants correctly described 76% of the action stimuli (range = 36–96%; SD = 17%; Communicative stimuli, M = 74%, SD = 18%; Individual stimuli, M = 80%, SD = 14%). Examples of very well recognized stimuli are “Stop” and “Imitate me” for the communicative stimuli, and “Jump” and “Look under the foot” for the individual stimuli. Bonferroni corrected binomial tests conducted on the full sample revealed that action identification was above chance level (proportion of correct responses of 0.20) for all the action stimuli (all ps < 0.001). Bonferroni corrected Chi-square tests performed on the single action stimuli (intention identification × Language) revealed that intention identification varied by Language only for the following two actions: “Go out of the way” (p < 0.001), and “No” (p < 0.001; see Supplementary Table S1). Bonferroni corrected Chi-square tests performed on the errors revealed no significant effect of Language on any of the action stimuli (all ps > 0.018), thus suggesting that, when the intention identification was incorrect, participants in the different language samples tended to select the same wrong action alternatives. Some response alternatives were thus more misleading than others in all languages.

Discussion

In the present paper we describe the multilingual CID-5, a database of 21 full-body point-light stimuli depicting two agents engaged in communicative interactions (N = 14) or performing non-communicative individual actions (N = 7) as seen from different viewpoints. For each stimulus, we provide five plausible response alternatives (only one being correct) translated into seven different languages (Chinese, Dutch, English, French, German, Italian, and Polish). Normative data collected from 140 naive participants (20 participants per language) confirmed that all the stimuli included in the multilingual CID-5 were classified as communicative vs. individual and recognized well above chance level from participants of all the seven language samples. Comparisons of global performance across different languages revealed no difference across samples in the ability to classify actions as communicative vs. individual, as indexed by the SDT parameter d′ calculated on the action classification question. Similarly, analyses on the proportion of correct responses divided by action stimulus revealed that all the 21 action stimuli were classified as communicative vs. individual in a comparable way in all language samples, with some actions being consistently very easy (e.g., ‘Stop’) and some others more difficult (e.g., ‘Look at the ceiling’) to classify. Overall, in the intention classification question participants showed a liberal criterion (negative c), that is a bias towards reporting the presence of a communicative interaction. This bias may be partially explained by the presence of a greater number of communicative action stimuli. The response bias showed a significant variation across language samples, and was especially evident in the Chinese-speaking and Polish-speaking samples. However, no bias towards reporting a communicative response alternative was found in the intention identification question, when participants were asked to select among multiple response alternatives. Thus, researchers interested in an unbiased measure of the ability to classify stimuli as communicative vs. individual may decide to rely on the intention identification question, after re-coding the response alternatives as communicative vs. individual.

For intention identification (selection of the correct response alternative), we found some individual variations in the proportion of correct responses across language samples, with French-speaking participants performing better compared to Dutch-speaking participants. Analyses divided by action stimulus revealed that 19 out of the 21 stimuli were identified in a comparable way across languages, while only two stimuli (“Go out of the way” and “No”) showed language-dependent variations. Furthermore, the error analysis showed that when the intention identification was incorrect, participants in the different language samples tended to select the same wrong alternative, suggesting that, for most of action stimuli, some response alternatives were more misleading than others in all languages.

These results provide evidence of instrument validity of the multilingual CID-5 as a new tool for the investigation of non-conventional communicative gestures in different languages. It is important to note that our data collection was designed to validate the alternatives in the different languages, and not to explore systematically cultural differences. Thus, from the present results, we cannot conclude that classification and identification of communicative gestures does not vary across cultures. First, participants in the different language samples were not balanced for age and education. Second, Chinese college students were tested in Italy and experienced thus mostly the same environment as Italian participants – a circumstance which might have well influenced their familiarity with the stimulus material. Third, and more importantly, the selected sample groups were not very distant in terms of shared cultural heritage. Future investigations should therefore remain open to the possibilities of systematic difference in nonverbal behavior across distant cultures. A final limitation relates to potential differences in social cognition and visuo-spatial abilities, not assessed in the present study. As individual differences in these abilities have been shown to correlate with recognition of social information from point-light stimuli (Okruszek et al., 2015), taking these variables into account may help to clarify the true nature of cross-linguistic and cross-cultural differences, if any, in intention-from-motion understanding.

Funding

This work received funding from the European Research Council under the European Union’s Seventh Framework Programme (FP7/2007-2013)/ERC grant agreement n. 312919.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Supplementary Material

The Supplementary Material for this article can be found online at: http://journal.frontiersin.org/article/10.3389/fpsyg.2015.01724

TABLE S1 | Percentage of participants who correctly responded to the classification question, and who reported each of the alternatives in the identification question. The “Com vs. Ind” column indicates the percentage of correct responses to the intention classification question (classification of the action as communicative vs. individual). The column “Action” indicates the percentage of responses provided for each of the five response alternatives. The first action alternative (in bold) reports the correct description. The column “Chi Square” reports the Chi Square values calculated on the proportion of correct vs. incorrect responses (intention classification and intention identification by Language). Values indicated in bold are statistically significant (∗∗∗p < 0.001).

References

Atkinson, A. P., Dittrich, W. H., Gemmell, A. J., and Young, A. W. (2004). Emotion perception from dynamic and static body expressions in point-light and full-light displays. Perception 33, 717–746. doi: 10.1068/p5096

PubMed Abstract | CrossRef Full Text | Google Scholar

Autodesk Inc (1998). Character Studio Max (Release 2) [Computer Software]. Sausalito, CA: Author.

Barrett, H. C., Todd, P. M., Miller, G. F., and Blythe, P. W. (2005). Accurate judgments of intention from motion cues alone: a cross-cultural study. Evol. Hum. Behav. 26, 313–331. doi: 10.1016/j.evolhumbehav.2004.08.015

CrossRef Full Text | Google Scholar

Brooks, A., Schouten, B., Troje, N., Verfaillie, K., Blanke, O., and van der Zwan, R. (2008). Correlated changes in perceptions of the gender and orientation of ambiguous biological motion figures. Curr. Biol. 18, R728–R729. doi: 10.1016/j.cub.2008.06.054

PubMed Abstract | CrossRef Full Text | Google Scholar

Casasanto, D. (2013). “Gesture and language processing,” in Encyclopedia of the Mind, ed. H. Pashler (Thousand Oaks, CA: Sage Publications), 372–374.

Google Scholar

Centelles, L., Assaiante, C., Etchegoyhen, K., Bouvard, M., and Schmitz, C. (2013). From action to interaction: exploring the contribution of body motion cues to social understanding in typical development and in autism spectrum disorders. J. Autism Dev. Disord. 43, 1140–1150. doi: 10.1007/s10803-012-1655-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Clarke, T. J., Bradshaw, M. F., Field, D. T., Hampson, S. E., and Rose, D. (2005). The perception of emotion from body movement in point-light displays of interpersonal dialogue. Perception 34, 1171–1180. doi: 10.1068/p5203

PubMed Abstract | CrossRef Full Text | Google Scholar

Dekeyser, M., Verfaillie, K., and Vanrie, J. (2002). Creating stimuli for the study of biological-motion perception. Behav. Res. Methods Instrum. Comput. 34, 375–382. doi: 10.3758/BF03195465

CrossRef Full Text | Google Scholar

Heberlein, A. S., Adolphs, R., Tranel, D., and Damasio, H. (2004). Cortical regions for judgments of emotions and personality traits from point-light walkers. J. Cogn. Neurosci. 16, 1143–1158. doi: 10.1162/0898929041920423

PubMed Abstract | CrossRef Full Text | Google Scholar

Johansson, G. (1973). Visual perception of biological motion and a model for its analysis. Percept. Psychophys. 14, 201–211. doi: 10.3758/BF03212378

CrossRef Full Text | Google Scholar

Kozlowski, L. T., and Cutting, J. E. (1977). Recognizing the sex of a walker from a dynamic point-light display. Percept. Psychophys. 21, 575–580. doi: 10.3758/BF03198740

CrossRef Full Text | Google Scholar

Loula, F., Prasad, S., Harber, K., and Shiffrar, M. (2005). Recognizing people from their movement. J. Exp. Psychol. Hum. Percept. Perform. 31, 210–220. doi: 10.1037/0096-1523.31.1.210

PubMed Abstract | CrossRef Full Text | Google Scholar

Macmillan, N. A., and Creelman, C. D. (2005). Detection Theory: A User’s Guide, 2nd Edn. Mahwah, NJ: Lawrence Erlbaum Associates.

Google Scholar

Manera, V., Becchio, C., Schouten, B., Bara, B. G., and Verfaillie, K. (2011a). Communicative interactions improve visual detection of biological motion. PLoS ONE 6:e14594. doi: 10.1371/journal.pone.0014594

PubMed Abstract | CrossRef Full Text | Google Scholar

Manera, V., Del Giudice, M., Bara, B. G., Verfaillie, K., and Becchio, C. (2011b). The second agent effect: communicative gestures increase the likelihood of perceiving a second agent. PLoS ONE 6:e22650. doi: 10.1371/journal.pone.0022650

PubMed Abstract | CrossRef Full Text | Google Scholar

Manera, V., Schouten, B., Becchio, C., Bara, B. G., and Verfaillie, K. (2010). Inferring intentions from biological motion: a stimulus set of point-light communicative interactions. Behav. Res. Methods 42, 168–178. doi: 10.3758/BRM.42.1.168

PubMed Abstract | CrossRef Full Text | Google Scholar

Manera, V., Schouten, B., Verfaillie, K., and Becchio, C. (2013). Time will show: real time predictions during interpersonal action perception. PLoS ONE 8:e54949. doi: 10.1371/journal.pone.0054949

PubMed Abstract | CrossRef Full Text | Google Scholar

Manera, V., von der Luhe, T., Schilbach, L., Verfaillie, K., and Becchio, C. (2015). Communicative interactions in point-light displays: choosing among multiple response alternatives. Behav. Res. Methods doi: 10.3758/s13428-015-0669-x [Epub ahead of print].

PubMed Abstract | CrossRef Full Text | Google Scholar

Marsh, A. A., Elfenbein, H. A., and Ambady, N. (2007). Separated by a common language: nonverbal accents and cultural stereotypes about Americans and Australians. J. Cross Cult. Psychol. 38, 284–301. doi: 10.1177/0022022107300275

CrossRef Full Text | Google Scholar

Montpare, J. M., and Zebrowitz-McArthur, L. (1988). Impressions of people created by age-related qualities of their gaits. J. Pers. Soc. Psychol. 55, 547–556. doi: 10.1037/0022-3514.55.4.547

PubMed Abstract | CrossRef Full Text | Google Scholar

Okruszek,Ł, Haman, M., Kalinowski, K., Talarowska, M., Becchio, C., and Manera, V. (2015). Impaired recognition of communicative interactions from biological motion in schizophrenia. PLoS ONE 10:e0116793. doi: 10.1371/journal.pone.0116793

PubMed Abstract | CrossRef Full Text

Pica, P., Jackson, S., Blake, R., and Troje, N. F. (2011). Comparing biological motion perception in two distinct human societies. PLoS ONE 6:e28391. doi: 10.1371/journal.pone.0028391

PubMed Abstract | CrossRef Full Text | Google Scholar

Pollick, F. E., Kay, J. W., Heim, K., and Stringer, R. (2005). Gender recognition from point-light walkers. J. Exp. Psychol. Hum. Percept. Perform. 31, 1247–1265. doi: 10.1037/0096-1523.31.6.1247

PubMed Abstract | CrossRef Full Text | Google Scholar

Pollick, F. E., Paterson, H. M., Bruderlin, A., and Sanford, A. J. (2001). Perceiving affect from arm movement. Cognition 82, B51–B61. doi: 10.1016/S0010-0277(01)00147-0

CrossRef Full Text | Google Scholar

Simion, F., Regolin, L., and Bulf, H. (2008). A predisposition for biological motion in the newborn baby. Proc. Natl. Acad. Sci. U.S.A. 105, 809–813. doi: 10.1073/pnas.0707021105

PubMed Abstract | CrossRef Full Text | Google Scholar

Sperber, D., and Wilson, D. (1986). Relevance: Communication and Cognition. Oxford: Blackwell.

Google Scholar

Vanrie, J., and Verfaillie, K. (2004). Perception of biological motion: a stimulus set of human point-light actions. Behav. Res. Methods Instrum. Comput. 36, 625–629. doi: 10.3758/BF03206542

CrossRef Full Text | Google Scholar

Keywords: point-light display, biological motion, communicative interaction, communicative intention, individual intention, cross-linguistic comparisons, forced choice

Citation: Manera V, Ianì F, Bourgeois J, Haman M, Okruszek ŁP, Rivera SM, Robert P, Schilbach L, Sievers E, Verfaillie K, Vogeley K, von der Lühe T, Willems S and Becchio C (2015) The Multilingual CID-5: A New Tool to Study the Perception of Communicative Interactions in Different Languages. Front. Psychol. 6:1724. doi: 10.3389/fpsyg.2015.01724

Received: 18 May 2015; Accepted: 26 October 2015;
Published: 17 November 2015.

Edited by:

Claudia Gianelli, University of Potsdam, Germany

Reviewed by:

Luke Edward Miller, University of California, San Diego, USA
Janny Christina Stapel, Uppsala University, Sweden

Copyright © 2015 Manera, Ianì, Bourgeois, Haman, Okruszek, Rivera, Robert, Schilbach, Sievers, Verfaillie, Vogeley, von der Lühe, Willems and Becchio. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Cristina Becchio, cristina.becchio@unito.it

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.