Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... ·...

20
Multisensory Research 26 (2013) 533–552 brill.com/msr Auditory Coding of Human Movement Kinematics Pia M. Vinken, Daniela Kröger, Ursula Fehse, Gerd Schmitz, Heike Brock and Alfred O. Effenberg Leibniz University Hanover, Institute of Sports Science, Hanover, Germany Received 12 July 2013; accepted 3 December 2013 Abstract Although visual perception is dominant on motor perception, control and learning, auditory infor- mation can enhance and modulate perceptual as well as motor processes in a multifaceted manner. During last decades new methods of auditory augmentation had been developed with movement sonification as one of the most recent approaches expanding auditory movement information also to usually mute phases of movement. Despite general evidence on the effectiveness of movement soni- fication in different fields of applied research there is nearly no empirical proof on how sonification of gross motor human movement should be configured to achieve information rich sound sequences. Such lack of empirical proof is given for (a) the selection of suitable movement features as well as for (b) effective kinetic–acoustical mapping patterns and for (c) the number of regarded dimensions of sonification. In this study we explore the informational content of artificial acoustical kinemat- ics in terms of a kinematic movement sonification using an intermodal discrimination paradigm. In a repeated measure design we analysed discrimination rates of six everyday upper limb actions to evaluate the effectiveness of seven different kinds of kinematic–acoustical mappings as well as short term learning effects. The kinematics of the upper limb actions were calculated based on inertial mo- tion sensor data and transformed into seven different sonifications. Sound sequences were randomly presented to participants and discrimination rates as well as confidence of choice were analysed. Data indicate an instantaneous comprehensibility of the artificial movement acoustics as well as short term learning effects. No differences between different dimensional encodings became evident thus indicating a high efficiency for intermodal pattern discrimination for the acoustically coded velocity distribution of the actions. Taken together movement information related to continuous kinematic parameters can be transformed into the auditory domain. Additionally, pattern based action discrim- ination is obviously not restricted to the visual modality. Artificial acoustical kinematics might be used to supplement and/or substitute visual motion perception in sports and motor rehabilitation. * To whom correspondence should be addressed. E-mail: [email protected] © Koninklijke Brill NV, Leiden, 2013 DOI:10.1163/22134808-00002435

Transcript of Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... ·...

Page 1: Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... · a repeated measure design we analysed discrimination rates of six everyday upper

Multisensory Research 26 (2013) 533–552 brill.com/msr

Auditory Coding of Human Movement Kinematics

Pia M. Vinken, Daniela Kröger, Ursula Fehse, Gerd Schmitz, Heike Brock and

Alfred O. Effenberg ∗

Leibniz University Hanover, Institute of Sports Science, Hanover, Germany

Received 12 July 2013; accepted 3 December 2013

AbstractAlthough visual perception is dominant on motor perception, control and learning, auditory infor-mation can enhance and modulate perceptual as well as motor processes in a multifaceted manner.During last decades new methods of auditory augmentation had been developed with movementsonification as one of the most recent approaches expanding auditory movement information also tousually mute phases of movement. Despite general evidence on the effectiveness of movement soni-fication in different fields of applied research there is nearly no empirical proof on how sonificationof gross motor human movement should be configured to achieve information rich sound sequences.Such lack of empirical proof is given for (a) the selection of suitable movement features as well asfor (b) effective kinetic–acoustical mapping patterns and for (c) the number of regarded dimensionsof sonification. In this study we explore the informational content of artificial acoustical kinemat-ics in terms of a kinematic movement sonification using an intermodal discrimination paradigm. Ina repeated measure design we analysed discrimination rates of six everyday upper limb actions toevaluate the effectiveness of seven different kinds of kinematic–acoustical mappings as well as shortterm learning effects. The kinematics of the upper limb actions were calculated based on inertial mo-tion sensor data and transformed into seven different sonifications. Sound sequences were randomlypresented to participants and discrimination rates as well as confidence of choice were analysed.Data indicate an instantaneous comprehensibility of the artificial movement acoustics as well as shortterm learning effects. No differences between different dimensional encodings became evident thusindicating a high efficiency for intermodal pattern discrimination for the acoustically coded velocitydistribution of the actions. Taken together movement information related to continuous kinematicparameters can be transformed into the auditory domain. Additionally, pattern based action discrim-ination is obviously not restricted to the visual modality. Artificial acoustical kinematics might beused to supplement and/or substitute visual motion perception in sports and motor rehabilitation.

* To whom correspondence should be addressed. E-mail:[email protected]

© Koninklijke Brill NV, Leiden, 2013 DOI:10.1163/22134808-00002435

Page 2: Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... · a repeated measure design we analysed discrimination rates of six everyday upper

534 P. M. Vinken et al. / Multisensory Research 26 (2013) 533–552

KeywordsAcoustical kinematics, auditory movement information, kinematic–acoustical mapping, motor learn-ing, movement sonification, multisensory integration, perceptual learning

1. Introduction

Although it is well known that the main sensory information guiding per-ception and recognition of movement comes from vision and proprioception(Abernethy et al., 2005; Williams et al., 2004), there exists prevailing evi-dence that object motions and actions of humans can also be identified by theauditory system. When a person knocks on a door the amplitude of the soundindicates a lower or higher amount of force, resulting in a more silent or louderknock, respectively. Based on the knocking sound the listener may be able tomake assumptions about who is knocking and which mood or intention thisperson might have. Such remarkable auditory discrimination skills may bebased partially on auditory and audiovisual mirror neurons, constituting an‘action–listening’ system as part of the perceptual system (Lahav et al., 2007)which is not only used during movement identification but also during thereenactment of distinct movements (Young et al., 2012). It is further assumed,that the perception of non-consciously represented movement features has aremarkable informational potential in the perception and execution of motortasks (Kibele, 2006). There is a substantial amount of information decodablefrom motion-dependent sound supporting processes of motor perception, con-trol and learning, which is realized by the auditory system in combination withmultisensory integration sites.

Human motions are low-frequent and commonly beyond the human hear-ing range. Hearable motion-dependent sounds are restricted to interactionswith external surfaces (in sport settings: surfaces of rackets, balls, grounds,etc.) or environmental situation-specific sounds (in daily life: rustling of cloth-ing, skin, joints, etc.). To generate additional motion acoustics, the methodof movement sonification — a technique to transform non-acoustic data intothe acoustic domain — can be used to support motor perception as well asmotor control. Even though there is initial evidence on the effectiveness ofmovement sonification (Effenberg, 2005; Rath and Rochesso, 2005; Vogt etal., 2010) it remains broadly unclear, how movement sonification becomeseffective on motor processes and how to achieve a high efficiency with re-spect to parameter selection as well as to parameter mapping. Thereto recentneurophysiological work of Schmitz and colleagues (2013) has revealed animpact of additional movement sonification of distal kinematic parameterson motor perception: When observing sonified breaststroke movements of ahuman avatar, audiovisual precision of motor perception was enhanced. Par-ticipants were better in judging swimming velocity when auditory information

Page 3: Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... · a repeated measure design we analysed discrimination rates of six everyday upper

P. M. Vinken et al. / Multisensory Research 26 (2013) 533–552 535

was given in combination with visual information compared to visual informa-tion alone. Enhanced perception correlated to enhanced activation of parts ofthe mirror neuron system. Finally it was concluded that movement sonifica-tion “amplifies the activity of the human action observation system includingsubcortical structures of the motor loop” (Schmitz et al., 2013, p. 1).

Based on these results of parameter selection we focused on a distal kine-matic parameter that is closely related to the effect of the movement: Theeffector endpoint trajectory of the acting hand. Thereby we pose the ques-tion if movement specific continuous kinematics of the acting body itselfhave comparable effects on movement perception — as has been addressedon auditory coded action effects like door-knocking sounds — when it istransferred into the acoustic domain via movement sonification. This idea issupported by previous research on movement sonification exploring behavior-neurophysiological correlations finally revealing that biological motion areas(superior temporal sulcus (STS), inferior parietal cortex (IPC), cerebellum)are involved in the integration of concordant audiovisual stimuli of complexwhole body movements (Scheef et al., 2009) and thus support motion percep-tion.

There is compelling evidence on visual motion perception indicating thathumans are able to recognize abstracted movements of their self and others(biological motion perception, Johannson, 1973; Thornton et al., 2002). Evenwhen original visual movement characteristics of human locomotion are dra-matically restricted, people are able to recognize locomotion patterns. Already1973 Johannson showed that the ability of observers to recognize abstract lo-comotion patterns depends on previous learning of similar motion patterns(Johannson, 1973). Meanwhile research on biological motion perception hasbeen expanded to other modalities, as to the auditory domain on natural mo-tion sounds (Bidet-Caulet et al., 2005). Current research is indicating that alsoartificial movement acoustics is feasible to address mechanisms of biologicalmotion perception enabling participants to assess qualitative and quantitativefeatures of gross motor movements (Schmitz and Effenberg, 2012; Schmitzet al., 2013). But it still remains unclear if complex movement patterns canbe identified based on artificial kinematic movement sonification alone andwithout respective auditory experience.

Research on audiovisual motion perception using environmental (Bidet-Caulet et al., 2005) and artificial (Làdavas, 2008) sounds complementingmovement behavior indicates that the integration of different sensory modal-ities occurs in similar and/or neighboring brain regions as unimodal percep-tion (e.g., STS, IPC, cerebellum) and thus enhances specific neural activation(Beauchamp, 2005; Bidet-Caulet et al., 2005; Làdavas, 2008). Bidet-Caulet etal. (2005) asked participants to indicate the sound direction of footsteps of awalking person presented to them during a fMRI study. Participants listened

Page 4: Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... · a repeated measure design we analysed discrimination rates of six everyday upper

536 P. M. Vinken et al. / Multisensory Research 26 (2013) 533–552

to the environmental sound of footsteps presented binaurally and were able toindicate when and to which direction (left or right of auditory field) the soundchanged from one side to the other (Bidet-Caulet et al., 2005). Following theirfMRI data, Schmitz et al. (2013) suppose that specific brain areas (STS, mirrorneuron system), which are involved in visual motion perception are activatedalso during auditory perception of artificial movement sounds as a part ofmultisensory integration (Lahav et al., 2007). Another line of argumentationrefers to a common code principle when explaining the advantages of multi-sensory stimuli compared to unisensory stimuli. It is argued that neighboringauditory and visual patches in the STS respond to modality-specific stimuli(auditory or visual) and convert them into a common code which is then passedinto intervening multisensory patches, resulting in enhanced neural responses(Beauchamp, 2005). Thus it can be assumed that the action observation systemis rather multi- than unimodal but might as well be addressed by single modal-ities. Given the impact of auditory movement information on perception andaction we wondered whether participants are able to discriminate actions justby artificial kinematic movement sounds which they had never heard beforeand never observed together with the visual stimuli within our study.

Hence, multisensory integration seems to provide mechanisms for the re-covery of sensory and spatial deficits in such a way that weakness of theprimary modality of information (e.g., vision) can be compensated by othersenses (e.g., auditory and proprioceptive system; Làdavas, 2008). Informationrepresented in each of the primary sensory systems is highly influenced fromthe other senses (Calvert et al., 2004). Studying the influence of the audi-tory system in visual detection tasks participants had to detect light spots intheir visual field, which were either presented alone or in combination with acongruent or incongruent auditory signal. Normal subjects and patients withvisual deficits showed enhanced multisensory neural responses, when stimuliof different sensory modalities were temporally (normal subjects) and spa-tially (patients with visual deficits) coincident. On the other hand there was noenhancement of multisensory responses when stimuli of different modalitieswere presented incongruently (Làdavas, 2008) and on single neuron level evenresponse reduction occurred (Stein and Meredith, 1993). Taken together it isgetting evident that different forms of auditory stimuli may not only result indifferent auditory movement information but may also alter multisensory inte-gration, clearly indicating the necessity of an appropriate intermodal stimulusconfiguration. The recovery of somatosensation after stroke e.g., can be sup-ported with multisensory training settings by integrating auditory informationin visuo-motor tasks (Carey et al., 2002). The innate ability of the human brainto perceive and integrate multisensory events is reinforced and thus facilitatesmotor learning and relearning.

Page 5: Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... · a repeated measure design we analysed discrimination rates of six everyday upper

P. M. Vinken et al. / Multisensory Research 26 (2013) 533–552 537

Additional empirical evidence comes from Parkinson’s disease therapy(Ringenbach et al., 2011): In a study on the accuracy of upper limb actions,younger and older adult patients were instructed verbally (up/down), visually(visual stimuli at top/bottom) or auditory (high/low tone) to draw lines. Over-all, participants received lower coefficient of variance of cycle time, lowervariability of amplitude and lower variability of relative phase in the audi-tory and visually instructed conditions compared to the verbally instructedcondition. In the auditory instructed condition participants received a moreconsistent coordination of movement (Ringenbach et al., 2011). The authorsargue that the benefit of the auditory instructed condition may be related tothe timing nature of the line drawing task. Especially on Parkinson’s diseaseauditory cues seem to support a kinesthetic focus on the sensation of executedmovements which may improve cognitive and motor functions (Ringenbach etal., 2011). Obviously auditory stimuli are well suitable for guiding movementexecution in timing tasks on both, healthy as well as challenged people.

When trying to create efficient artificial auditory movement informationtwo main aspects have to be considered:

(1) Movement acoustics should possess a good correlation between the finetemporal structure of acoustical and optical stimuli to address multisen-sory integration sites selectively (Parise et al., 2012).

(2) Movement acoustics should mediate auditory information about the re-lated action pattern efficiently.

Ad (1): The mapping of sound to action patterns can be realized by measur-ing kinematic features of the movement and transforming them into sound,as has been realized with ‘movement sonification’ on rowing (Effenberg etal., 2011) or with the ‘Bodycoder System’ (Wilson-Bokowiec and Bokowiec,2006). Movement sonification is a method to expand movement acoustics tosilent phases of actions which are usually not evoking attending sounds (e.g.,actions or gestures of the limbs). Such artificial action-specific sounds containinformation about the related movement kinematics, like the course of a tra-jectory, creating a new kind of movement information that we call acousticalkinematics. In contrast to discrete environmental sounds the selected move-ment features can be transformed continuously to acoustics — independentlyof ecological distractors. In our study velocity and position information ofthe effectors’ endpoint trajectory is mapped to amplitude, frequency, spectralcomposition (of the sound) and stereo in different combinations (cf. Table 1and Fig. 1). However, when listening to acoustical kinematics associated tim-bre information is missing because an electronic sound sample is used.

Ad (2): The assessment of the informational content of acoustical kine-matics can be realized by comparing the impact of the different kinematic

Page 6: Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... · a repeated measure design we analysed discrimination rates of six everyday upper

538 P. M. Vinken et al. / Multisensory Research 26 (2013) 533–552

Table 1.Composition of the seven parameter combinations used for the sound sequences. Italic abbrevi-ations in the left column refer to the combination of parameters in the other columns, e.g., ASCfor amplitude and spectral composition, etc.

Parameter Amplitude (A) Spectral composition (SC) Frequency (F) Stereo (S)combination (absolute velocity) (anterior–posterior) (cranial–caudal) (medio-lateral)

ASC � �AF � �AS � �ASCF � � �ASCS � � �AFS � � �ASCFS � � � �

Figure 1. Description of sonified action plane and corresponding sound parameter of the distalend effector, the third, right metacarpophalangeal joint.

movement sonifications in a discrimination task. If reliability of auditory dis-crimination of action patterns correlates with the amount of available move-ment information, as many information of that pattern as possible should becoded by sonification.

Page 7: Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... · a repeated measure design we analysed discrimination rates of six everyday upper

P. M. Vinken et al. / Multisensory Research 26 (2013) 533–552 539

Summed up, when studying the effects of intersensory processing and mul-tisensory integration there is compelling evidence that combined congruentinformation can enhance perceptual learning (Beauchamp, 2005; Croix et al.,2010; Làdavas, 2008; Shams and Seitz, 2008) as well as training efficiencyin adult learners (Seitz et al., 2006). Furthermore research shows that multi-sensory information is most beneficial when presented congruently with goodintermodal correlation (Parise et al., 2012; Shams and Seitz, 2008) and al-locates the effectiveness of either environmental sounds or simple forms ofmusic on recognition (Lahav et al., 2007), differentiation (Bidet-Caulet etal., 2005), reenactment (Young et al., 2012) and learning (Shams and Seitz,2008) of human movement. Additional movement sonification enhances theperception and reproduction accuracy of gross motor movements (Effenberg,2005) pointing to a broad application variety of artificial auditory move-ment information in sports and rehabilitation when addressing motor functionsand tuning the multisensory framework. A goal could be to support ath-letes in technically oriented sport disciplines like gymnastics or high diving.In neuro-motor rehabilitation a main goal could be to enhance the recoveryof every day-actionability of the upper limbs when being affected e.g. afterstroke.

In our exploratory study different forms of kinematic movement sonifica-tions are compared in terms of enabling a reliable auditory discrimination andcross-modal recognition (visual instruction and auditory discrimination) of sixeveryday upper limb actions. In detail, we wanted to find out if people are ableto discriminate movement specific artificial sound sequences never heard be-fore. Movement sonification was used to transform kinematic data of the distalend effector into the acoustic domain. Velocity and position information of theacting right arm was calculated based in inertial motion sensor data. Six differ-ent everyday upper limb actions were sonified using seven different parametercombinations each to explore possible differences of the discrimination fre-quencies. The seven parameter combinations were used to grade the sonifiedparameter complexities (from two up to four sonified parameters). Addition-ally it is hypothesized that upper limb actions can be differentiated accordingto their sonified kinematic parameters. For example, drawing a circle shouldbe discriminated best, when its main action plane (e.g., medio-lateral andanterior–posterior plus velocity) is represented in the sound sequences of themovement sonification, compared to sound sequences where one or even bothparameters are not represented (e.g., cranial–caudal plus velocity). Thereforewe establish three categories of the six upper limb actions based on the simi-larity of the action structure: Action Category I — drawing a circle and stirringin a pot. This category is representing slow, circular actions in a common mainaction plane with structural differences in terms of velocity, smoothness andnumber of circular motions. These structural differences are detectable within

Page 8: Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... · a repeated measure design we analysed discrimination rates of six everyday upper

540 P. M. Vinken et al. / Multisensory Research 26 (2013) 533–552

the three sonified parameters, namely spectral composition (anterior–posteriorplane), frequency (cranial–caudal plane) and stereo (medio-lateral plane). Ac-tion Category II — pouring water from one glass to another and drinking froma glass. This category is representing slow, linear actions which are detectableon their main action plane with structural differences in terms of the heightof the key phase (pouring takes place at the level of the table, drinking atthe level of the mouth). These structural differences are detectable within thetwo sonified parameters spectral composition (anterior–posterior plane) andstereo (medio-lateral plane). Action Category III — rasping one’s nails andbrushing one’s teeth. This category is representing fast, cyclical actions whichare detectable on the main action plane with structural differences in terms ofmovement frequency and the height of the actions’ main phase. These struc-tural differences are detectable within only one sonified parameter namelyfrequency (cranial–caudal plane).

2. Material and Method

2.1. Participants

In expectation of a medium effect (Cohen’s f � 0.25) we recruited N = 28right-handed students (mean score and SE of Edinburgh Inventory = 90.34 ±2.12; Oldfield, 1971) of the Leibniz University Hanover to participate in ourstudy (n = 14 males, age = 24.71 ± 2.60 yrs; n = 14 females, age = 24.53 ±3.28 yrs). All participants indicated that they had normal hearing. None ofthem reported to have any prior experiences concerning the experimental pro-cedures used in our study (movement sonification). Participants were asked todiscriminate different sonification sequences and rate confidence of choice foreach of the 126 trials (3 repetitions × 6 actions × 7 parameter combinations;see also Section 2.3). Participants were informed about the general procedureand gave their written consent prior to the study. The experimental task wascarried out regarding the guidelines of the Central Ethical Committee of theLeibniz University Hanover.

2.2. Material

2.2.1. Preparation of SonificationsA person was equipped with five inertial sensors (MTx miniature inertial3DOF orientation tracker; Xsens Technologies BV, Enschede, The Nether-lands) on the right arm. The sensors (size: 38 mm × 53 mm × 21 mm, weight:30 g) were attached on the centers of the right scapula, the right upper arm,the right lower arm, the right back of the hand and the center of the right backsof the fingers. Inertial sensors provide 3D acceleration (up to 18 g), 3D rateof turn (up to 1.200°/s), three degrees of freedom orientation and a samplingrate up to 100 Hz depending on the number of sensors used (Fong and Chan,

Page 9: Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... · a repeated measure design we analysed discrimination rates of six everyday upper

P. M. Vinken et al. / Multisensory Research 26 (2013) 533–552 541

Figure 2. Pictures and sound sequences’ hypsogrammes of amplitude of our six everyday upperlimb actions.

2010; Helten et al., 2011). The person was videotaped when executing the sixdifferent everyday upper limb actions: (1) drawing a circle and (2) stirring ina pot (Action Category I); (3) pouring water from one glass to another and(4) drinking from a glass (Action Category II); (5) rasping one’s nails and(6) brushing one’s teeth (Action Category III; cf. Fig. 2), while raw data fromthe inertial sensors were recorded. The videotaped sequences were used forlater visual instruction (see Section 2.3). In a next step, a four-segment for-ward kinematic model of the right arm was used to determine position andvelocity information of the right hand needed for the sonification (Brock etal., 2012).

Four specific kinematic parameters from the four-segment model were se-lected to modulate four different sound parameters based on the sound ‘JupiterLead’, which was generated by the Synth Modul ‘SonicCell’ (Roland Ger-many GmbH, Nauheim, Germany; Brock et al., 2012). Key features of soundsare their frequency, amplitude or intensity, spectral patterns and — in the caseof an azimuth angle different from zero — temporal and intensity differencesbetween ears. In the present study positions of the right metacarpophalangealjoint in medio-lateral axis were mapped on interaural intensity differences(stereo) and positions in the anterior–posterior axis on spectral patterns (spec-tral composition of the sound). The complete spectral composition of thesound was provided at the lowest distance between the right metacarpopha-langeal joint and the body center (highest partial at 12 481 Hz). A movementto the largest distance reduced the partials linearly to a maximum value of6020 Hz. Both features are still related to the ecologic mechanisms of soundlocalization (Kapralos et al., 2003). Movements in the cranial–caudal axis

Page 10: Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... · a repeated measure design we analysed discrimination rates of six everyday upper

542 P. M. Vinken et al. / Multisensory Research 26 (2013) 533–552

were mapped on frequency, taken into account profound perceptual discrimi-nation abilities of this feature in humans (Carlile, 2011). Movements modifiedthe sound ‘Jupiter Lead’ linearly between frequencies of 73 Hz and 294 Hz.Velocity changes were displayed as amplitude changes, assuming an eco-logical relationship between energy of movement and energy of sound (cf.Table 1 and Fig. 1). Overall seven different parameter combinations wererealized consisting of three two-dimensional parameter combinations, threethree-dimensional parameter combinations, and one four-dimensional param-eter combination (cf. Table 1). This procedure resulted in 6 (upper limb ac-tions) × 7 (parameter combinations) = 42 different sonification sequences.Overall durations as well as duration of movement’s main phases of all sonifi-cation sequences were identical.

2.3. Procedure

Participants were seated at a table 450 cm in front of a 200 × 270 cm widevideo screen wearing circumaural headphones (Beyer Dynamic, DT 100, 30–20 000 Hz). Auditory stimuli were displayed with pleasant overall amplitudeup to 65 dB. An instruction video of the six upper limb actions was initiallypresented to each participant for three times. Additionally a chart with picturesand denominations of each action was placed in front of the participant. Noaudiovisual instruction was given at any time. After watching the instructionvideo participants were instructed to a six-item forced choice task to discrim-inate the presented sonification sequence to one of the six upper limb actions.Additionally the confidence of choice had to be ranked on a percentage scale(0–100%, Likert scale). Participants received no feedback at any time.

One hundred and twenty six trials were presented to participants in a stan-dardized setting: Each trial consisted of (1) the trial number presented on thevideo screen (2.0 s), (2) the sonification sequence presented via headphones(overall duration 12.0 s; action duration 8.0 s), (3) a timeslot for the discrimi-nation decision and confidence rating (6.5 s) completed by (4) a pink noisesequence (1.0 s) leading to an overall duration of a single trial of 21.5 s.The study consisted of seven blocks, each containing 6 (actions) × 3 (rep-etitions) = 18 trials. Each block represented a parameter combination (seeSection 2.2). Order of blocks were counterbalanced over all participants andtrial sequences were pseudo-randomized including the rule not to present aparticular action more than twice in a row.

2.4. Data Analysis

A significance criterion of α = 5% was established for all results reported. Inorder to assess the discrimination rates and chance level t-tests for dependentsamples were calculated, including participant’s numbers of correct answersas dependent variable. Cohen’s d was calculated as an effect size for all re-

Page 11: Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... · a repeated measure design we analysed discrimination rates of six everyday upper

P. M. Vinken et al. / Multisensory Research 26 (2013) 533–552 543

ported t-values (Cohen, 1988). Separate univariate ANOVAs with repeatedmeasures for the factors Block, Parameter Combination, Action Category andAction with number of correct answers as dependent variable were computed.The dependent variable confidence rating was subjected to the same statisticalprocedure. Post-hoc analyses were calculated using Tukey’s HSD post-hoctest. Cohen’s f was calculated as an effect size for all reported F -values(Cohen, 1988). Additionally a correlation analysis was conducted in order toexplore the relationships in correct answers between the parameter combina-tions.

3. Results

Participant’s numbers of correct answers are shown in Fig. 3. Participantswere able to discriminate upper limb actions above chance level from thefirst block on. Effect sizes ranged from Cohen’s d = 1.761 (block 1) to Co-hen’s d = 2.402 (block 6), thus indicating large effects. Additionally par-ticipants increased their discrimination rate with the duration of the studyF(6,162) = 16.444, p < 0.05, Cohen’s f = 0.780, indicating a large effect.A post-hoc analysis on the factor Block revealed significant differences be-tween the first block and blocks 3 to 7 as well as between the second blockand blocks 3 to 7, thus leading to the assumption that participants needed 36trials to learn the mapping between sound sequences and corresponding ac-tions (Fig. 3a). Figure 3b shows participant’s number of correct answers overthe seven different parameter combinations. Discrimination rates did not differsignificantly between parameter combinations F(6,162) = 0.573, p = 0.752,Cohen’s f = 0.145. Thus higher dimensional sonifications did not lead to en-hanced performance. Figure 3c shows participant’s number of correct answersover the three different action categories. There was a significant main ef-fect on Action Category F(2,54) = 23.280, p < 0.05, Cohen’s f = 0.929indicating a large effect. A post-hoc analysis revealed significant differencesbetween rasping/brushing (Action Category III) and drawing/stirring (ActionCategory I) as well as between rasping/brushing (Action Category III) andpouring/drinking (Action Category II), thus leading to the assumption thataction structure may influence detection ability in our discrimination taskand fast, cyclical actions can be discriminated best. Furthermore there wasa significant main effect on Action, F(5,135) = 7.216, p < 0.05, Cohen’sf = 0.517, indicating a large effect. A post-hoc analysis on the factor Ac-tion revealed significant differences between pouring and stirring, pouring anddrawing, pouring and drinking, pouring and rasping as well as between stirringand rasping (Fig. 3d). ANOVAs (Action × Parameter Combinations, ActionCategory × Parameter Combinations) did not reveal significant interactions

Page 12: Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... · a repeated measure design we analysed discrimination rates of six everyday upper

544 P. M. Vinken et al. / Multisensory Research 26 (2013) 533–552

Fig

ure

3.Pa

rtic

ipan

t’snu

mbe

rsof

corr

ect

answ

ers

(mea

SE)

of(a

)th

ese

ven

bloc

ks(1

–7)

ofth

est

udy

prog

ress

,ind

epen

dent

ofpa

ram

eter

com

-bi

natio

n,(b

)th

ese

ven

diff

eren

tpar

amet

erco

mbi

natio

ns,(

c)th

eth

ree

diff

eren

tact

ion

cate

gori

esan

d(d

)th

esi

xdi

ffer

ente

very

day

uppe

rlim

bac

tions

.∗ i

ndic

ates

sign

ifica

ntdi

ffer

ence

sac

cord

ing

toT

ukey

HSD

post

-hoc

test

,+in

dica

tes

sign

ifica

ntdi

ffer

ence

toch

ance

leve

l(da

shed

line)

.

Page 13: Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... · a repeated measure design we analysed discrimination rates of six everyday upper

P. M. Vinken et al. / Multisensory Research 26 (2013) 533–552 545

F(30,810) = 1.326, p = 0.115, Cohen’s f = 0.222 and F(12,324) = 1.571,p = 0.099, Cohen’s f = 0.241.

Figure 4a shows participant’s confidence ratings during study progress. Par-ticipants increased their confidence rate during study progress F(6,162) =12.031, p < 0.05, Cohen’s f = 0.668, indicating a large effect. A post-hocanalysis on the factor Block revealed ten significant differences, whereas par-ticipants in the first two blocks are significantly less confident than in blocks 4to 7. The same is true for the third block compared to blocks 6 and 7, support-ing the assumption that participants increased confidence about discriminat-ing actions from each other. Figure 4b shows participant’s confidence ratingover all different parameter combinations. No differences between parame-ter combinations became significant (F(6,162) = 0.700, p = 0.650, Cohen’sf = 0.161). Figure 4c shows participant’s confidence rating over the threedifferent action categories documenting a significant main effect on ActionCategory F(2,54) = 17.883, p < 0.05, Cohen’s f = 0.815, indicating a largeeffect. A post-hoc analysis revealed significant differences between draw-ing/stirring (Action Category I) and rasping/brushing (Action Category III)as well as between drawing/stirring (Action Category I) and pouring/drinking(Action Category II). Additionally there was a significant main effect on Ac-tion, F(5,135) = 18.728, p < 0.05, Cohen’s f = 0.833, indicating a largeeffect. A post-hoc analysis on the factor Action revealed significant differencesbetween drawing and brushing, drawing and rasping, drawing and drinking,stirring and brushing, stirring and rasping, stirring and drinking as well as be-tween pouring and brushing, pouring and rasping, and pouring and drinking(Fig. 4d).

Results show that participants were able to discriminate six everyday upperlimb actions above chance level from the first block on and to increase theirdiscrimination and confidence rate during the 126 trials. Furthermore resultsindicate differences between the discrimination and confidence rate of the sixdifferent upper limb actions chosen for this study. Additionally, we calculateda correlation analysis for the number of correct answers of our seven parame-ter combinations revealing that 18 of 21 possible combinations correlated witheach other (Product Moment Correlation ranging between r = 0.40, . . . ,0.77).Because no differences in discrimination and confidence rate between param-eter combinations became evident and all parameter combinations correlatedwith each other, there seems to be an overall action specific information en-abling participants to discriminate upper limb actions from each other.

4. Discussion

The aim of the study was to explore if artificial movement acoustics based onkinematic movement parameters can be decoded by naïve listeners. Results

Page 14: Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... · a repeated measure design we analysed discrimination rates of six everyday upper

546 P. M. Vinken et al. / Multisensory Research 26 (2013) 533–552

Fig

ure

4.Pa

rtic

ipan

t’sco

nfide

nce

ratin

gs(m

ean

±SE

)of

(a)

the

seve

nbl

ocks

(1–7

)of

the

stud

ypr

ogre

ss,

inde

pend

ent

ofpa

ram

eter

com

bina

tion,

(b)

the

seve

ndi

ffer

entp

aram

eter

com

bina

tions

,(c)

the

thre

edi

ffer

enta

ctio

nca

tego

ries

and

(d)

the

six

diff

eren

teve

ryda

yup

per

limb

actio

ns.∗

indi

cate

ssi

gnifi

cant

diff

eren

ces

acco

rdin

gto

Tuk

eyH

SDpo

st-h

octe

st.

Page 15: Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... · a repeated measure design we analysed discrimination rates of six everyday upper

P. M. Vinken et al. / Multisensory Research 26 (2013) 533–552 547

show that participants were able to discriminate six everyday upper limb ac-tions even when being naïve about the utilized kinematic–acoustic mappingstructure and being inexperienced with movement sonification at all. Addi-tionally participants obtained high discrimination rates already from the firstblock on (42.46% score compared to 16.67% chance level) and furthermorerapidly increased within 40 min of treatment duration (up to 68.65% withinthe last block) indicating short term learning.

Discrimination rates were obviously independent from the complexity ofthe sonification and treatments were intercorrelated suggesting that there is anoverall action specific information enabling participants to discriminate up-per limb actions from each other. Previous research has indicated that theauditory system is capable of supplementing visual perception by decodingmotion related sounds and addressing mechanisms of multisensory integra-tion (Beauchamp, 2005) thus integrating visual as well as auditory biologicalmotion information (Bidet-Caulet et al., 2005). Results of the present studyhighlight the importance of exclusively presented artificial movement acous-tics which can obviously provide action specific information by presumablyenriching biological movement perception audiovisually.

Based on the study of Young et al. (2012), who had participants perceiveand reenact stride length and cadence of natural and synthesized walkingsounds, and the results of our study we argue that the auditory identificationof human movements, based on a continuously auditory coded kinematic tra-jectory, could be realized in a similar manner as visual identification of humanmovements, restricted to point-light displays. Neuroscientific research revealsthat mirror neurons may cause and/or support identification, recognition andcomprehension of observed human movements (Rizzolatti, 2005; Young etal., 2012). When dealing with new and unnatural sound sequences Lahav andcolleagues (2007) argue that the degree of motor activation is modulated by alearned mapping between the movement and its acoustics. This is not trans-ferable to our setting: There was no feedback given to the participants at anytime and they were not asked for any specific motor activity. The enhance-ment of the discrimination rate during continued treatment therefore shouldbe based on increasing knowledge about sound sequences of concurring ac-tion patterns enabling the participants to a more reliable discrimination. Thison one hand might be reflected in our data in terms of an increase of correctanswers during study progress (Fig. 3a). However, on the other hand, par-ticipants in our study were already able to correctly identify actions abovechance from the first block on when being naïve about the mapping underlyingour sound sequences. That might indicate that action specific-multimodal-representations could be addressed also by an acoustic transformation of theirkinematics.

Page 16: Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... · a repeated measure design we analysed discrimination rates of six everyday upper

548 P. M. Vinken et al. / Multisensory Research 26 (2013) 533–552

In contrast to our hypothesis identification rates were not related to thedifferent dimensional sonifications. Even listening to low-dimensional sonifi-cations resulted in high identification rates, and it might be concluded that eachkind of sonification or treatment, resp. provided similar amounts of informa-tion about a certain action pattern. The intercorrelations between treatmentsimply that performances are at least partially based on the same factor. Thisfactor might be interpreted in terms of inter-individual differences of the par-ticipants, but data do not support this view. The correlation might rather berooted in the design of the kinematic sonifications: All sonifications containedthe velocity of the action that was mapped onto amplitude, and Fig. 2 illus-trates that the velocity differed between submovements of actions. Thus theamplitude profiles informed about action structures and might have been suf-ficient for their discrimination. Actions of the same category differed less intheir amplitude profile than actions of different categories. Accordingly errorswere larger within than between categories, confirming that discriminationrates depend on the (dis)similarity of action structures. Directional informationof actions provided by frequency, spectral composition and stereo modifica-tions might have been of less importance for the present perceptual task, butit should not be concluded that the display of a single dimension is sufficientfor action discrimination. Considering that velocity was defined as the lengthof the action vector in three-dimensional space within a given time, it reflectspositional changes in time independent of action direction. Therefore it mightbe understood as an integrating four-dimensional parameter.

The aspect that participants in our study were able to identify everyday up-per limb actions based on the acoustical kinematics is supported by fMRI dataillustrating that listening to movement specific sound (acquired piano pieces)activates motor-related brain networks (Broca’s area, premotor region, intra-parietal sulcus and the inferior parietal region). Such areas are accommodatingthe human mirror neuron system and they are part of a ‘hearing–doing sys-tem’ that seems to be highly dependent of the participant’s motor repertoire(Lahav et al., 2007). Therefore, when people’s perception might be restrictedby perceptual deficits (e.g., rehabilitation after stroke) or perception underlieshigh demands (e.g., high performance sports) auditory information guidingmovement behavior may have a supportive role in motor learning and control(Làdavas, 2008; Schmidt and Lee, 2011). Resulting from that and the currentresearch discussed above we want to highlight two specific practical impli-cations: First, if the so-called ‘hearing–doing system’ depends on the humanmotor repertoire, auditory stimuli may play a significant role in the relearningof movements. This is of keen interest for neuro-motor rehabilitation, sen-sory therapy and/or high performance sport, when formerly mastered actionscould not be performed accurately anymore and thus have to be or shouldbe reacquired. Second, Gonzalez and colleagues (2010) could show that pre-

Page 17: Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... · a repeated measure design we analysed discrimination rates of six everyday upper

P. M. Vinken et al. / Multisensory Research 26 (2013) 533–552 549

motor auditory information based on kinematic aspects of objects to grasphelps to anticipate prehension and individuals can use auditory cues to pre-dict the forces required to lift the object. These findings expand the possibleforms of implementing auditory information in the relearning of movementin such a way that pre-motor auditory cues may be applied in a priming con-text and thus establish new forms for developing precision in rehabilitationand/or high performance sport (e.g., shooting) and expand research and appli-cation in multisensory information integration. For example, during recoveryof somatosensation after stroke the implementation of multisensory trainingregimen — like for instance integrating auditory information in visuo-motortasks — might reinforce the innate ability of the human brain to perceive andintegrate multisensory events and thus facilitates motor learning and relearn-ing (Carey et al., 2002).

We are aware of several limitations of our study and want to highlight threespecific ones. First, due to technical limitations and the novelty of implement-ing this method we decided in this study to focus on single-arm upper limbactions, thus sonifying only one arm. For future studies it would be of highinterest to expand our methodology and implement both arms and even thewhole body for movement sonification. Second, the question remains openhow different target groups of individuals may respond in comparable stud-ies. In our study we asked healthy participants to evaluate a healthy model.However, it would be interesting for future studies to contrast healthy andchallenged individuals and their responses to our healthy model, or to exam-ine if rates of action discrimination could be modulated by presenting differentstimuli, generated by individuals themselves compared to our model or evenother (well known vs. unknown) individuals. Recently Schmitz and Effenberg(2012) have shown that subjects identify own movements above chance levelwhen they hear own movements, but are not able to correctly reject own move-ments when they hear movements of other persons. Thus the discrimination ofa person is better, when sonified movements correspond to the own movement.It would be interesting to investigate whether the same holds for action identi-fication. If so, auditory displays of actions should be adapted to the individualskills or action repertoire to optimize the effects in rehabilitation and sports.Third, future studies in this field of applied sonification should additionallyfocus on simple actions, like for instance pointing tasks. Thereby it would beof high interest to see if for example auditory position information itself issufficient to support motor precision and accuracy of the upper limbs. Addi-tionally it is of high interest and relevance to question if there are other, maybeimproving, mappings of kinematic acoustics.

Based on that we argue that auditory information seem to enhance the judg-ment accuracy of human actions and exhibit potential for improving motorlearning and control. Thereby future studies should try to focus on the amount,

Page 18: Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... · a repeated measure design we analysed discrimination rates of six everyday upper

550 P. M. Vinken et al. / Multisensory Research 26 (2013) 533–552

the function and the content of the auditory information in (multi-)sensorylearning and control.

Acknowledgements

We would like to thank the European Union for financially supporting this re-search by the European Regional Development Fund (ERDF, Project Nr. W2-80118660).

References

Abernethy, B., Hanrahan, S. J., Kippers, V., Mackinnon, L. T. and Pandy, M. G. (2005). TheBiophysical Foundations of Human Movement, 2nd edn. Human Kinetics, Champaign, IL,USA.

Beauchamp, M. S. (2005). See me, hear me, touch me: multisensory integration in lateraloccipital–temporal cortex, Curr. Opin. Neurobiol. 15, 145–153.

Bidet-Caulet, A., Voisin, J., Bertrand, O. and Fonlupt, P. (2005). Listening to a walking humanactivates the temporal biological motion area, Neuroimage 28, 132–139.

Brock, H., Schmitz, G., Baumann, J. and Effenberg, A. O. (2012). If motion sounds: movementsonification based on inertial motion data, in: Proceedia Engineering 34, P. Drahne andJ. Sherwood (Eds), pp. 556–561. Elsevier, Amsterdam, Netherlands.

Calvert, G., Spence, C. and Stein, B. E. (2004). The Handbook of Multisensory Processes. MITPress, Cambridge, MA, USA.

Carey, L. M., Abbott, D. F., Puce, A., Jackson, G. D., Syngeniotis, A. and Donnan, G. A.(2002). Reemergence of activation with poststroke somatosensory recovery: a serial fMRIcase study, Neurology 59, 749–752.

Carlile, S. (2011). Psychoacoustics, in: The Sonification Handbook, T. Hermann, A. Hunt andJ. Neuhoff (Eds), pp. 549–552. Logos, Berlin.

Cohen, J. (1988). Statistical Power Analysis for the Behavioral Sciences, 2nd edn. LawrenceErlbaum, New York, NY, USA.

Croix, G., Lejeune, L., Anderson, D. I. and Thouvarecq, R. (2010). Light fingertip contact onthigh facilitates handstand balance in gymnasts, Psychol. Sport Exerc. 11, 330–333.

Effenberg, A. O. (2005). Movement sonification — effects on perception and action, IEEEMultimed. 12, 53–59.

Fong, D. T.-P. and Chan, Y.-Y. (2010). The use of wearable inertial motion sensors in humanlower limb biomechanics studies: a systematic review, Sensors 10, 11556–11565.

Gonzalez, D. A., Dubrowski, A. and Carnahan, H. (2010). The use of auditory cues in anticipa-tory control of grasping forces, Motor Control 14, 255–264.

Helten, T., Brock, H., Müller, M. and Seidel, H.-P. (2011). Classification of trampoline jumpsusing inertial sensors, Sports Eng. 14, 155–164.

Johannson, G. (1973). Visual perception of biological motion and a model for its analysis,Percept. Psychophys. 14, 201–211.

Kapralos, B., Jenkin, M. R. M. and Milios, E. (2003). Auditory Perception and Spatial (3D)Auditory Systems. Technical report CS-2003-07, York University, UK.

Page 19: Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... · a repeated measure design we analysed discrimination rates of six everyday upper

P. M. Vinken et al. / Multisensory Research 26 (2013) 533–552 551

Kibele, A. (2006). Non-consciously controlled decision making for fast motor reactions insports — a priming approach for motor responses to non-consciously perceived movementfeatures, Psychol. Sport Exerc. 7, 591–610.

Làdavas, E. (2008). Multisensory-based approach to the recovery of unisensory deficit, Ann.N.Y. Acad. Sci. 1124, 98–110.

Lahav, A., Saltzman, E. and Schlaug, G. (2007). Action representation of sound: audiomotorrecognition network while listening to newly acquired actions, J. Neurosci. 27, 308–314.

Oldfield, R. C. (1971). The assessment and analysis of handedness: the Edinburgh inventory,Neuropsychologia 9, 97–113.

Parise, C. V., Spence, C. and Ernst, M. O. (2012). When correlation implies causation in multi-sensory integration, Curr. Biol. 22, 46–49.

Rath, M. and Rocchesso, D. (2005). Continuous sonic feedback from a rolling ball, IEEE Mul-timed. 12, 60–69.

Ringenbach, S. D. R., van Gemmert, A. W. A., Shill, H. A. and Stelmach, G. E. (2011). Auditoryinstructional cues benefit unimodal and bimanual drawing in Parkinson’s disease patients,Hum. Mov. Sci. 30, 770–782.

Rizzolatti, G. (2005). The mirror neuron system and its function in humans, Anat. Embryol.210, 419–421.

Scheef, L., Boecker, H., Daamen, M., Fehse, U., Landsberg, M. W., Granath, D.-O., Mechling,H. and Effenberg, A. O. (2009). Multimodal motion processing in area V5/MT — evidencefrom an artificial class of audio-visual events, Brain Res. 2152, 94–104.

Schmidt, R. A. and Lee, T. (2011). Motor Control and Learning — A Behavioral Emphasis.Human Kinetics, Champaign, IL, USA.

Schmitz, G. and Effenberg, A. O. (2012). Perceptual effects of auditory information about ownand other movements, in: Proceedings of the 18th International Conference on AuditoryDisplay, Atlanta, GA, USA, pp. 89–94.

Schmitz, G., Mohammadi, B., Hammer, A., Heldmann, M., Samii, A., Münte, T. F. and Effen-berg, A. O. (2013). Observation of sonified movements engages a basal ganglia frontocorti-cal network, Neuroscience 14, 1–11.

Seitz, A. R., Kim, R. and Shams, L. (2006). Sound facilitates visual learning, Curr. Biol. 16,1422–1427.

Shams, L. and Seitz, A. R. (2008). Benefits of multisensory learning, Trends Cogn. Sci. 12,411–417.

Stein, B. E. and Meredith, M. A. (1993). The Merging of the Senses. MIT Press, Cambridge,MA, USA.

Thornton, I. M., Rensink, R. A. and Shiffrar, M. (2002). Active versus passive processing ofbiological motion, Perception 31, 837–853.

Vogt, K., Pirro, D., Kobenz, I., Höldrich, R. and Eckel, G. (2010). PhysioSonic — evaluatedmovement sonification as auditory feedback in physiotherapy, in: Auditory Display, 6th In-ternational Symposium, CMMR/ICAD 2009, Lecture Notes in Computer Science, S. Ystad,M. Aramaki, R. Kronland-Martinet and K. Jensen (Eds), pp. 103–120. Springer, Berlin, Ger-many.

Williams, A. M., Ward, P. and Smeeton, N. J. (2004). Perceptual and cognitive expertise insport — implications for skill acquisition and performance enhancement, in: Skill Acquisi-

Page 20: Auditory Coding of Human Movement Kinematicssonification-online.com/.../Vinken-et-al_Auditory... · a repeated measure design we analysed discrimination rates of six everyday upper

552 P. M. Vinken et al. / Multisensory Research 26 (2013) 533–552

tion in Sport — Research, Theory and Practice, A. M. Williams and N. J. Hodges (Eds),pp. 328–347. Routledge, New York, NY, USA.

Wilson-Bokowiec, J. and Bokowiec, M. A. (2006). Kinaesonics: the intertwining relationshipof body and sound, Contemp. Music Rev. 25, 47–57.

Young, W., Rodger, M. and Craig, C. M. (2012). Perceiving and reenacting spatiotemporalcharacteristics of walking sounds, J. Exp. Psychol. Hum. Percept. Perform. 39, 464–476.