multisensory integration at peak limb velocity · 2015. 12. 2. · localization when the limb is...

87
Multisensory Integration at Peak Limb Velocity by Tristan Defrancesco Loria A Master’s thesis submitted in conformity with the requirements for the degree of Master of Science Exercise Sciences University of Toronto © Copyright by Tristan Loria 2015

Upload: others

Post on 19-Jan-2021

2 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

Multisensory Integration at Peak Limb Velocity

by

Tristan Defrancesco Loria

A Master’s thesis submitted in conformity with the requirements

for the degree of Master of Science

Exercise Sciences

University of Toronto

© Copyright by Tristan Loria 2015

Page 2: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

ii

Abstract

Multisensory Integration at Peak Limb Velocity

Tristan Defrancesco Loria

Master of Science

Graduate Department of Exercise Sciences

University of Toronto

2015

The current thesis focused on multisensory integration during rapid upper-limb

movements, specifically at peak limb velocity (PLV). Participants were required to “fling” their

limb through a virtual target while aligning PLV with the target. In Experiment 1, participants

were provided with augmented: auditory, visual, audiovisual, or no feedback at peak limb

velocity. In Experiment 2, participants completed the same flinging task. At PLV, participants

were exposed to either two: auditory beeps, visual flashes, or audiovisual events. Participants

then indicated which sensory cue was presented first (i.e., left or right of the target) in a within

modality temporal order judgment task. This judgment was also made following stimuli

presentation at rest. Across both experiments, the results failed to reveal an audiovisual (i.e.,

bimodal) advantage over unisensory stimuli presentation (i.e., auditory or visual cues alone),

suggesting that the optimal integration of audiovisual information may not occur at peak limb

velocity.

Page 3: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

iii

Acknowledgments

Without the support of my parents Rosy and Tony this work would not have been

possible. I am grateful everyday for their love and support. In addition, Dr. Luc Tremblay has

provided invaluable guidance throughout this process. Thank you Luc for your patience, hard

work, and suggestions over the last two years. The growth I have experience both personally and

academically under your supervision has not gone unappreciated. To my committee members Dr.

Tim Welsh and Dr. Matthias Niemeier, your guidance and suggestions during the early phases of

this project were greatly appreciated. To my fellow graduate students in the PMB and AA labs,

thank you for the countless hours of conversation, good times, and support. Thank you to

everyone at Positano Restaurant. Lastly, I’d like to say a very special thank you to Shanna. Time

will pass and it may go fast, but we’ll still be together.

Page 4: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

iv

Table of Contents

Table of Contents ........................................................................................................................... iv

List of Figures ................................................................................................................................ vi

Chapter 1 – Multisensory Integration and Goal-directed Action .................................................... 1

Introduction .............................................................................................................................. 1 1

Overview of the Current Thesis ....................................................................................................... 1 1.1

Chapter 2 – Literature Review .................................................................................................. 3 2

Processing Sensory Information ...................................................................................................... 3 2.1

Multisensory Processing .................................................................................................................. 5 2.2

2.2.1 Multisensory Processing at the Cortical Level ......................................................................... 6 2.2.2 Multisensory Integration and Perception .................................................................................. 8 2.2.3 Multisensory Integration During Action .................................................................................. 9

2.2.3.1 Sensory Gating During Goal-directed Action .................................................................................. 9 Experimental Aims and Rationale ................................................................................................. 11 2.3

Chapter 3 – Common Methods ............................................................................................... 12 3

Participants ..................................................................................................................................... 12 3.1

Apparatus ....................................................................................................................................... 12 3.2

General Procedure .......................................................................................................................... 14 3.3

3.3.1 Data Analysis Experiment 1 ................................................................................................... 15 3.3.2 Experiment 2 Data Analysis ................................................................................................... 15

Chapter 4 – Can you hear that peak? Utilization of auditory and visual feedback at peak limb 4

velocity .......................................................................................................................................... 17

4.1.1 Abstract ................................................................................................................................... 18 4.1.2 Introduction ............................................................................................................................ 19 4.1.3 Methods .................................................................................................................................. 23 4.1.4 Participants ............................................................................................................................. 23 4.1.5 Apparatus ................................................................................................................................ 24 4.1.6 Procedure ................................................................................................................................ 25 4.1.7 Data Analysis .......................................................................................................................... 27 4.1.8 Results .................................................................................................................................... 28 4.1.9 Discussion ............................................................................................................................... 31

Page 5: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

v

4.1.10 What Does This Article Add? .............................................................................................. 34 4.1.11 References ............................................................................................................................ 35 Multisensory Integration at Peak Limb Velocity in a Within-Modality Temporal Order Judgment 4.2

Task.. ....................................................................................................................................................... 39 4.2.1 Abstract ................................................................................................................................... 40 4.2.2 Introduction ............................................................................................................................ 41 4.2.3 Methods .................................................................................................................................. 46 4.2.4 Participants ............................................................................................................................. 46 4.2.5 Apparatus ................................................................................................................................ 46 4.2.6 Procedure ................................................................................................................................ 48 4.2.7 Results .................................................................................................................................... 50 4.2.8 Discussion ............................................................................................................................... 53

Chapter 5 – General Discussion ............................................................................................. 61 5

Experiment 1 Summary ................................................................................................................. 61 5.1

Experiment 2 Summary ................................................................................................................. 61 5.2

General Discussion Overview ....................................................................................................... 62 5.3

5.3.1 Critical Temporal Window For Visual Processing ................................................................ 63 5.3.2 Multisensory Integration At Peak Limb Velocity .................................................................. 65 5.3.3 The Advantage of Auditory Feedback ................................................................................... 68 5.3.4 Limitations of the Current Thesis ........................................................................................... 69 5.3.5 Concluding Remarks .............................................................................................................. 70

Chapter 6 – References ........................................................................................................... 71 6

Appendices ............................................................................................................................. 77 7

Appendix A – Block Analysis of Experiment 1 ............................................................................ 78 7.1

Appendix B – Block Analysis of Experiment 2 ............................................................................ 80 7.2

Page 6: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

vi

List of Figures

Figure 1. A depiction of the experimental apparatus used in the feedback and TOJ experiments

with the virtual target shown.........................................................................................................12

Figure 2. An example of the flinging movement performed in both experiments........................14

Figure 3. Participants completed the experiment while their heads rested in a chin rest and a

virtual target was displayed in front of them.................................................................................23

Figure 4. An example of the flinging movement performed by participants. The velocity profiles

correspond to the approximate limb velocities illustrated above………………………………..25  

Figure 5. Variability in resultant Displacement. Also shown is the MLE predicted variability in

the audiovisual condition………………………………………………………………………...31  

Figure 6. A depiction of the experimental apparatus.....................................................................46  

Figure 7. An illustration of the flinging movement performed in the experiment………………48  

Figure 8. Number of correct responses as a function of sensory condition……………………..53  

Page 7: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

1

Chapter 1

Multisensory Integration and Goal-directed Action

Introduction 1

We inhabit a dynamic world, rich in sensory information. Considering that we often

actively engage with multisensory stimuli through goal-directed action, it is critical to understand

how the central nervous system processes and utilizes this information. In the context of goal-

directed action, the major empirical focus to date has largely concerned how sensory information

is utilized at the critical moment in the trajectory, specifically movement end (Woodworth,

1899). However, for various motor skills such as overarm throwing, tossing, and putting, the

critical moment in the trajectory is not movement end, but rather at or near peak limb velocity.

How is sensory information used for accurate motor execution when peak limb velocity is the

critical kinematic moment? How are multisensory cues integrated at this kinematic marker?

Further, can augmented sensory feedback be employed to facilitate the spatial and temporal

occurrence of peak limb velocity? In line with these questions, the current thesis sought to

elucidate some of the underlying mechanisms governing sensory utilization, integration, and

localization when the limb is traveling at peak velocity.

Overview of the Current Thesis 1.1

The overarching goal of the current thesis is to understand how sensory information from

multiple modalities is integrated and utilized during goal-directed actions where peak limb

velocity is critical in determining the outcome. What follows is a review of literature pertaining

Page 8: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

2

to processing sensory information, how cues from multiple modalities are combined and

integrated on the behavioural as well as neurological level, and a brief overview of sensory

gating during goal-directed action. Emphasis will be placed on how sensory information is used

throughout an ongoing movement to facilitate accurate execution of a task.

Following this literature review, two experiments are reported. In the first experiment

(i.e., feedback experiment), participants were required to fling their limb through a virtual target

with the goal of aligning their peak limb velocity with the intersection of a virtual target. At peak

limb velocity (i.e., as measured by the velocity of the index finger), participants were presented

with auditory, visual, or audiovisual feedback from a piezo-LED device affixed to the right index

finger. The results of this experiment showed that auditory feedback significantly reduced the

variability in peak limb velocity occurrence when compared to movements performed with no

feedback. When analyzing performance in the visual and audiovisual conditions in contrast, no

significant reduction in constant error or variable error was found. This pattern of results was

interpreted as evidence that participants failed to reweight sensory information in statistically

optimal fashion.

In the second experiment (i.e., TOJ experiment) participants performed the same flinging

movement with the goal of reaching peak limb velocity (i.e., as measured by the velocity of the

wrist) when the index finger intersected the virtual target. At peak limb velocity, participants

were presented with two auditory, visual, or audiovisual cues via two piezo-LED devices

positioned on either side of the virtual target. Following the trial, participants reported which side

of the virtual target the first sensory cue was presented, in a temporal order judgment (TOJ).

Participants also completed the TOJ following stimuli presentation at rest. The results indicated

that participants were more accurate in the TOJ task when viewing the stimuli at rest compared

Page 9: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

3

to presentation at peak limb velocity. When TOJs were completed at rest, participants were more

accurate in the auditory, and audiovisual conditions relative to the visual condition. Critically,

when comparing between TOJs made at rest or following presentation at peak limb velocity, it

was found that participant’s accuracy significantly decreased only in the audiovisual condition.

Overall, the results of this experiment suggest that at peak limb velocity, visual information is

not particularly salient. In addition, the results of this experiment further suggest that the optimal

integration of audiovisual information may not occur when the limb is travelling at peak

velocity.

Chapter 2 – Literature Review 2

Processing Sensory Information 2.1

Precision in aiming movements is facilitated when afferent information is available.

During instances of rapid goal-directed action, the window for processing incoming sensory

feedback is considerably narrow (i.e., Elliott, Carson, Goodman, & Chua, 1991; Elliott, Hansen,

& Khan, 2010). Because visual information is rapidly processed (Lesevre, 1982), it provides

critical feedback as a movement unfolds. Woodworth conducted one of the first studies on the

topic of visual processing during action in 1899. In one experiment, participants were required to

trace a pencil along a rotating drum at various speeds with the lights in the room illuminated or

extinguished. Based on the participant’s performance in this task, Woodworth (1899) concluded

that visual information could be processed and used to inform movement corrections within 450

ms. However, various experiments conducted since have shown that Woodworth’s reciprocal

aiming task induced an overestimation of visual processing times.

Page 10: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

4

Keele and Posner (1968) had participants perform a discrete movement task in which a

stylus was moved from a home position to a target under a specific movement time bandwidth.

By measuring whether participants were hitting the target zone with the lights in the room

illuminated or extinguished, Keele and Posner (1968) proposed the minimum duration for

processing visual feedback was between 190-260 ms. However, Zelaznik, Hawkins, and

Kisselburgh (1983) reduced the required movement time bandwidth, and when analyzing

endpoint variability, found that visual feedback could be utilized in as little as 100-250 ms. In

light of this, the generally accepted visual processing delay has been estimated at 100 ms

(Carlton, 1992). In addition to vision, afferent sensory feedback from the remaining modalities

can also provide valuable information relating to an ongoing movement.

Because of the rapid processing of auditory information within the central nervous

system (Celesia, 1976; Warren, Wise & Warren, 2005), this information can be used with

temporal and spatial precision. In rhesus monkeys for example, behavioural responses to

auditory stimuli have been reported to be as short as 60 ms for a saccade to an auditory target

(Russo & Bruce, 1994). In addition, reaching movements performed to auditory targets have

been shown to have endpoint errors of less than 0.5 cm (e.g., Levy-Tzedek, Hanassy, Abboud,

Maidenbaum, & Amedi, 2012). Overall, when the critical moment in the trajectory is movement

end, both visual and auditory information can be utilized during movement execution. However,

what remains to be determined is how the use of visual and auditory information changes when

the critical moment is not movement end, but rather peak limb velocity.

Besides sensory processing specifically during action, another approach often utilized

when examining how sensory information is processed is the temporal order judgment task

(TOJ). In this task, participants are presented with auditory and visual cues. These cues are not

Page 11: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

5

always presented simultaneously; rather their presentation is staggered according to

predetermined stimulus onset asynchronies (SOA). The participant’s task is to indicate which

sensory cue they perceived to have been presented first (e.g., Hirsh and Sherrick 1961; Kanabus,

Szelag, Rojek, Pöppel, 2002; Spence, Shore, & Klein, 2001). A general finding from this

literature is that for auditory and visual information to be perceived as simultaneous, the visual

stimulus must lead the auditory stimulus by ~ 75-100 ms (e.g., Zampini, Shore, & Spence, 2003;

Jaekl & Harris, 2007). It is critical to note however that many of the TOJ studies previously

conducted presented auditory and visual cues from different spatial locations while the

participant remained at rest (e.g., Bushara, Grafman, & Hallett, 2001; Jaskowski, Jaroszyk, &

Hojan-Jezierska, 1990). As such, the processing of spatially compatible visual and auditory

information in the TOJ paradigm remains to be determined.

Although the above literature has focused on the processing and utilization of unimodal

information, it is also important to consider that sensory cues rarely exist independently of one

another. At any given time, multiple sources of sensory information are available to an individual

and can be used to guide action. Specifically, the central nervous system (CNS) is to able to

formulate a stable percept of the environment as well as plan movements by concurrently

utilizing sensory cues from the vestibular, visual, proprioceptive, and auditory modalities (for

reviews see van Atteveldt, Murray, Thut, & Schroeder, 2014; Driver & Noesselt, 2008).

Multisensory Processing 2.2

When sensory cues from multiple modalities are presented concurrently, the central

nervous system is typically able to formulate a more robust estimate of a sensory event through

multisensory integration. The Maximum Likelihood Estimation Model (MLE) put forward by

Page 12: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

6

Ernst and Bülthoff (2004) provides a theoretical framework for how the CNS accomplishes this

task. In this model, the CNS integrates sensory information across multiple modalities in a

statistically optimal way. Depending on the context of a particular event, the quality of the

sensory information varies. For example, when reaching for a glass of water under very low

luminance conditions, the visual cue can be less reliable than the proprioceptive cue associated

with limb position. As such, proprioception would be more heavily relied upon given its

reliability relative to vision (e.g., van Beers, Sittig, & Denier van der Gon, 1999; van Beers,

Wolpert, & Haggard, 2002). However, if the lights in the room were fully illuminated, a reliance

on vision would be expected. This variability in stimulus reliability affects the weights given to

each sensory modality during the integration process, with lesser weight assigned to unreliable

cues. By taking a weighted average during the sensory integration process, sensory information

from differing modalities is combined such that ambiguity is reduced, and a more stable percept

of a sensory event is formed. In addition, numerous experiments focused on the neuronal basis of

sensory processing have shown that the CNS is largely organized for multisensory integration.

2.2.1 Multisensory Processing at the Cortical Level

Early physiological work as well as more recent neuroimaging data has shed light on the

implicit nature of multisensory processing. Seminal work conducted by Meredith, Nemitz, and

Stein (1987) as well as Stein, Meredith, Hunneycutt, and McDade (1989) showed visual and

auditory information converging on single neurons within the superior colliculus of cats. Stein et

al. (1989) observed that multisensory stimuli yielded a greater magnitude of firing (i.e., a

superadditive effect) among this population of neurons relative to unisensory stimuli, as well as a

greater number of neural responses overall. Critically, it was observed that this effect is time

sensitive, with the superadditive effect only occurring if auditory and visual information were

Page 13: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

7

presented within a temporal window, spanning approximately 150 ms (Meredith, Nemitz, &

Stein, 1987). This convergence of multisensory information in the cortex has also been supported

in experiments involving humans.

Audiovisual convergence and multisensory neurons have been observed throughout the

cortex as well as in areas once considered to be unisensory. Bremmer et al. (2001) exposed

participants to moving auditory, visual, and tactile stimuli while measuring neural activity using

functional magnetic resonance imaging (fMRI). Of particular interest in this experiment was

neural activity within the motion sensitive areas of the cortex (i.e., the intraparietal sulcus and

ventral premotor cortex). The authors report the results of a conjunction analysis that revealed

significant activation common to all three modalities within the posterior parietal cortex (PPC),

right ventral premotor cortex, and lateral inferior postcentral cortex.

In addition, multisensory information has been shown to influence processing within the

primary sensory areas. For example, Calvert et al. (1999) exposed participants to bimodal (i.e.,

audiovisual) and unimodal (i.e., auditory or visual) stimuli in an fMRI study. Under bimodal

stimulus presentation, Calvert et al. (1999) reported significantly greater activity in the primary

auditory (i.e., A1) and visual cortices (i.e., V1) when compared to unimodal presentation. This

finding led the authors to suggest that the primary sensory areas may also process information in

the non-preferred modality. Further work also supports activation within A1 and V1 from the

non-preferred modality. In a study conducted by Lakatos et al. (2009), macaques performed an

intermodal selective attention task where they were required to attend to either auditory of visual

cues on alternating trials. While performing this task, recordings were taken from groups of

neurons in A1 and V1. When analyzing phase coherence across trials, as measured by intertrial

coherence, it was found that neuronal activity within A1 and V1 was influenced specifically

Page 14: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

8

when the monkey attended the non-preferred modality. As such, it appears that the primary

sensory cortices may be multisensory in nature (for reviews see Ghazanfar, & Schroeder, 2006;

van Atteveldt et al., 2014).

Overall, the neuronal evidence suggests that multisensory cues are inherently processed

as such, and that the primary sensory areas are organized to process multisensory information.

How does the robust representation of multisensory stimuli in the cortex manifest in the realm of

perception and action? Multisensory perception and the integration of sensory cues during

voluntary action will be discussed in the following sections.

2.2.2 Multisensory Integration and Perception

As discussed above, the superadditive effect of multisensory input is strongly represented

at the neuronal level. Perceptually, multisensory stimulus presentation can at times be

facilitative, whereas in other instances bias perception creating multisensory illusions. One

example of facilitation is when compared to unisensory stimulus presentation, multisensory

information can yielded shorter reaction times (e.g., Hershenson, 1962). Beyond this facilitative

effect, the presence of multiple sensory modalities can influence the perception of a sensory

event as evidenced by the McGurk effect (McGurk & MacDonald, 1976). Additionally, the

ventriloquist effect (e.g., Alais & Burr, 2004) is a well-known perceptual illusion that occurs

independent of laboratory manipulations. This effect can be characterized by the example of

watching television. In this instance, the voice of an actor on screen appears to emanate

seamlessly from the actor’s mouth, despite the fact that the auditory information stems from the

television speakers. Alais and Burr (2004) examined the spatial nature of this effect by

investigating the localizability of auditory and visual stimuli. The authors report that the illusion

arises from near optimal sensory integration, with the more reliable sensory input (i.e., vision in

Page 15: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

9

this case) biasing the spatial position of the lesser reliable sensory cue (i.e., audition).

Interestingly, when the spatial localizability of the visual cue was degraded, the effect is reversed

such that audition biases the spatial position of vision. As such, sensory information is optimally

integrated when an individual observes an audiovisual event.

2.2.3 Multisensory Integration During Action

It is important to note that the studies discussed above most often consider the

integration and utilization of multisensory information when the participant is a passive observer,

and the motor responses are only made after stimulus presentation. The results of the

experiments discussed in the previous section strongly suggest that at rest, optimal integration

across sensory modalities occurs (e.g., Alais & Burr, 2004; Ernst & Banks, 2002; Ernst &

Bültoff, 2004). It is important to note here that humans hardly remain stationary in their

multisensory environments. Interestingly, studies examining multisensory integration during

action have revealed that the integration of sensory information across multiple modalities differs

during goal-directed action. The discussion below pertains to instances of multisensory

integration during goal-directed action.

2.2.3.1 Sensory Gating During Goal-directed Action

The perception of tactile stimulation can be gated as function of movement. Chapman,

Bushnell, Miron, Duncan, and Lind (1987) noted a reduction of tactile sensitivity as a function of

passive and active limb movements. Further work conducted by Juravle, Deubel, Tan, and

Spence (2010) sought to determine the changes in tactile sensitivity throughout the time course

of a goal-directed movement. Of particular interest are the results from Experiment 1 in which

participants performed reaching movements between two effectors (i.e., computer mice), with

Page 16: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

10

tactile stimulation being applied to the moving hand prior to and during the movement.

Specifically, discrete tactile stimulation was applied to the hand at: movement preparation,

initiation (i.e. immediately following the ‘go’ signal), movement execution, and finally after

movement completion. Juravle et al. (2010) reported decreased tactile sensitivity during the

movement with the largest reduction occurring when stimulation had been applied during

movement preparation and execution. Thus, the data strongly suggests that tactile gating is

modulated as a function of movement phase, with the strongest suppression occurring following

movement initiation. When taken as whole, the results of Chapman et al. (1987) and Juravle et

al. (2010) suggest that sensory processing at rest fundamentally differs from that during

movement preparation and execution.

More pertinent to the current thesis is the sensory gating that occurs for audiovisual

processing during goal-directed action. In an experiment conducted by Tremblay and Nguyen

(2010), participants performed a 30 cm reaching movement. Under the terminal point (i.e.,

target) of this movement, participants were exposed to the fission/ fusion illusion (Shams,

Kamitani, & Shimojo, 2000; Shams, Kamitani, & Shimojo, 2002). In this illusion, the presence

of auditory beeps biases the number of perceived flashes. For example, when two flashes are

present and one beep is heard, these events are fused such that the participant will erroneously

report seeing only one flash. However, Tremblay and Nguyen (2010) reported that participants

became less susceptible to the fusion illusion at the highest velocity stages of their upper-limb

movement. Tremblay and Nguyen (2010) suggested that this modulation in multisensory

integration might be a reflection of the task-relevant optimal integration of multisensory cues, or

sensory gating of auditory information.

Page 17: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

11

Experimental Aims and Rationale 2.3

The empirical focus throughout the current literature has been on sensory processing

during movements where the critical moment in the trajectory is movement end (Keele &

Posner, 1968; Carlton, 1981; Ernst & Bülthoff, 2004). Unfortunately, this focus has neglected

movements where peak limb velocity is the critical moment in the trajectory such as overarm

throwing. As such, how sensory information is processed at this specific kinematic marker has

remained largely understudied. This approach is certainly warranted in order to fully understand

how sensory information is processed across the entire spectrum of motor skills. As such, the

current thesis sought to understand how sensory information was utilized with emphasis on

movements where peak limb velocity is the critical kinematic moment, as opposed to movement

end. To this end, participants in Experiments 1 and 2 were tasked with aligning the moment peak

limb velocity was reached with the intersection of a virtual target. Sensory processing and

utilization was assessed through augmented sensory feedback as well as within a within-modality

temporal order judgment.

Page 18: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

12

Chapter 3 – Common Methods 3

Participants 3.1

Participants in Experiment 1 (N = 13) and Experiment 2 (N = 13) were recruited from the

graduate and undergraduate population at the University of Toronto. Inclusion criterion was

being right-hand dominant with normal or corrected to normal visual acuity. Participants were

financially compensated for participation (i.e., $10).

Apparatus 3.2

Participant’s heads rested on a chin rest (38 cm tall) fastened to a 61 cm by 75 cm table

for the duration of the experiments (see Figure 1). A custom built wooden frame (48 cm tall)

with a reflective surface 12 cm in diameter was positioned 13 cm away from the home position.

Attached to the upper most portion of the chin

rest was a yellow LED light (1 cm in diameter),

which served as the target for the experiment.

The participant viewed the target via the

reflective surface. As such, the target appeared

elevated from the surface of the table (i.e., a

virtual target). Each trial began with the

participant’s finger on a 1.5 cm by 1.5 cm piece

of Velcro™, which served as the home position

for the experiments. Participants were cued to

begin their movement when the target became visible via the reflection. Participant’s limb

Figure 1. A depiction of the experimental apparatus used

in the feedback and TOJ experiments with the approximate

position of the virtual target shown. Note: This figure is

not to scale.

Page 19: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

13

movements were measured along the Y-axis (i.e., primary movement axis), as and the X and Z-

axes (i.e., secondary movement axes). These movement axes are plotted in Figure 1.

The participant’s movements were recorded throughout the experiments using an

Optotrak Certus (Northern Digital Inc., Waterloo, ON, Canada) motion capture system. An

infrared emitting diode (IRED) was used to measure the participant’s movement. The Optotrak

real-time sampling rate was set at 500 Hz and controlled using a custom Matlab script (The

Mathworks, Inc., Natwick, MA, USA) to trigger real-time stimuli presentations through an

analog-to-digital board (PCI-6024E, National Instruments Corp., Austin, TX, USA).

In both experiments, instantaneous velocity was calculated using the position and time

data collected from the Optotrak system while recording from an IRED positioned on the index

finger (e.g., the feedback experiment) and on the wrist (i.e., TOJ experiment). Movement start

location was defined when the limb first exceeded a velocity of 0.03 m/s. Peak velocity (i.e., as

measured by the unique IRED position in each experiment) was selected as two consecutive

samples with a velocity decrease, following a minimum velocity of 0.8 m/s. Lastly, the resultant

displacement between movement start and peak limb velocity occurrence was calculated and

analyzed. In order to ensure that peak limb velocity occurred at the same moment across

experiment (i.e., differing IRED locations), a pilot study was carried out with three participants

who did not participate in either the feedback or the TOJ experiment. In this pilot study,

participants performed the flinging movement while limb velocity was measured via an IRED

positioned on the index finger (i.e., experiment 1) as well as the wrist (i.e., experiment 2). Using

a Pearson correlation, the sample at which peak limb velocity occurred for both IREDs was

compared. This analysis revealed an average correlation of .99 between the samples in which

peak limb velocity occurred between the two IREDs. As such, although limb velocity was

Page 20: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

14

measured at different locations on the moving limb, peak velocity occurrence was consistent

across both experiments.

General Procedure 3.3

Before the experimental trials, the participant estimated the target and home position. The

home position was estimated by having the participant rest their index finger on the home

position for a trial. When estimating the target position, the participant placed their finger at the

position of the virtual target. Participants were instructed to “fling” their index finger through the

target while moving as quickly and accurately as possible during the experimental trials (see

Figure 2). As such, the participant aimed to reach their peak limb velocity while passing through

the centre of the virtual target. After, participants returned to the home position and awaited the

signal for the next trial.

Figure 2. An example of the flinging movement performed in both experiments. Participants were instructed to align PLV

occurrence with the intersection of the target while propelling their limb to its terminal position.

Page 21: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

15

3.3.1 Data Analysis Experiment 1

Performance was assessed across feedback conditions by analyzing: movement time (i.e.,

MT), average peak limb velocity, constant error (CE), variable error (VE), and resultant

displacement (i.e., R). Constant error was defined as the sum of the differences between the

position of peak limb velocity occurrence and target position divided by the number of trials.

Variable error was calculated as the square root of the sum of the squared differences between

the position of peak limb velocity occurrence and mean peak limb velocity location divided by

the number of trials. Resultant displacement refers to the distance from the home position

travelled when peak limb velocity was reached.

To determine the effects of feedback on performance accuracy, separate one-way

repeated measures ANOVA’s were conducted for constant and variable error, as well as resultant

displacement for all movement axes. Significant main effects were followed up with Tukey HSD

post hoc comparisons. Violations of spherecity were corrected using Hyunh-Feldt corrections

and the degrees of freedom reported to two decimal places.

3.3.2 Experiment 2 Data Analysis

As in experiment 1, constant error, variable error, and resultant displacement were

calculated as dependent measures. Although participants completed their entire flinging

movement, for the purposes of constant error and variable error calculations, peak limb velocity

position was utilized. In addition, the main dependent variable in this experiment was the

accurate reporting of which sensory cue was presented first. To determine how response

accuracy varies as a function of movement condition and sensory modality, a 2 Presentation (at

rest, at peak limb velocity) by 3 Sensory Cue (Auditory, Visual, Audiovisual) repeated measures

Page 22: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

16

ANOVA was used. Tukey’s HSD post hoc comparisons were conducted on any main effect

associated sensory cue and stimulus presentation. Deviations from sphericity were corrected

using Hyunh-Feldt corrections, with the degrees of freedom reported to two decimal places.

Page 23: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

17

Chapter 4 – Can you hear that peak? Utilization of 4auditory and visual feedback at peak limb velocity

Page 24: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

18

4.1.1 Abstract

Purpose: At rest, the central nervous system combines and integrates multisensory cues to yield

an optimal percept. When engaging in action, the relative weighing of sensory modalities has

been shown to be altered. Because the timing of peak velocity is the critical moment in some

goal-directed movements (e.g., overarm throwing), the current study sought to test whether

visual and auditory cues are optimally integrated at that specific kinematic marker when it is the

critical part of the trajectory. Methods: Participants performed an upper-limb movement in

which they were required to reach their peak limb velocity when the right index finger

intersected a virtual target (i.e., a flinging movement). Brief auditory, visual, or audiovisual

feedback (i.e., 20 ms in duration) was provided to participants at peak limb velocity.

Performance was mainly assessed by analyzing the spatial bias and variability of the limb’s

position when it reached peak velocity. Results: Relative to when no feedback was provided,

auditory feedback significantly reduced the spatial variability of the finger position at peak limb

velocity. However, no such reductions were found for the visual feedback condition. Further,

providing both auditory and visual cues concurrently also failed to yield the predicted

improvements in endpoint variability. Conclusions: Overall, the central nervous system can

make significant use of an auditory cue, but may not optimally integrate a visual and auditory

cue at peak limb velocity, when peak velocity is the critical part of the trajectory.

Page 25: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

19

Can you hear that peak?

Utilization of auditory and visual feedback at peak limb velocity

4.1.2 Introduction

Our environment is filled with multisensory stimuli. Incoming sensory information from

the various modalities creates a cohesive perception of our environment and, as such, is critically

relied upon for activities of daily living. Within the context of goal-directed action, studies have

primarily focused on skills where movement end was the critical moment of the trajectory (i.e.,

reach-to-grasp: e.g., Jeannerod, 1984). However, there are voluntary motor skills, such as

throwing a baseball or hitting a golf ball, for which the critical moment is ideally peak limb

velocity (PLV). In research contexts, these movements include tossing (e.g., Fleischauer &

Sherwood, 2008), flicking/flinging (e.g., Dulberg, Amant, & Zettlemoyer, 1999), and punching

(e.g., Cavanagh & Landa, 1976). Such movements can provide novel perspectives on the use of

sensory feedback during action because the predominantly utilized sensory modality (i.e., vision)

is not particularly salient at the critical kinematic marker (i.e., PLV). Indeed, the percept of the

limb when it reaches peak limb velocity is a blur, at best. As such, how can sensory feedback

from the non-visual modalities (e.g., audition) be utilized when the limb is travelling at high

velocities? Prior to addressing this question, selected studies were reviewed to ascertain current

knowledge on the use of vision, audition, and audio-visual information, specifically for

movements where the limb must stop onto a target (e.g., discrete reaches/ reaching).

During rapid upper-limb reaches, visual information gathered during the movement can be

used to perform online trajectory amendments (e.g., Keele & Posner, 1968). For example,

Proteau, Roujoula and Messier (2009) had participants perform a video-aiming task in which

vision of a cursor representing the participant’s limb was manipulated. Proteau et al. (2009)

Page 26: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

20

reported that participants were able to initiate a correction soon after the cursor jumped, leading

them to suggest that visual information is monitored throughout discrete reaching movements

(see also Saunders & Knill, 2003). However, in these studies, the cursor jumps were

implemented early in the trajectory (i.e., soon after movement initiation). To test when visual

feedback typically contributes to online control during discrete reaches, Kennedy, Bhattacharjee,

Hansen, Reid, and Tremblay (2015) asked participants to perform 30 cm movements while

vision was provided during different windows along real-time limb velocity profiles.

Specifically, vision was provided early (i.e., limb velocity of 0.8-1.4 m/s, before PLV), in the

middle (i.e., limb velocity above 1.4 m/s, before and after PLV), or late (i.e., limb velocity of

1.4-0.8 m/s, after PLV) in the movement. Kennedy et al. (2015) reported that endpoint

consistency was comparable to full vision in the early and middle visual windows. However, the

middle visual feedback condition led to longer limb deceleration phase durations compared to

full vision. As such, the early window alone, which lasted only 43 ms, led to endpoint precision

and limb deceleration duration comparable to full vision. Thus, the spatiotemporal window for

processing online visual feedback could be considerably narrow. As a result, the gathering of

visual feedback can vary significantly during voluntary reaching movements and seems to

predominantly take place quite early in the trajectory.

Sensory information from other modalities (e.g., audition) can also provide valuable spatial

feedback during reaching movements. In the experimental trials of a study conducted by Levy-

Tzedek, Hanassy, Abboud, Maidenbaum, and Amedi (2012) sighted participants performed

upper-limb reaching movements while blindfolded and received auditory feedback via a sensory

substitution device (SSD). When provided with auditory feedback only, participant’s movements

did not differ significantly in terms of movement time, peak velocity, or amplitude as compared

to reaches performed with visual feedback. Although participants were more accurate when

Page 27: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

21

visual information was available, reaches performed using the SSD had endpoint errors of less

than 0.5 cm on average (Levy-Tzedek et al. 2012). Overall, auditory feedback can also be used to

complete upper-limb reaches with excellent levels of precision.

Visual and auditory feedback can also be combined and integrated to yield an optimal

percept of spatial position. The Maximum Likelihood Estimation Model (MLE) forwarded by

Ernst and Bülthoff (2004) has provided a theoretical framework for how the central nervous

system accomplishes this task. According to the MLE, the central nervous system combines

incoming sensory information across multiple modalities in a statistically optimal fashion (e.g.,

Alais & Burr, 2004). Stimulus reliability affects the “weight” given to each sensory modality

during the integration process, with lesser weight assigned to unreliable cues (e.g., Ernst &

Banks, 2002). By taking a weighted average during the sensory integration process, sensory

estimates from different modalities are combined such that ambiguity is reduced, and a more

stable percept of a sensory event is formed. A well-cited example of multisensory integration is

spatial ventriloquism. This illusion can be characterized by watching television in which the

spatial location of auditory signals is biased towards the spatial location of the visual signal

(Alais and Burr, 2004). In fact, multisensory processing is so robust at the perceptual,

behavioural, and neuronal level (see Driver & Noesselt, 2008 for a review) that some researchers

hypothesize that the brain is largely organized for multisensory integration (e.g., Ghazanfar, &

Schroeder, 2006).

However, it is important to note here that support for the MLE and other Bayesian models

of sensory utilization have primarily stemmed from tasks requiring little or no movement (see

Witten & Knudsen, 2005 for a review; cf. Körding & Wolpert, 2006). Considering that we

hardly remain stationary in daily life (or at least we should not), it is important to consider how

sensory information from multiple modalities is combined and integrated during goal-directed

Page 28: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

22

action. Interestingly, studies that have examined multisensory integration have revealed

considerable differences between multisensory integration at rest and multisensory integration

during action.

It has been reported that audiovisual processing is modulated during fast and accurate

reaching movements. In an experiment conducted by Tremblay and Nguyen (2010), participants

performed a 30 cm movement to a 0.5 cm target. Beneath the terminal point of this movement,

participants were presented with an audiovisual fission/ fusion illusion (see Shams, Kamitani, &

Shimojo, 2000). In this audiovisual illusion, the presence of auditory beeps has been reported to

bias the number of perceived visual flashes. For example, when two flashes and one beep were

presented, these events were fused such that participants often erroneously report seeing only one

flash (i.e., fusion illusion). Tremblay and Nguyen (2010) reported that participants became less

susceptible to the fusion illusion during goal-directed action. Critically, it was found the degree

to which participants fused the two visual stimuli was minimized at the highest limb velocity

stage of the movement. When taken as a whole, limb velocity can be a useful proxy to test

multisensory combination and integration (see Tremblay & de Grosbois, 2015), especially

because engaging in voluntary action significantly alters the integration of multisensory cues (see

also Juravle, Deubel, Tan, & Spence, 2010).

The experiments cited above (e.g., Proteau et al., 2009; Kennedy et al., 2015) employed

tasks where the participant’s goal was to terminate their movement onto a target (i.e., upper-limb

reaches). In such instances, visual feedback gathered early in the trajectory was utilized to bring

the limb to a halt, on its final position. However, the critical position of voluntary movements

can be during the trajectory, which could alter when sensory information is primarily gathered.

For example, to properly execute an overarm throw or golf swing, release or contact with the ball

at peak velocity is absolutely critical (e.g., Jegede, Watts, Stitt, & Hore, 2005). To the best of our

Page 29: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

23

knowledge, however, no experiments have been conducted on the use of feedback presented at

peak limb velocity in order to control the limb position at peak limb velocity (i.e., PLV being the

critical portion of the trajectory).

The aim of the current study was to test whether the utilization and integration of

multisensory cues was optimal when auditory, visual, or audiovisual feedback were presented at

peak limb velocity, and that is, when peak limb velocity was the critical part of the trajectory

(i.e., a flinging movement). Although providing a visual cue normally facilitates performance,

this improvement may be modest with a flinging task because visual information about the limb

at peak limb velocity less certain than at rest (re.: blurred image of the limb). In addition, because

of audition’s excellent temporal sensitivity (Celesia, 1976), providing auditory cues could also

yield an improvement in performance, especially at the fastest stage of a rapid limb movement.

Finally, based on the presumption that the central nervous system still attempts to integrate

multisensory cues at peak limb velocity (Ernst & Bülthoff, 2004), visual and auditory cues

presented together should yield better endpoint precision than with visual or auditory feedback

alone. Alternatively, if the integration of the cues is suboptimal at peak limb velocity,

performance in one of the unisensory conditions (i.e., vision or audition alone) should yield

better, or at least as good, performance than the combined condition.

4.1.3 Methods

4.1.4 Participants

The University of Toronto Research Ethics Board approved the experimental protocol

reported herein. Thirteen individuals (6 males) with an average age of 23.7 years (SD = 2.2)

provided informed consent prior to participating in the experiment. Participants were recruited

from the graduate and undergraduate student populations at the University of Toronto. All

Page 30: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

24

participants self-reported to be right-handed with normal or corrected-to-normal vision.

Participants were financially compensated $10.

4.1.5 Apparatus

Participant’s heads rested on a chin rest (38 cm tall) fastened to a 61 cm by 75 cm table

(see Figure 3 for a depiction of the experimental apparatus). A custom built wooden frame (48

cm tall) with a reflective surface 12 cm in diameter was positioned 13 cm away from the home

position. Attached to the upper most portion

of the chin rest was a yellow LED light (1

cm in diameter), which served as the target

for the experiment (see Figure 3). The

participant viewed the target via the

reflective surface such that the target

appeared elevated from the surface of the

table (i.e., a virtual target). Each trial began

with the participant’s finger on a 1.5 cm by

1.5 cm piece of Velcro, which served as the

home position for the experiment. The movements of the participant were measured along the Y-

axis (i.e., primary movement axis), as well as the X and Z-axes (i.e., secondary movement axes).

The resultant distance between the home position and observed target location was

approximately 40 cm. Along the X, Y, and Z axes, the target was approximately 15 cm to the

right, 27 cm away and 25 cm above the home position, respectively (see axes in Figure 3).

The position of the participant’s right index finger was monitored throughout the

experiment using an Optotrak Certus (Northern Digital Inc., Waterloo, ON, Canada) motion

Figure 3. Participants completed the experiment while their heads

rested in a chin rest and a virtual target was displayed in front of

them. This schematic is not drawn to scale.

Page 31: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

25

capture system. An infrared light emitting diode (IRED) was attached to a banjo pick and placed

on the tip of the participant’s right index finger. The Optotrak real-time sampling rate was set at

500 Hz and controlled using a custom Matlab script (The Mathworks Inc., Natick, MA, USA).

Likewise, the Matlab script controlled the real-time feedback through an analog output board

(PCI-6024E, National Instruments Corp., Austin, TX, USA). The experiment was conducted in a

dark room to prevent the participants from viewing their limb. Also, participants wore a black

arm sleeve and the lights in the room were illuminated every five minutes to minimize the effects

of retinal and pupil adaptation to darkness (Fedorov & Mkrticheva, 1938; Hecht, 1920).

A custom piezo electric buzzer (2350 Hz, 75 dB) with a green LED light affixed to it was

positioned proximal to the IRED on the participant’s right index finger. This device (i.e., piezo-

LED) was used to provide participants with augmented feedback during the movement (i.e., for

20 ms only). Instantaneous velocity was calculated using the position data collected from two

subsequent samples gathered by the Optotrak system. Movement start location was defined when

the limb first exceeded a velocity of 0.03 m/s. Peak velocity was marked as two samples with a

velocity decrease, after the limb reached a minimum velocity of 0.8 m/s. The IRED position at

peak velocity was used to calculate constant and variable error, as well as movement time. The

resultant distance between movement start and peak limb velocity location was also calculated.

4.1.6 Procedure

Prior to beginning the experimental trials, the target and home position were estimated by

the participant and recorded using the Optotrak. The home position was estimated by having the

participant rest their index finger on the home position for a single trial. In order to prevent

participants from receiving augmented tactile feedback that would yield an improvement to

performance, participants performed their movement towards a virtual target. To estimate the

Page 32: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

26

position of the virtual target, the participant placed their finger where the virtual target was

observed. At this time, the Optotrak recorded the position of the IRED and saved this location as

the target position. In the experimental trials, participants were instructed to “fling” their index

finger through the target as quickly and accurately as possible (see Figure 4). That is, the

participant’s task was to reach their peak limb velocity (i.e., as measured by the velocity of the

index finger) as they passed through the centre of the virtual target (i.e., no tactile feedback

provided). Following completion of the flinging movement, participants returned their finger to

the home position and awaited the signal for the next trial.

Figure 4. The flinging movement performed by participants and the corresponding velocity profiles.

Page 33: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

27

Due to the novel dynamics of the task, participants first completed twenty familiarization

trials, followed by thirty baseline trials, where no feedback was given (i.e., no feedback

condition). The no feedback conditions were designed as a baseline and were thus not

counterbalanced with the experimental feedback conditions. After, participants received auditory

(i.e., piezo beep), visual (i.e., LED flash), or audiovisual feedback for thirty trials each in a

blocked and counter-balanced order. Real-time feedback was provided for 20 ms when and

where peak limb velocity was reached via the custom built piezo-LED apparatus affixed to the

finger. Based on the sampling frequency of 500 Hz, Matlab processing delays, and hardware

transmission delays; the onset of the flash or beep was no more than 10 ms after peak limb

velocity was reached, which was deemed to be in real-time. In the audiovisual condition, the

auditory and visual cues were presented simultaneously. Although previous work has shown that

to be perceived as simultaneous, the visual stimulus must lead the auditory stimulus by roughly

75-100 ms (e.g., Zampini, Shore, & Spence, 2003), the auditory and visual cues were presented

simultaneously in this study. This was done in part because peak limb velocity was detected in

real-time. As well, auditory and visual cues are presented simultaneously in the real world.

4.1.7 Data Analysis

To assess the flinging performance across the feedback conditions, movement time (i.e.

MT), average peak limb velocity, constant error (i.e., CE), variable error (i.e., VE) and resultant

displacement (i.e., DISP) were analyzed. Constant error was defined as the sum of the

differences between peak limb velocity position and the estimated target position divided by the

number of trials. Variable error refers to the square root of the sum of the squared differences

between peak limb velocity occurrence and mean peak limb velocity position divided by the

number of trials. In addition, resultant displacement and movement time were calculated as the

Page 34: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

28

distance from the home position traveled when peak limb velocity was reached. To determine the

effect of feedback condition on these performance measures, separate one-way repeated

measures ANOVAs with four levels (i.e., no feedback, auditory, visual, audiovisual) were

performed for constant and variable error, across all movement axes. Alpha was set at .05 and

Huynh-Feldt corrections were applied when the sphericity assumption was violated. In all

instances of sphericity assumption violations, the degrees of freedom were reported to one

decimal place. Effects sizes are reported using partial eta squared. Tukey’s HSD comparisons

were used for post hoc analyses, when required. Additionally, the predicted weights of variability

in the audiovisual condition were computed using the equation:

𝜎!" =!!!  !  !!

!

!!!!  !!

!

(1)

In this equation (e.g., Ernst & Bülthoff, 2004), the product of the squared variability in the

spatial position of PLV occurrence in the auditory (A) and visual (V) condition is divided by the

sum of squared variability of PLV occurrence in the auditory and visual conditions.

4.1.8 Results

Means and standard deviations for all dependent variables can be found in Table 1. On

average, participants completed the movement in 150 ms, with an average peak velocity of 2.92

m/s. The ANOVAs performed for MT and PLV revealed that movement time did not vary

significantly across feedback conditions, F (3, 36) = 1.5, p = .23, 𝜂!! = .03, nor did peak velocity,

F (3, 36) = .18, p = .91, 𝜂!! = .02. When analyzing constant error in the primary movement axis,

no significant effect of feedback condition was found, F (3, 36) = 2.9, p = .11, 𝜂!! = .14. This

result was also found in the X-axis, F (3, 36) = .21, p = .88, 𝜂!! = .03, and Z-axis F (3, 36) = .43,

Page 35: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

29

p = .73, 𝜂!! = .04, respectively. Lastly, when analyzing constant error for resultant displacement,

no significant effect of feedback condition was found, F (3, 36) = 1.3, p = .30, 𝜂!! = .11.

Table 1

Means and Between-subject Standard Deviations for Movement Time (MT), Average Peak Limb Velocity (PLV),

Constant error (CE) and Variable error (VE) Across the Three Movement Axes (i.e., Y-X-Z), and CE and VE for

Resultant Displacement (R). Values in Bold Indicate a Statistically Significant Difference Between Feedback

Conditions.

MT (ms)

PLV (m/s)

CE-Y (mm)

CE-X (mm)

CE-Z (mm)

CE-R (mm)

VE-Y (mm)

VE-X (mm)

VE-Z (mm)

VE-R (mm)

Feedback Condition

M (SD)

M (SD)

M (SD)

M (SD)

M (SD)

M (SD)

M (SD)

M (SD)

M (SD)

M (SD)

No Feedback

152.4 (24.8)

2.8 (0.7)

-110.3 (73.6)

-3.2 (32.1)

-64.9 (48.5)

54.6 (64.1)

48.4 (20.2)

17.2 (11.3)

34.5 (14.1)

56.2 (24.5)

Auditory

149.9 (19.2)

2.9 (0.6)

-93.5 (49.2)

-3.2 (33.6)

-58.7 (55.5)

66.6 (59.2)

27.7 (11.5)

10.3 (3.8)

20.1 (8.4)

30.8 (14.8)

Visual

159.1 (19.1)

2.9 (0.5)

-90.1 (55.7)

-3.8 (34.3)

-59.9 (40.7)

71.4 (53.2)

37.1 (18.7)

12.1 (8.2)

25.2 (16.3)

43.2 (25.5)

Audiovisual

157.9 (24.2)

2.9 (0.5)

-86.3 (48.3)

-6.7 (30.4)

-57.9 (40.1)

80.6 (45.1)

38.3 (14.4)

14.1 (7.6)

22.1 (7.8)

41.1 (17.2)

In contrast, the analyses of variable error in the primary, secondary, and tertiary axis (i.e.,

Y-axis, X-axis, Z-axis, respectively) yielded main effects of feedback condition: Y-axis, F (3,

36) = 6.5, p = .01, 𝜂!! = .35; X-axis, F (2.4, 29.1) = 3.3, p = .01, 𝜂!! = .21; and Z-axis, F (3, 36) =

4.1, p = .09, 𝜂!! = .27. Tukey’s HSD post hoc contrast thresholds were calculated for the three

axes: X-axis: HSD = 6.83 mm; Y-axis: HSD = 12.89 mm; Z-axis: HSD = 13.05 mm. The post

hoc contrasts revealed that variability of the finger position at peak limb velocity was

significantly lower in the auditory condition than in the no feedback condition across all three

axes: Y-axis: p < .01; X-axis: p = .03; Z-axis: p = .03 (see Table 1). As well, these variable error

Page 36: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

30

values in the visual and audiovisual conditions were not significantly lower than in the no

feedback condition while being not significantly worse than in the auditory condition (i.e., all ps

> .09, see Table 1). Accordingly, these main effects and post hoc contrasts were replicated for

the variable error values of the resultant displacement, F (3, 36) = 6.4, p = .01,  𝜂!! = .35, HSD =

17.5 mm (see Figure 5).

To follow up these variable error findings, the predicted weights of the auditory and visual

cues in the audiovisual condition were computed using Equation 1. In this equation, the observed

variability in peak limb velocity occurrence was calculated for the auditory and visual conditions

and used to predict the variability in peak limb velocity occurrence in the audiovisual condition.

This analysis revealed the mean predicted variability in spatial position of peak limb velocity

occurrence along the Y-axis (i.e., primary axis) was 21.1 mm (SD = 9.3). This mean predicted

variability significantly differed from the mean observed variability t (12) = 4.08, p = .002, Mean

Observed = 38.3 mm (SD = 14.4). This relationship was also found for the secondary axes: X-

axis; Mean Predicted = 7.4, SD = 2.8, Mean Observed = 14.1 (SD = 7.6), t (12) = 3.66, p = .003,

as well as the Z-Axis; Mean Predicted = 14.7 (SD = 7.6), Mean Observed = 22.1 mm (SD = 7.8),

t (12), 2.7, p = .02. Finally, as shown in Figure 5, when analyzing resultant displacement, the

predicted value of 23.7 mm (SD = 12.5 mm) significantly differed from the observed variability

in displacement of 41.1 mm, t (12) = 3.44, p = .005.

Page 37: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

31

Figure 5. Variability in resultant displacement across all experimental conditions as well as MLE predicted

variability in the audiovisual condition.

4.1.9 Discussion

The processing and integration of multisensory stimuli at peak limb velocity to control

limb position at that kinematic marker was the primary focus of the current experiment. To this

end, participants were tasked with flinging their index finger through a virtual target, and

aligning the moment their right index finger reached its peak velocity with the position of the

virtual target (i.e., participants did not make contact with the target thus avoiding terminal

feedback). At peak limb velocity, participants received auditory, visual, or audiovisual feedback.

In addition, trials where no feedback was given were also completed. Based on the MLE (Ernst

& Bülthoff, 2004), it was hypothesized that performance would be better in the audiovisual

condition relative to at least one of the unisensory conditions. Contrary to this, it was found that

only auditory feedback yielded a significant reduction in performance variability as compared to

when no feedback was provided. The results thus indicate that the most reliable sensory cue to

Page 38: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

32

detect the spatial position of the limb at peak velocity can be audition, when peak limb velocity

is the critical part of a fast and accurate voluntary action.

An explanation for the lack of differences between the no feedback and visual feedback

conditions may be related to the utilization of visual feedback. Although many studies have

shown that visual information facilitates the execution of voluntary action (e.g., Zelaznik,

Hawkins, & Kisselburgh, 1983; Keele & Posner, 1968), augmented visual feedback failed to

yield a significant improvement to the flinging task employed in this study. This may be related

to the differences in visual processing between traditional tasks where movement end is the

critical moment for sensory feedback utilization and tasks where peak limb velocity is of greater

importance, as in the current study. As introduced above, feedback utilization before peak limb

velocity for an endpoint control task such as those employed by Proteau et al. (2009) and

Kennedy et al. (2015) may occur very early on in the trajectory. In contrast, for tasks where peak

limb velocity is the critical part of a movement, the use of visual information may be

significantly limited.

An alternate yet not mutually exclusive explanation for the lack of visual feedback

facilitation may be related to the quality of the visual cue. As noted in the Methods section, the

entire experiment was conducted in darkness. Because the limb was traveling at such a high

velocity (i.e., approximately 2.9 m/s) when the visual feedback was presented, the LED created a

streak of light instead of a discrete indication of where peak limb velocity occurred. Although

peak limb velocity was associated with the first point along this streak, the participants may have

been unable to form a stable representation of where peak limb velocity occurred and thus

augmented visual feedback did not reliably improve their performance. Nevertheless, visual

feedback did not yield significant improvements in performance for the flinging task employed.

Page 39: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

33

A consideration regarding the audiovisual condition is related to the simultaneity in the

presentation of auditory and visual cues. One might surmise that because auditory information is

processed faster than visual cues (e.g., Celesia, 1976; Zampini et al., 2003), presenting the

auditory and visual information together would have led to asynchronous integration of the two

modalities. This could explain why the auditory information led to the best performance. If this

were the case, it would have been expected that the position of peak limb velocity occurrence in

both the visual and audiovisual conditions would reflect the approximate 100 ms difference

associated with the temporal window of integrating visual and auditory cues (e.g., Colonius &

Diederich, 2004). That is, when traveling at 2.92 m/s when peak limb velocity was reached, it

would be expected that the position of peak limb velocity occurrence would be biased roughly

0.29 m further than observed in the resultant axis. These differences between the actual values,

(Mean Visual = 71.4 mm, Mean Auditory = 66.6 mm), however are negligible when compared to

the predicted values.

When considering the results of the auditory and visual conditions alone, the MLE (Ernst

& Bülthoff, 2004) would also predict that the results of the audiovisual condition would be at

least as good as the auditory condition. However, the results suggest that in the audiovisual

condition, participants have heavily weighted the visual cue instead of the most reliable sensory

cue (i.e., audition) and thus performance was not facilitated. Support for this claim can be found

in the significant differences found for the predicted performance in the audiovisual condition.

Specifically, the predicted values across the movement axes, and well as resultant displacement

were found to be significantly less than the observed values, suggesting that the weights assigned

to the auditory and visual cues at peak limb velocity was suboptimal. This finding has

implications for the optimal integration of multisensory cues during goal-directed action.

Specifically, when peak limb velocity is the critical moment in the trajectory, optimal

Page 40: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

34

multisensory integration may not necessarily occur. Finally, the experimental task employed here

(i.e., a flinging task) provides a novel approach to studying sensory processing during an

ongoing movement. This particular focus certainly has practical applications for skills such as

throwing a baseball or hitting a golf ball.

4.1.10 What Does This Article Add?

This study adds to the literature in two ways. The flinging task we employed provides a

useful and novel approach to studying sensory processing throughout an ongoing movement.

Although there is predominant focus in the literature on sensory processing during reaches to

terminal targets, it is important to focus on sensory processing at peak limb velocity because of

practical applications in sports contexts, such as baseball and golf. Also, the main findings add to

the growing literature on how the central nervous system integrates sensory cues across various

modalities during voluntary movements. Overall, this study supports the position that the CNS

combines and integrates sensory information in a flexible and task specific manner, but such

integration during voluntary action is not necessarily optimal.

Page 41: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

35

4.1.11 References

Alais, D., & Burr, D. (2004). The ventriloquist effect results from near-optimal bimodal

integration. Current Biology, 14(3), 257–262. doi:10.1016/j.cub.2004.01.029

Cavanagh, P. R., & Landa, J. (1976). A biomechanical analysis of the karate chop. Research

Quarterly, 47(4), 610–618. doi:10.1080/10671315.1976.10616718

Celesia, G. G. (1971). Organization of auditory cortical areas in man. Brain, (99), 403–414.

Colonius, H., & Diederich, A. (2004). Multisensory interaction in saccadic reaction time: A

time-window-of-integration model. Journal of Cognitive Neuroscience, 16(6), 1000–9.

doi:10.1162/0898929041502733

Driver, J., & Noesselt, T. (2008). Multisensory interplay reveals crossmodal influences on

“sensory-specific” brain regions, neural responses, and judgments. Neuron, 57(1), 11–23.

doi:10.1016/j.neuron.2007.12.013

Dulberg, M. S., Amant, R. S., & Zettlemoyer, L. S. (1999). An imprecise mouse gesture for the

fast activation of controls. In A. M. Sasse, & C. Johnson (Eds.), Proceeding of Human-

Computer Interaction – INTERACT 1999 (pp. 375-385) Amsterdam, NL: IOS Press.

Ernst, M. O., & Banks. M. S. (2002). Humans integrate visual and haptic information in a

statistically optimal fashion. Nature, 415, 429–434. doi:10.1038/415429a

Ernst, M. O., & Bülthoff, H. H. (2004). Merging the senses into a robust percept. Trends in

Cognitive Sciences, 8(4), 162–169. doi:10.1016/j.tics.2004.02.002

Page 42: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

36

Fedorov, N. T., & Mkrticheva, L. (1938). Mechanism of light flicker fusion during the course of

dark and light adaptation. Nature, 142, 750–751.  doi:10.1038/142750a0

Fleischauer, M. A., & Sherwood, D. E. (2008). Development of motor error detection capability

in an underhand throwing task. In Beaulieu, N. P. (Ed) Physical Activity and Children:

New Research, 195-210e. New York: Nova Science Publishers.

Gepshtein, S., Burge, J., Ernst, M. O., & Banks, M. S. (2005). The combination of vision and

touch depends on spatial proximity. Journal of Vision, 5(11), 1013–23. doi:10.1167/5.11.7

Ghazanfar, A. A, & Schroeder, C. E. (2006). Is neocortex essentially multisensory? Trends in

Cognitive Sciences, 10(6), 278–285. doi:10.1016/j.tics.2006.04.008

Hecht, S. (1920). The dark adaptation of the human eye. The Journal of General Physiology,

2(5), 499–517. doi:10.1085/jgp.2.5.499

Jeannerod, M. (1984). The timing of natural prehension movements. Journal of Motor Behavior,

16(3), 235–254. doi:10.1080/00222895.1984.10735319

Jegede, E., Watts, S., Stitt, L., & Hore, J. (2005). Timing of ball release in overarm throws

affects ball speed in unskilled but not skilled individuals. Journal of Sports Sciences, 23(8),

805–16. doi:10.1080/02640410400021393

Juravle, G., Deubel, H., Tan, H. Z., & Spence, C. (2010). Changes in tactile sensitivity over the

time-course of a goal-directed movement. Behavioural Brain Research, 208(2), 391- 401.

doi:10.1016/j.bbr.2009.12.009

Page 43: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

37

Keele, S. W., & Posner, M. I. (1968). Processing of visual feedback in rapid movements. Journal

of Experimental Psychology, 77(1), 155–158. doi:10.1037/h0025754

Kennedy, A., Bhattacharjee, A., Hansen, S., Reid, C., & Tremblay, L. (2015). Online vision as a

function of real-time limb velocity: another case for optimal windows. Journal of Motor

Behavior, 47(6), 465-475. doi:10.1080/00222895.2015.1012579

Körding, K. P., & Wolpert, D. M. (2006). Bayesian decision theory in sensorimotor control.

Trends in Cognitive Sciences, 10(7), 319–326. doi:10.1016/j.tics.2006.05.003

Levy-Tzedek, S., Hanassy, S., Abboud, S., Maidenbaum, S., & Amedi, A. (2012). Fast, accurate

reaching movements with a visual-to-auditory sensory substitution device. Restorative

Neurology and Neuroscience, 30(4), 313–323. doi:10.3233/RNN-2012-110219

Proteau, L., Roujoula, A., & Messier, J. (2009). Evidence for continuous processing of visual

information in a manual video-aiming task. Journal of Motor Behavior, 41(3), 219–231.

doi:10.3200/JMBR.41.3.219-231

Saunders, J. A., & Knill, D. C. (2003). Humans use continuous visual feedback from the hand to

control fast reaching movements. Experimental Brain Research, 152(3), 341–352.

doi:10.1007/s00221-003-1525-2

Shams, L., Kamitani, Y., & Shimojo, S. (2000). What you see is what you hear. Nature, 408,

788. doi:10.1038/35048669

Tremblay, L., & de Grosbois, J. (2015). Why encode limb and body displacements in the

velocity domain? Neurophysiological and behavioral evidence. In T. Heinen, (Ed.),

Page 44: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

38

Advances in visual perception research (pp. 279-292). Hauppauge, NY, USA: Nova

Science Publishers.

Tremblay, L., & Nguyen, T. (2010). Real-time decreased sensitivity to an audio-visual illusion

during goal-directed reaching. PLoS One, 5(1), e8952. doi:10.1371/journal.pone.0008952

Witten, I. B., & Knudsen, E. I. (2005). Why seeing is believing: merging auditory and visual

worlds. Neuron, 48(3), 489–496. doi:10.1016/j.neuron.2005.10.020

Zelaznik, H. N., Hawkins, B., & Kisselburgh, L. (1983). Rapid visual feedback processing in

single-aiming movements. Journal of Motor Behavior, 15(3), 217–236.

doi:10.1080/00222895.1983.10735298

Page 45: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

39

Multisensory Integration at Peak Limb Velocity in a Within-4.2

Modality Temporal Order Judgment Task

Page 46: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

40

4.2.1 Abstract

During goal-directed action, sensory processing is critical. When sensory information

from the various modalities is concurrently available, the cues are integrated in a statistically

optimal fashion. However, previous reports have shown that limb velocity may have a

modulatory effect on unimodal and multisensory processing. The current study focused

specifically on multisensory integration at peak limb velocity. Participants were required to fling

their limb through a virtual target and align their peak limb velocity with the intersection of the

target. At peak limb velocity, participants were presented with two auditory, visual, or

audiovisual cues. Following completion of the flinging movement, participants reported which

side of the target the first sensory cue was presented (i.e., a temporal order judgment task [TOJ]).

Participants also made this TOJ when the sensory cues were presented at rest. It was found that

TOJs were more accurate following presentation at rest than at peak limb velocity. At rest,

participants were more accurate in the auditory and audiovisual condition relative to the visual

condition. In addition, when comparing accuracy in the audiovisual condition at rest to peak limb

velocity, it was found that performance significantly decreased only in this condition. Overall,

the results suggest that multisensory estimates obtained at peak limb velocity may be less robust

than those obtained under unimodal conditions.

Page 47: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

41

4.2.2 Introduction

Integrating information across the sensory modalities shapes the perception of our

environment. Temporal information from the auditory modality and spatial information from the

visual modality are relied upon not only to perceive our environment, but also to execute goal-

directed actions (Sedda, Monaco, Bottini, & Goodale, 2011). Within the context of goal-directed

action, visual information is heavily relied upon to execute spatially driven tasks (e.g., Keele &

Posner, 1968), while auditory information dominates the temporal domain (e.g., Recanzone,

2003). When auditory and visual information are concurrently available, the central nervous

system typically integrates the information in a statistically optimal way (e.g., Ernst & Bülthoff,

2004). Although optimal audiovisual integration has been reported at rest (e.g., Alais & Burr,

2004), engaging in action has been shown to have a modulatory effect on multisensory

integration (e.g., Tremblay & Nguyen, 2010). In addition, the results of shown in Chapter 4.1

further suggest that optimal integration of audiovisual feedback may not occur when presented

specifically at peak limb velocity (PLV). In the experiment reported herein, we provide further

evidence that the weighting of auditory and visual cues presented at peak limb velocity may be

suboptimal.

A classic approach used to study sensory processing has been the temporal order

judgment task. In this task, participants are exposed to a pair of sensory cues presented at various

stimulus onset asynchronies (SOA), with the participant reporting which cue was presented first

(e.g., Hirsh & Sherrick 1961; Kanabus, Szelag, Rojek, & Pöppel, 2002; Spence, Shore, & Klein,

2001). Due to the processing latencies across the senses (e.g., Jaekl & Harris, 2007), an

audiovisual event is perceived as simultaneous, if the visual stimulus is presented before the

Page 48: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

42

auditory stimulus by ~75-100 ms (e.g., Zampini, Shore, & Spence, 2003; Jaekl & Harris, 2007).

Temporal order judgment tasks are an excellent paradigm to assess sensory processing because

the temporal and spatial constraints of the task can be manipulated independently. By orienting

participants to attend to which cue was presented first (i.e., temporal constraints) as opposed to

which side was presented first (i.e., spatial constraints) it is possible to assess how temporal and

spatial information are taken into account when perceiving audiovisual events (e.g., Zampini et

al., 2003).

Despite the long-standing interest in temporal order judgment tasks (e.g., Hamlin, 1895),

this paradigm is not without its shortcomings and confounds. One of the most commonly cited

issues is the lack of spatial compatibility between the presentation of the auditory and visual

cues. Specifically, visual information is sometimes presented directly in front of the participant,

whereas auditory information is presented via headphones worn by the participant (e.g., Bushara,

Grafman, & Hallett, 2001; Jaskowski, Jaroszyk, & Hojan-Jezierska, 1990; cf: Spence et al.,

2001; Zampnini et al., 2003). As such, there exists a lack of precision in understanding how

sensory information is processed within the TOJ task due to this spatial confound (Zampini et al.,

2003). Further, the experiments cited above had participants perform the TOJ task only at rest,

neglecting sensory processing during action. Considering that we spend a large portion of time

executing goal-directed actions towards multisensory stimuli, it is pertinent to understand how

sensory information is processed during action. In the current study, participants were exposed to

spatially congruent auditory and visual information and completed a TOJ following stimulus

presentation at peak limb velocity or at rest.

Outside of the temporal order judgment literature, many experiments have examined how

sensory information is processed and utilized specifically during goal-directed action. During

Page 49: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

43

action, vision is deemed to be heavily relied upon throughout an ongoing movement (e.g., Keele

& Posner, 1968; Zelaznik, Hawkins, & Kisselburgh, 1983; Proteau, Roujoula & Messier, 2009).

In fact, some researchers have proposed that visual information is critical specifically during the

latter phases of a movement (e.g., Carlton, 1981; Woodworth, 1899), while others suggest that

early trajectory information is critical (e.g., Kennedy Bhattacharjee, Hansen, Reid, & Tremblay,

2015).

Auditory information can also be utilized to during goal-directed actions. Levy-Tzedek,

Hanassy, Abboud, Maidenbaum, and Amedi (2012) provided sighted participants with a sensory

substitution device (SSD) conveying movement related information via auditory feedback. When

comparing movement time, peak velocity, and amplitude, no differences were found for

movements performed with visual feedback, or with the SSD. In a similar study conducted by

Rosati, Oscari, Spagnol, Avanzini, and Masiero (2012), participants tracked a target on screen

while receiving continuous auditory feedback. Of particular interest are the results from

Experiment 1 and 3 of Rosati et al. (2012). These authors found that auditory feedback

significantly reduced tracking errors relative to when no auditory feedback was available. In

addition, auditory and visual feedback yielded greater performance on a novel visuomotor

transformation task relative to visual feedback alone (Rosati et al., 2012). Overall, this evidence

suggests that task specific auditory information can facilitate sensorimotor control.

When both auditory and visual information are available, the central nervous system

integrates this information in a statistically optimal way (Ernst & Bülthoff, 2004). Redundant

sensory information within a modality is combined, and information across multiple modalities

is integrated (i.e., in a body-centred frame of reference) to yield an optimal percept (Ernst &

Banks, 2002). An example of optimal audiovisual integration is the ventriloquist effect whereby

Page 50: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

44

the perceived spatial position of auditory information is biased towards the spatial position of the

simultaneously presented visual information (Alais & Burr, 2004; Howard & Templeton, 1966).

Multisensory integration has been reported at the neuronal level (e.g., Meredith, Nemitz & Stein,

1987) and numerous experiments have shown that the central nervous system is largely

organized for multisensory integration (see Driver & Nosselet, 2003; Ghazanfar & Schoerder,

2006 for reviews).

Despite the evidence that optimal integration occurs whenever auditory and visual events

are simultaneously presented, recent evidence suggests this may not always be the case.

Critically, many of the experiments reporting optimal multisensory integration involve a

participant whom passively observed an audiovisual event (see Witten & Knudsen, 2005 for a

review). Interestingly, studies examining multisensory integration during action have revealed

optimal multisensory integration may not always occur. In a study conducted by Tremblay and

Nguyen (2010), participants were exposed to the audiovisual fission/ fusion illusion (Shams,

Kamitani, & Shimojo, 2000) during a 30 cm reaching movement. The authors reported that the

degree to which participants were fusing the audiovisual event was modulated by limb velocity,

such that the least amount of fusion was reported at the highest limb velocities. Based on these

findings, it appears that multisensory integration may not always be optimal during goal-directed

action (see also Juravle, Deubel, Tan, & Spence, 2010).

It is important to note that the experiments cited in the preceding paragraph examined

motor skills where movement end was the critical moment in the trajectory. However, for other

movements such as overarm throwing and hitting a golf ball, the critical moment in the trajectory

is ideally peak limb velocity. As such, how visual information is utilized specifically when peak

limb velocity is the critical moment in the trajectory has gone relatively understudied. In the

Page 51: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

45

current study, sensory processing and integration was assessed specifically at this kinematic

marker.

In line with this approach, the experiment reported in Chapter 4.1 had participants

perform an upper-limb flinging movement where their goal was to align peak limb velocity with

the intersection of a virtual target (i.e., in the absence of terminal feedback). At peak limb

velocity, participants received augmented auditory, visual, or audiovisual feedback in an attempt

to facilitate performance. Relative to when no feedback was provided, it was found that

providing auditory feedback significantly reduced the variability in position of the peak limb

velocity occurrence along the primary movement axis, as well as the displacement traveled when

peak limb velocity was reached. Further, no significant improvements were found when visual or

audiovisual feedback was provided. Critically, the performance in the audiovisual condition was

significantly worse than the predicted MLE values (e.g., Ernst and Bülthoff, 2004). Overall, the

results from this study, in conjunction with those reported by Tremblay and Nguyen (2010),

suggest that when the limb is traveling at its peak velocity, optimal integration of audiovisual

cues may not occur.

In line with the Tremblay and Nguyen (2010) results, the current experiment sought to

further test whether optimal integration occurs when the limb is traveling at peak limb velocity.

To this end, participants performed an upper-limb flinging movement with the goal of aligning

their peak limb velocity with the intersection of a virtual target. At peak limb velocity,

participants were presented with two auditory, visual, or audiovisual cues and reported which

cue was presented first in a within-modality temporal order judgment. Participants also

completed this judgment while remaining stationary. During action, it was expected that

performance across conditions would decrease relative to TOJs performed at rest (e.g., Juravle et

Page 52: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

46

al., 2010). Based on the MLE (Ernst & Bülthoff, 2004) it was hypothesized that participants

would be more accurate at rest in the audiovisual condition relative to the auditory and visual

conditions. If audiovisual information is integrated optimally during action, performance in the

TOJ should still be best in the audiovisual condition. Alternatively, if audiovisual cues are not

integrated optimally at peak velocity, performance should significantly decrease following

presentation at peak velocity compared to at rest.

4.2.3 Methods

4.2.4 Participants

The University of Toronto Research Ethics Board approved the experiment reported

herein. Of the sixteen participants recruited, three were excluded because their TOJ accuracy at

rest was below chance (i.e., average of 30 or less correct responses across conditions). As such,

data from thirteen individuals (3 males) with an average age of 22.75 years (SD = 4.5) was

analyzed. All participants provided informed consent prior to participation. Participants were

recruited from the undergraduate student population at the University of Toronto. All

participants were right-handed (i.e., self-reported), and had normal or corrected-to-normal vision

at the time of participation. For completing the experiment, participants were compensated $10.

4.2.5 Apparatus

Participant’s heads rested on a chin rest (38 cm tall) fastened to a 61 cm by 75 cm table

(see Figure 6 for a depiction of the experimental apparatus) for during the experimental trials. A

custom built wooden frame (48 cm tall) with a reflective surface 12 cm in diameter was

positioned 13 cm away from the home position. Attached to the forehead mount portion of the

Page 53: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

47

chin rest was a yellow LED light (1 cm in diameter), which served as the target for the

experiment. The target was visible to the participant through the reflection such that the target

appeared suspended in the air above the table. Two piezo-LED devices were placed on the

wooden frame 1.5 cm to the left and right of the virtual target. Both devices consisted of a green

LED light affixed to a piezo-buzzer (2350 Hz, 75 dB).

Participants wore a Life Brand model 225ZZ wrist brace with splint (Shoppers Drug

Mart, Toronto, ON, Canada) to restrict movement of the hand and wrist. Each trial began with

the participant’s finger on the home position (i.e., a 1.5 cm by 1.5 cm piece of Velcro). The

position of the participant’s right limb was

recorded throughout the experiment using an

Optotrak Certus (Northern Digital Inc.,

Waterloo, ON, Canada) motion capture system.

An infrared emitting diode (IRED) was

positioned on the distal epiphysis of the

participant’s ulna. The Optotrak real-time

sampling rate was set at 500 Hz and controlled

using a custom Matlab script (The MathWorks

Inc., Natwick, MA, USA). This Matlab script

also controlled the piezo-LED devices through an analog output board (PCI-6024E, National

Instruments Corp., Austin, TX, USA). Participant’s movements were measured along the Y-axis

(i.e., primary movement axis), as well as the X and Z-axes (i.e., secondary movement axes).

The piezo-LED devices presented sensory cues to participants either at peak limb

velocity or while the participant remained at rest. On every trial, both piezo-LED devices would

Figure 6. A depiction of the experimental apparatus. This figure

is not to scale.

Page 54: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

48

present two auditory beeps, two visual flashes, or two audiovisual events. The stimulus onset

asynchrony (SOA) was + or - 20 ms. When the cues were presented at peak limb velocity,

instantaneous velocity was calculated using the position data collected from two subsequent

samples gathered by the Optotrak. Movement start location was defined when the limb first

exceeded a velocity of 0.03 m/s. Peak velocity (i.e., as measured by the IRED position on the

wrist) was marked as two consecutive samples with a decrease in velocity, after the limb had

exceeded a minimum velocity of 0.8 m/s. Because participants were aiming with their index

finger while velocity was measured on the wrist, it was pertinent to ensure the moment of peak

velocity occurrence was consistent between the finger and wrist. To this end, three participants

who did not participate in this experiment completed a brief pilot study. In this study,

participants performed the flinging task for thirty trials where no feedback was given. In this

pilot study, limb velocity was measured via an IRED positioned on the index finger (i.e., Chapter

4.1) and the ulna (i.e., Chapter 4.2). Pearson correlations were used to compare the sample at

which peak limb velocity occurred for both IRED locations. This analysis revealed a .99

correlation between the sample at which peak limb velocity occurred on the finger and ulna.

Thus, although limb velocity was measured on the wrist, the point at which peak limb velocity

occurred was consistent with the moving finger.

4.2.6 Procedure

The target and home position were estimated by the participant and recorded using the

Optotrak prior to beginning the experiment. The home position was estimated by having the

participant rest their index finger on the home position for a trial. To estimate the target position,

the participants then moved their finger out to the observed position of the virtual target. During

the experimental trials, participants were instructed to “fling” their index finger through the

Page 55: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

49

virtual target (i.e., participants did not make physical contact with a target) and align their peak

velocity when the finger intersected the target (see Figure 7). A virtual target was used because

physical contact with an object would have yielded tactile feedback, hence obscuring the actual

effects of the investigated visual and auditory information processing. Upon completion of the

flinging movement, participants returned to the home position and awaited the signal for the next

trial.

Figure 7. An illustration of the flinging movement performed by participants. The target appeared roughly 40 cm away from the

starting position.

Page 56: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

50

Participants first completed 20 familiarization trials to habituate to the task dynamic (i.e.,

viewing a target via a reflection). During these trials, participants were given audiovisual

feedback when their limb reached peak velocity via the piezo-LED devices positioned on either

side of the virtual target. Both devices were triggered simultaneously, with the participant

instructed to have their finger as close to the target as possible when the audiovisual feedback

was given. After, participants completed 120 experimental trials per sensory condition in a

blocked and counter-balanced order. Within each block (i.e., auditory, visual, or audiovisual),

participants were presented with two auditory beeps, visual flashes, or audiovisual events from

either side of the virtual target as described above. As mentioned above, the onset of the stimuli

was held constant at a SOA of + or - 20 ms. The side of the first cue was pseudo-randomized

during each block. Specifically, participants were exposed to 30 trials where the first cue was

presented to the left of the target, and 30 where the first cue was presented from the right of the

target, at rest and at peak limb velocity.

Participants also completed TOJs following stimuli presentation at rest. During these

trials, the virtual target would become illuminated, followed 100 ms after by the first sensory cue

on either side of the target as described above. Participants were instructed to maintain fixation

on the remembered location of the virtual target. Following completion of all trials, participants

responded “Left first” or “Right first” via a button box held in their left hand. Lastly, the

calculation of our dependent variables (i.e., constant error, variable error, resultant displacement,

and movement time) was based on the IRED position at peak limb velocity.

4.2.7 Results

Means and standard deviations, as well as accuracy results can be found in Table 2.

Participants completed the flinging movement in 203 ms on average. Across conditions, the

Page 57: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

51

average peak limb velocity (i.e., as measured by the IRED location) achieved was 1.84 m/s.

Movement time and average peak velocity were analyzed using separate one-way repeated

measures ANOVAs. These tests confirmed that movement time, F (2, 24) = .28, p = .76, 𝜂!! =

.02, and peak velocity, F (2, 24) = 1.1, p = .37, 𝜂!! = .08, did not significantly vary as a function

of sensory cue. In addition, the ANOVA conducted for end position along the primary movement

axis (i.e., Y-axis) yielded no significant effects of sensory cue, F (2, 24) = 1.5, p = .24, 𝜂!! = .21.

This effect was also found when analyzing end position along the secondary axes: X-axis: F (2,

24) = 1.7, p = .85, 𝜂!! = .01; Z-axes: F (2, 24) = .30, p = .97, 𝜂!! = .02. Finally, the ANOVA

conducted for resultant displacement yielded no significant effect of sensory cue, F (2, 24) = .72,

p = .50, 𝜂!! = .01.

Table 2

Means and Between-subject Standard Deviations for Movement Time (MT), Average Peak Limb

Velocity (PLV), Constant Error (CE) and Variable Error (VE) Across the Three Movement Axes

(Y-X-Z), and CE and VE for Resultant Displacement (R).

MT (ms)

PLV (m/s)

CE-Y (mm)

CE-X (mm)

CE-Z (mm)

CE-R (mm)

VE-Y (mm)

VE-X (mm)

VE-Z (mm)

VE-R (mm)

Sensory Condition

M (SD)

M (SD)

M (SD)

M (SD)

M (SD)

M (SD)

M (SD)

M (SD)

M (SD)

M (SD)

Auditory

206.1

(64.1)

1.9

(0.1)

-46.79

(31.7)

-43.8

(62.1)

-54.3

(28.7)

99

(32.8)

29.5

(10.7)

16.1

(6.3)

23.4

(10.1)

35.4

(19.1)

Visual

198.3

(33.6)

1.8

(0.1)

-49.31

(40.4)

-46.1

(69.4)

-53.4

(33.6)

99.3

(43.3)

24.3

(7.1)

19.3

(10.3)

25.6

(8.8)

43.2

(19)

Audiovisual

204.5

(44.3)

1.9

(0.1)

-40.1

(30.7)

-44.4

(58.6)

-54.1

(29.5)

104.9

(33)

23.6

(7)

14.2

(4.5)

18.9

(5.3)

25.5

(7.2)

Page 58: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

52

In contrast, when analyzing accuracy (i.e., the number of correct responses) when the

participant remained at rest, a significant main effect of sensory cue was found, F (2,24) = 7.7, p

= .01, 𝜂!! = .49. Tukey’s HSD post hoc comparisons revealed that participants were more

accurate in the auditory condition than the visual condition (HSD = 5.3, Mean Auditory = 41.6,

Mean Visual = 33.7, p = .01). Post hoc comparisons also revealed participants were more

accurate in the audiovisual condition compared to the visual condition (HSD = 4.4, Mean

Audiovisual = 39.1, Mean Visual = 33.7, p = .05). The analysis conducted on TOJ accuracy

following stimuli presentation at rest versus at peak velocity yielded a significant main effect, t

(38) = 2.7, p = .01, with participants being more accurate at rest (M = 38.2, SD = 7.9) than at

peak limb velocity (M = 35.1, SD = 6.9). In addition, the interaction between sensory cue and

movement failed to reach significance, F (2, 24) = 2.1, p = .15, 𝜂!! = .13. However, because of

the large effect size associate with the interaction, it was deemed theoretically relevant to carry

out post hoc comparisons (e.g., Field, 2009). When analyzing within the sensory modalities, it

was also found that accuracy only decreased significantly between movement conditions in the

audiovisual condition (HSD = 5.4, Mean Audiovisual Rest = 39.1, Mean PLV = 33.6, p = .05).

These findings are depicted in Figure 8 (see also Table B).

Page 59: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

53

Figure 8. Number of correct responses as a function of movement and sensory conditions.

4.2.8 Discussion

The processing of unimodal cues, and the integration of multisensory cues at rest and at

peak limb velocity (i.e., as measured by the velocity of the wrist) was examined in the current

experiment. Participants were required to fling their index finger through a virtual target, while

reaching peak limb velocity when intersecting the virtual target (i.e., without terminal feedback).

At peak limb velocity, participants were presented with two auditory, visual, or audiovisual cues.

After the trial, their task was to discern which side of the virtual target the first sensory cue was

presented. This temporal order judgment task was also completed while the participant remained

stationary. In the resting condition, participants were most accurate in the auditory and

audiovisual conditions relative to the visual condition. However, when comparing within sensory

conditions and between movement conditions, accuracy only significantly decreased in the

audiovisual condition between the rest and flinging conditions. These results from the TOJ task

Page 60: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

54

being secondary to the flinging task actually reflect those of Chapter 4.1 where the cues were

relevant for the flinging task. The results thus suggest that engaging in goal-directed action has a

modulatory effect on multisensory integration.

The results of the current experiment suggest that visual information gathered at peak

limb velocity may not be particularly useful when peak limb velocity is the critical moment in

the trajectory, as evidenced by the poor performance in the visual condition. As such, the results

of the current experiment further suggest that visual information collected early on in the

trajectory is critical to control the amplitude of a rapidly moving limb. This finding is also in

agreement with Kennedy et al. (2015), who found that visual information provided early on in

the trajectory yielded comparable end-point variability to when vision was provided throughout

the trajectory.

Participants were most accurate at rest in the auditory and audiovisual conditions. As

such, the results are in agreement with the MLE (Ernst & Bültoff, 2004) whereby audiovisual

stimuli presentation facilitated performance in the TOJ relative to unimodal cues (e.g., vision).

One potential explanation for the advantage of performing TOJs when auditory information was

available comes from the Modality Appropriateness Hypothesis put forward by Welch and

Warren (1980). In this view, the influence of each modality for sensory integration is dependent

on the modality’s appropriateness within the task. Perhaps, participants took advantage of the

temporal precision of the auditory modality (e.g., Celesia, 1976). When visual information was

presented alone, the estimate obtained was less robust than audiovisual estimates due to the

temporal ambiguity of the visual estimate.

When auditory and visual information were presented at peak limb velocity, the results

suggest these cues were not optimally integrated. One potential explanation for this finding may

Page 61: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

55

be that participants failed to re-weight bimodal sensory information collected at peak limb

velocity in a statistically optimal fashion. Specifically, the estimate of the limb obtained at peak

limb velocity is particularly noisy due to the speed of the moving limb (e.g., Ghez & Sainburg,

1995). When this information is integrated with the audiovisual information, a more reliable

percept is not yielded due to weight assigned to the visual information. As such, an optimal

strategy would be to lower the weight of the visual information. However, it appears that

participants actually increased the weight on visual information when engaging in the visually

guided flinging movement (e.g., Sober & Sabes, 2005). Thus, performance in the TOJ task may

have decreased only when audiovisual information was presented at peak limb velocity because

of the increased weight on visual information, despite its worse reliability as compared to the

auditory cue during a visuomotor task. As such, it would be of interest to perform this

experiment with an auditory spatial target to see how target modality influences sensory

processing.

Overall, the current study provides further evidence that sensory processing is modulated

during voluntary movement. Specifically, the results suggest that the integration of audiovisual

information yields less robust estimates than if the estimates would be optimally integrated.

Although the inability to optimally re-weight sensory information presented at peak limb

velocity in a statistically optimal fashion may help explain the results, further work is needed to

uncover specifically how the central nervous system processes information during rapid goal-

directed actions. Overall, the results of this study in conjunction with those found in Chapter 4.1,

suggest that when a limb is traveling at high velocities during visuomotor tasks, the integration

of audiovisual information is suboptimal.

Page 62: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

56

References

Adhikari, M, B., Goshorn, E. S., Lamichhane, B., & Dhamala, M. (2013). Temporal-order

judgment of audiovisual events involves network activity. Brain Connectivity, 3(5), 536–

545. doi:10.1089/brain.2013.0163

Alais, D., & Burr, D. (2004). The ventriloquist effect results from near-optimal bimodal

integration. Current Biology, 14(3), 257–262. doi:10.1016/j.cub.2004.01.029

Bushara, K. O., Grafman, J., & Hallett, M. (2001). Neural correlates of auditory – visual

stimulus onset asynchrony detection. Journal of Neuroscience, 21(1), 300–304.

Carlton, L. G. (1981). Visual information: The control of aiming movements. The Quarterly

Journal of Experimental Psychology, 33(1), 87-93.

Celesia, G. G. (1976). Organization of auditory cortical areas in man. Brain, 99(3), 403–414.

Driver, J., & Noesselt, T. (2008). Multisensory interplay reveals crossmodal influences on

“sensory-specific” brain regions, neural responses, and judgments. Neuron, 57(1), 11–23.

doi:10.1016/j.neuron.2007.12.013

Ernst, M. O., & Banks. M. S. (2002). Humans integrate visual and haptic information in a

statistically optimal fashion. Nature, 415, 429–434. doi:10.1038/415429a

Ernst, M. O., & Bülthoff, H. H. (2004). Merging the senses into a robust percept. Trends in

Cognitive Sciences, 8(4), 162–169. doi:10.1016/j.tics.2004.02.002

Field, A. P. (2009). Discovering statistics using SPSS. London, England: SAGE Publications

LTD

Page 63: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

57

Ghazanfar, A. A, & Schroeder, C. E. (2006). Is neocortex essentially multisensory? Trends in

Cognitive Sciences, 10(6), 278–285. doi:10.1016/j.tics.2006.04.008

Ghez, C., & Sainburg, R. (1995). Proprioceptive control of interjoint coordination. Canadian

Journal of Physiology and Pharmacology, 73(2), 273–284.

Hamlin, A. J. (1895) On the least observable interval between stimuli addressed to disparate

senses and to different organs of the same sense. American Journal of Psychology, 6, 564–

575

Hirsh, I. J., & Sherrick, C. E. J. (1961). Perceived order in different sense modalities. Journal of

Experimental Psychology, 62(5), 423–432. doi:10.1037/h0045283

Howard, I. P., & Templeton, W. B. (1966). Human spatial orientation. Oxford, England: Wiley

and Sons.

Jaekl, P. M., & Harris, L. R. (2007). Auditory-visual temporal integration measured by shifts in

perceived temporal location. Neuroscience Letters, 417(3), 219–24.

doi:10.1016/j.neulet.2007.02.029

Jaśkowski, P., Jaroszyk, F., & Hojan-Jezierska, D. (1990). Temporal-order judgments and

reaction time for stimuli of different modalities. Psychological Research, 52(1), 35–38.

doi:10.1007/BF00867209

Juravle, G., Deubel, H., Tan, H. Z., & Spence, C. (2010). Changes in tactile sensitivity over the

time-course of a goal-directed movement. Behavioural Brain Research, 208(2), 391- 401.

doi:10.1016/j.bbr.2009.12.009

Page 64: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

58

Kanabus, M., Szelg, E., & Rojek, E. (2002). Temporal order judgment for auditory and visual

stimuli. Acta Neurobiologiae Experimentalis, 62, 263–270.

Keele, S. W., & Posner, M. I. (1968). Processing of visual feedback in rapid movements. Journal

of Experimental Psychology, 77(1), 155–158. doi:10.1037/h0025754

Kennedy, A., Bhattacharjee, A., Hansen, S., Reid, C., & Tremblay, L. (2015). Online vision as a

function of real-time limb velocity: another case for optimal windows. Journal of Motor

Behavior, 47(6), 465-475. doi:10.1080/00222895.2015.1012579

Levy-Tzedek, S., Hanassy, S., Abboud, S., Maidenbaum, S., & Amedi, A. (2012). Fast, accurate

reaching movements with a visual-to-auditory sensory substitution device. Restorative

Neurology and Neuroscience, 30(4), 313–323. doi:10.3233/RNN-2012-110219

Meredith, M. A., Nemitz, W. J., & Stein, B., E. (1987). Determinants of multisensory integration

neurons. I. temporal factors in superior colliculus. Journal of Neuroscience, 7(10), 3215–

3229.

Proteau, L., Roujoula, A., & Messier, J. (2009). Evidence for continuous processing of visual

information in a manual video-aiming task. Journal of Motor Behavior, 41(3), 219–231.

doi:10.3200/JMBR.41.3.219-231

Recanzone, G. H. (2003). Auditory Influences on Visual Temporal Rate Perception. Journal of

Neurophysiology, 89(2), 1078–1093.

Rosati, G., Oscari, F., Spagnol, S., Avanzini, F., & Masiero, S. (2012). Effect of task-related

continuous auditory feedback during learning of tracking motion exercises. Journal of

Neuroengineering and Rehabilitation, 9(79), 2-13. doi:10.1186/1743-0003-9-79

Page 65: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

59

Sedda, A., Monaco, S., Bottini, G., & Goodale, M. A. (2011). Integration of visual and auditory

information for hand actions: Preliminary evidence for the contribution of natural sounds to

grasping. Experimental Brain Research, 209(3), 365–74. doi:10.1007/s00221-011-2559-5

Shams, L., Kamitani, Y., & Shimojo, S. (2000). What you see is what you hear. Nature, 408,

788. doi:10.1038/35048669

Spence, Charles, Shore, I., D., Klein, M, R. (2001). Multisensory prior entry. Journal of

Experimental Psychology: General, 130(4), 799–832.

Tremblay, L., & de Grosbois, J. (2015). Why encode limb and body displacements in the

velocity domain? Neurophysiological and behavioral evidence. In T. Heinen (Ed.),

Advances in visual perception research (pp. 279-292). Hauppauge, NY, USA: Nova

Science Publishers.

Tremblay, L., & Nguyen, T. (2010). Real-time decreased sensitivity to an audio-visual illusion

during goal-directed reaching. PLoS One, 5(1), e8952. doi:10.1371/journal.pone.0008952

Welch, R. B., & Warren, D. H. (1980). Immediate perceptual response to intersensory

discrepancy. Psychological Bulletin, 88(3), 638-667

Witten, I. B., & Knudsen, E. I. (2005). Why seeing is believing: Merging auditory and visual

worlds. Neuron, 48(3), 489–496. doi:10.1016/j.neuron.2005.10.020

Woodworth, R. S. (1899). The accuracy of voluntary movement. Psychological Review, 3, 1-

119.

Page 66: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

60

Zampini, M., Shore, D. I., & Spence, C. (2003). Audiovisual temporal order judgments.

Experimental Brain Research, 152(2), 198–210. doi:10.1007/s00221-003-1536-z

Zelaznik, H. N., Hawkins, B., & Kisselburgh, L. (1983). Rapid visual feedback processing in

single-aiming movements. Journal of Motor Behavior, 15(3), 217–236.

doi:10.1080/00222895.1983.10735298

Page 67: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

61

Chapter 5 – General Discussion 5

Experiment 1 Summary 5.1

The use of unisensory and multisensory feedback was the focus of Experiment 1 (i.e.,

feedback experiment). In this experiment, participants were tasked with flinging their right limb

through the centre of a virtual target. Their goal was to reach peak limb velocity (i.e., as

measured by index finger velocity) when the finger intersected a virtual target. At peak limb

velocity, participants were presented with auditory, visual, or audiovisual feedback via a piezo-

LED device affixed to the right index finger. It was hypothesized that performance would be best

in the multisensory conditions as evidenced by reductions in constant and variable errors as well

as resultant displacement. Contrary to this hypothesis, it was found that auditory feedback alone

reduced variability in peak velocity occurrence across all movement axes (see Figures 1 & 5),

relative to when no feedback was provided. No significant effects were observed for the visual or

audiovisual feedback conditions. The results from this experiment suggest that when the critical

moment in the trajectory is peak limb velocity (i.e., as opposed to movement end), the use of

visual feedback may be limited. In addition, performance in the audiovisual condition suggested

that participants failed to re-weight unreliable sensory information in a statistically optimal

fashion, as evidenced by the significant differences between MLE predicted and observed

variable errors.

Experiment 2 Summary 5.2

Experiment 2 (i.e., TOJ experiment) further assessed whether audiovisual cues were

optimally integrated, or not, at peak limb velocity when peak limb velocity is the critical portion

of the trajectory. In this experiment, participants were again required to fling their limb through

Page 68: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

62

the centre of a virtual target. At peak limb velocity (i.e., as measured by wrist velocity)

participants were presented with two auditory, visual, or audiovisual cues via two piezo-LED

devices affixed on either side of the virtual target. Participants had to report which side of the

virtual target was presented first in a temporal order judgment (TOJ) task. Participants also

completed the TOJ task while remaining stationary. It was found that at rest, participants were

more accurate in the auditory and audiovisual condition relative to the visual condition. When

the sensory cues were presented at peak limb velocity, it was found that participant’s TOJ

accuracy was significantly worse in the audiovisual condition compared to at rest. The results of

this experiment suggested that participants prioritized the processing of the least reliable cue

(i.e., vision) when performing the flinging movement towards a visual target.

General Discussion Overview 5.3

Sensory processing at peak limb velocity was assessed across two experiments in the

current thesis. The results reported here have implications for sensory processing throughout an

ongoing movement and the integration of audiovisual information, specifically at peak limb

velocity when this kinematic marker was the critical moment in the trajectory. The discussion

below pertains to sensory processing during goal-directed action with emphasis on how sensory

processing differs when the critical moment in the trajectory is peak limb velocity, not

movement end. In addition, how sensory information is integrated at peak limb velocity will be

discussed with emphasis on the implications of the current results on the use of sensory

information during goal-directed action.

Page 69: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

63

5.3.1 Critical Temporal Window For Visual Processing

The current thesis sought to determine how sensory information is processed and

integrated during movements where the critical moment for feedback utilization is peak limb

velocity, as opposed to movement end. Although the empirical focus on movement end is

certainly warranted to understand the fundamental aspects of common upper-limb movements

(i.e., reach-to-grasp and pointing), greater emphasis could be placed on movements with

alternative critical moments. By investigating how sensory information is utilized specifically at

peak limb velocity, the results of the current thesis have implications for feedback processes

occurring throughout an ongoing movement.

Visual information provides important feedback at the end of a movement. Although

previous studies have shown the utility of visual feedback at movement end (e.g., Zelaznik et al.,

1983; Chua & Elliott, 1993; Ma-Wyatt & McKee, 2007) the current thesis provides evidence that

visual feedback at peak limb velocity does little to facilitate precision when the limb is traveling

at peak velocity. This finding may be due to differing feedback processes between the critical

moments in the trajectories (i.e., pointing versus flinging). In a study conducted by Proteau,

Roujoula and Messier (2009), participants performed a manual-aiming task where the position of

the limb was represented by a cursor on screen. During the trajectory, the position of the cursor

was perturbed (i.e., lateral cursor jump soon after movement start). Proteau et al. (2009) reported

that participants were able to perform fast and accurate corrections to the cursor jump and

complete the movement accurately. This finding led the authors to conclude that when

movement end is the critical moment in the trajectory, visual information is continuously

monitored. However, other work has shown that visual processing is critical during specific

phases of a movement.

Page 70: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

64

Visual information gathered at specific limb velocities may be critical when executing

goal-directed actions. In a study conducted by Kennedy et al. (2015), participants performed a 30

cm reaching movement within a 270-350 ms movement time bandwidth. Visual information was

provided to participants at certain limb velocities. That is, visual information was provided early

(i.e., limb velocity of 0.8-1.4 m/s, before peak limb velocity), in the middle (i.e., limb velocity

above 1.4 m/s, up to peak limb velocity), or late (i.e., limb velocity of 1.4-0.8 m/s, after peak

limb velocity) in the trajectory. It was found that endpoint precision and time after peak limb

velocity were comparable to full vision when visual information was provided during the early

phase of the trajectory. In contrast, the middle window did exhibit comparable endpoint

precision to full vision, but at a significant temporal cost. This finding in conjunction with those

reported by Proteau et al. (2009) suggest that when the critical moment in the trajectory is

movement end, visual processing within certain phases of the movement are more salient than

others. The current thesis was able to extend this knowledge by manipulating the accuracy

requirements of the task (i.e., make peak limb velocity the critical moment instead of movement

end). The results from the current thesis highlight the importance of visual feedback early in the

trajectory, when controlling movement amplitude. This is evidenced by the lack of facilitation of

the visual feedback in the feedback experiment (see Figure 5), and the poor performance in the

visual condition of the TOJ experiment (see Figure 8). Thus, it appears that visual information at

peak limb velocity is not very salient during upper-limb movements, but rather visual

information gathered early on in the trajectory is critical for the successful execution of goal-

directed actions.

Page 71: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

65

5.3.2 Multisensory Integration At Peak Limb Velocity

The experiments conducted in the current thesis assessed whether audiovisual integration

is optimal, specifically when the limb is travelling at peak velocity. This pursuit is warranted

given previous work reporting that multisensory integration was modulated by limb velocity

(e.g., Tremblay & Nguyen, 2010). In addition, the perception of unisensory stimuli has been

shown to decline during a movement (e.g., Juravle et al., 2010). The results of the current thesis

suggest that during action, audiovisual integration may not be optimal. The discussion below

pertains to why multisensory integration may not be optimal at peak limb velocity. As well,

alternative sensory processing mechanisms the central nervous system may be employing at peak

limb velocity will be discussed.

At peak limb velocity, sensory information may not be weighted in a statistically optimal

fashion. A potential explanation for the lack of audiovisual facilitation found in the experiments

is that an unreliable cue, and its weight in the integrated percept, reduced the overall reliability of

the perceptual estimate. Based on the results, it would appear that the least reliable stimulus was

the visual stimulus. In the feedback experiment for example, when the LED was illuminated in

both the visual and audiovisual conditions, it created a streak of light due to the high velocity of

the moving limb. This streak of light would most certainly reduce the quality of the visual

estimate, given that the participant could associate peak limb velocity with any point along this

streak. According to the spatial nature of the task, it was predicted that participants would

heavily weight the visual cue (e.g., Rock & Victor, 1964). However, the quality of this estimate

is limited at peak limb velocity. As such, an optimal strategy would have been to re-weight the

visual estimate (i.e., reduce its overall weight during integration) in the audiovisual condition.

However, the current results do not suggest that participants were able to do so, as evidenced by

Page 72: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

66

the significant difference between observed and MLE predicted performance in the audiovisual

condition.

Previous work on sensory weighting has shown the weights assigned to each modality are

flexible. In a study conducted by Fetsch, Pouget, DeAngelis, and Angelaki (2011), monkeys

performed a heading discrimination task (i.e., judging the direction of linear self-motion)

utilizing visual and vestibular cues. On a trial-by-trial basis, the reliability of the visual and

vestibular cues was manipulated such that one cue would become more reliable than the other.

Of particular interest was neuronal activity in the dorsal medial superior temporal area (MSTd),

an area of the cortex believed to encode heading perception (i.e., integrating visual and vestibular

cues). Fetsch et al. (2011) reported that when the reliability of either cue was manipulated, the

monkeys were able to adapt and perform the task successfully, while corresponding fluctuations

in the firing rate of neurons within the MSTd was observed. This pattern of results was

interpreted as evidence that the monkeys re-weighted the sensory information on a trial-by-trial

basis based on the reliability of the cue.

A potential explanation for the lack of facilitation found for the visual and audiovisual

conditions is that participants fail to re-weight this cue (i.e., reduce its overall weight during

integration), when it is presented specifically at peak limb velocity. It is important to note that in

the study conducted by Fetsch et al. (2011), the animal was a passive observer of the sensory

events and did not actively engage in any sort of action. In the current thesis, participants were

required to integrate auditory and visual cues while the limb is travelling at a high velocity

during a goal-directed action. As would be expected given the results of Fetsch et al. (2011),

participants should have become aware of the visual estimate’s unreliability and adjusted its

weighted within the first few trials of the block. In order to further determine how participant’s

Page 73: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

67

performance changed over time in the feedback and TOJ experiment, an additional block

analysis was conducted (see Appendices A & B). In this analysis, trials were divided into thirds

(i.e., first third, middle third, and final third) with the interest here being whether participants

improved their performance over time. However, when analyzing the first third, middle third,

and final third trials of the block for the feedback experiment (see Appendix A) no significant

differences in variable error were found (i.e., all ps > .12). In addition, when conducting the same

analysis for the TOJ experiment, participant’s accuracy in the TOJ task again did not improve

over time (see Appendix B). This analysis suggests that the participant’s strategy, and weights

assigned to the sensory cues, did not change throughout the task. As such, it appears that when

unreliable sensory information is presented at peak limb velocity, the cues are not re-weighted in

a statistically optimal fashion.

An alternative explanation for participant’s performance across experiments may be

related to the reliance of visual information during reaches performed to visual targets. Sober and

Sabes (2005) contend that the sensory modality congruent with the target modality will be relied

upon during action. In both experiments, participants were presented with a visual target. Based

on Sober and Sabes (2005), it would be expected that visual processing would be prioritized

because it is congruent with the target modality. Within both experiments however, visual

information did not provide very salient information in comparison to the auditory cue (i.e., an

incongruent modality). As such, no facilitation in performance was found when visual

information was available across both experiments because participants would have upregulated

the processing of the lesser reliable sensory cue. Conducting the same study as described in

Chapter 4.1 with an auditory target could test this hypothesis. In this follow up study,

participants would perform the flinging movement towards a localizable auditory target. If

sensory information congruent with the target modality were upregulated, it would be expected

Page 74: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

68

that performance would be best in the audiovisual conditions, due to the upregulation of the most

reliable modality within the task.

Overall, more work is required to understand how the central nervous system processes

information specifically during instances of rapid goal-directed action. Although the results of

this thesis suggest that the weights assigned to auditory and visual information may be

suboptimal, it does not provide a definitive explanation as to why. As such, future work could

focus on how sensory information is processed within movement phases prior to peak limb

velocity. This approach could yield valuable insight into the acute moments of a movement that

are critical to its outcome. In addition, it is also important to study multisensory integration

during goal directed action in general, as this focus has gone relatively overlooked in the

literature (see Witten & Knudsen, 2005 for a review).

5.3.3 The Advantage of Auditory Feedback

The results of both experiments emphasize that auditory information is particularly useful

when peak limb velocity is the critical moment in the trajectory. Therefore, providing augmented

auditory feedback may be beneficial when learning certain motor skills. One such skill where

peak limb velocity is the critical moment in the trajectory is overarm throwing. Empirical work

on overarm throwing has shown that when a series of throws are performed, ball speed tends to

vary from throw to throw as a result of variability in ball release. Given that limb velocity peaks

at a specific moment in the trajectory, ball release either before or after this point negatively

affects ball speed (Timmann, Citron, Watts, & Hore, 2001). Because the feedback experiment

showed that auditory feedback reduced variability in displacement of the limb when presented at

peak velocity, auditory feedback may be a particularly useful to convey movement related

feedback during overarm throws.

Page 75: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

69

Auditory feedback may be particularly useful when learning ball release point in the

novice throwing population. In a study directly comparing release points between skilled and

unskilled throwers, Jegede, Watts, Stitt, and Hore (2005) found strong correlations between

release timing and forearm velocity for skilled, but not unskilled throwers. As a result of

variability in release, the unskilled throwers achieved a lower ball speed when compared to the

skilled throwers, while skilled throwers were timing their release close to peak limb velocity

(Jegede et al., 2005). The results of that study suggest that the optimal release window for

overarm throwing centres on peak limb velocity and that novice throwers fail to release the ball

reliably at this kinematic marker. As shown in Chapter 4.1, auditory feedback reduced the

variability in peak limb velocity occurrence across all movement axes as well as resultant

displacement. Therefore, auditory feedback provided at peak limb velocity may be useful for

novice throwers when learning ball release timing during a throwing task.

5.3.4 Limitations of the Current Thesis

A limitation of the feedback experiment was that participants always completed the no

feedback condition before the counterbalanced auditory, visual, and audiovisual blocks. This

block order certainly warrants the criticism that participants were more accurate in the feedback

trials relative to no feedback, simply due to more practice with the task. As such, in future

experiments involving the same flinging task, the no feedback condition should be

counterbalanced with the feedback conditions. Another limitation in both experiments is the

location of the virtual target. As noted under General Methods, each participant estimated the

location of the virtual target. As such, the observed position of the target may have varied

between participants. To correct for any potential variation, future experiments should calculate

the position of the virtual target by measuring the distance between the participant’s eyes and

Page 76: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

70

the reflective surface as well as the LED to the reflective surface. This position should then be

used when calculating dependent variables such as constant and variable error.

5.3.5 Concluding Remarks

Processing sensory information accurately is critical to successfully execute actions.

Although sensory information is typically combined in a statistically optimal way, the current

thesis suggests that weights assigned to auditory and visual information when the limb is

travelling at peak velocity is suboptimal. These findings may be related to the inability of

participants to reweight sensory information presented at peak limb velocity in a statistically

optimal fashion. As such, future work is required to understand how sensory information is

processed specifically when the limb is traveling at high velocities, such as when performing

overarm throws or hitting a golf ball. Pursuing this research could yield valuable applied and

fudamental knowledge. For example, understanding how non-visual sensory information is

processed specifically during action could lead to advances in augmented feedback provided

during motor learning (e.g., Effenberg, 2005). Finally, this focus may also lead to further

knowledge regarding sensory processing mechanisms utilized during movement planning and

execution.

Page 77: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

71

Chapter 6 – References 6

Alais, D., & Burr, D. (2004). The ventriloquist effect results from near-optimal bimodal

integration. Current Biology, 14(3), 257–262. doi:10.1016/j.cub.2004.01.029

Bremmer, F., Schlack, A., Shah, N. J., Zafiris, O., Kubischik, M., Hoffmann, K.-P., Ziles, K., &

Fink, R., G. (2001). Polymodal motion processing in posterior parietal and premotor

cortex: A human fMRI study strongly implies equivalencies between humans and

monkeys. Neuron, 29, 287–296.

Bushara, K. O., Grafman, J., & Hallett, M. (2001). Neural correlates of auditory – visual

stimulus onset asynchrony detection. Journal of Neuroscience, 21(1), 300–304.

Calvert, G. A., Brammer, M. J., Bullmore, E. T., Campbell, R., Iversen, S. D., & David, A. S.

(1999). Response amplification in sensory‐specific cortices during crossmodal binding.

Neuroreport, 10(12), 2619-2623.

Carlton, L. G. (1981). Processing visual feedback information for movement control. Journal of

Experimental Psychology: Human Perception and Performance, 7(5), 1019–1030.

Carlton, L. G. (1992). On the specificity of learning and the role of visual information for

movement control. In L. Proteau & D. Elliott (Eds.), Vision and motor control. Advances in

psychology, (pp. 3-31). New York, NY, USA: Elsevier Science & Technology.

Celesia, G., G. (1976). Organization of the auditory cortical areas of man. Brain, 99(3), 403-414.

Page 78: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

72

Chapman, C., Bushnell, M., Miron, D., Duncan, G., & Lund, J. (1987). Sensory perception

during movement in man. Experimental Brain Research, 68(3), 516-524.

Chua, R., & Elliott, D. (1993). Visual regulation of manual aiming. Human Movement Science,

12(4), 365–401. doi:10.1016/0167-9457(93)90026-L

Driver, J., & Noesselt, T. (2008). Multisensory interplay reveals crossmodal influences on

“sensory-specific” brain regions, neural responses, and judgments. Neuron, 57(1), 11–23.

doi:10.1016/j.neuron.2007.12.013

Effenberg, A. O. (2005). Movement sonification: Effects on perception and action. IEEE

Multimedia, 12(2), 53–59. doi:10.1109/MMUL.2005.31

Elliott, D., Carson, R. G., Goodman, D., & Chua, R. (1991). Discrete vs. continuous visual

control of manual aiming. Human Movement Science, 10, 393–418.

Elliott, D., Hansen, S., & Khan, M. A. (2010). The optimization of speed, accuracy, and energy

in goal-directed aiming. In D. Elliott & M. Khan (Eds.), Vision and goal-directed

movement: Neurobehavioral perspectives (pp. 21-38). Champaign, IL, USA: Human

Kinetics

Ernst, M. O., Banks. M. S. (2002). Humans integrate visual and haptic information in a

statistically optimal fashion. Nature, 415, 429–434.

Ernst, M. O., & Bülthoff, H. H. (2004). Merging the senses into a robust percept. Trends in

Cognitive Sciences, 8(4), 162–169. doi:10.1016/j.tics.2004.02.002

Page 79: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

73

Fetsch, C. R., Pouget, A., DeAngelis, G. C., & Angelaki, D. E. (2012). Neural correlates of

reliability-based cue weighting during multisensory integration. Nature Neuroscience,

15(1), 146–54. doi:10.1038/nn.2983

Ghazanfar, A. a, & Schroeder, C. E. (2006). Is neocortex essentially multisensory? Trends in

Cognitive Sciences, 10(6), 278–85. doi:10.1016/j.tics.2006.04.008

Hershenson, M. (1962). Reaction times as a measure of intersensory facilitation. Journal of

Experimental Psychology, 63(3), 289–293.

Hirsh, I. J., & Sherrick, C. E. J. (1961). Perceived order in different sense modalities. Journal of

Experimental Psychology, 62(5), 423–432. doi:10.1037/h0045283

Kanabus, M., Szelg, E., & Rojek, E. (2002). Temporal order judgment for auditory and visual

stimuli. Acta Neurobiologiae Experimentalis, 62, 263–270.

Jaekl, P. M., & Harris, L. R. (2007). Auditory-visual temporal integration measured by shifts in

perceived temporal location. Neuroscience Letters, 417(3), 219–24.

doi:10.1016/j.neulet.2007.02.029

Jaśkowski, P., Jaroszyk, F., & Hojan-Jezierska, D. (1990). Temporal-order judgments and

reaction time for stimuli of different modalities. Psychological Research, 52(1), 35–38.

doi:10.1007/BF00867209

Juravle, G., Deubel, H., Tan, H. Z., & Spence, C. (2010). Changes in tactile sensitivity over the

time-course of a goal-directed movement. Behavioural Brain Research, 208 (2), 391- 401.

Keele, S. W., & Posner, M. I. (1968). Processing of visual feedback in rapid movements. Journal

of Experimental Psychology, 77(1), 155–158.

Page 80: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

74

Kennedy, A., Bhattacharjee, A., Hansen, S., Reid, C., & Tremblay, L. (2015). Online vision as a

function of real-time limb velocity: another case for optimal windows. Journal of Motor

Behavior, 47(6), 465-475. doi:10.1080/00222895.2015.1012579

Lakatos, P., O’Connell, M. N., Barczak, A., Mills, A., Javitt, D. C., & Schroeder, C. E. (2009).

The leading sense: Supramodal control of neurophysiological context by attention. Neuron,

64(3), 419–30. doi:10.1016/j.neuron.2009.10.014

Lesevre, N. (1982). Chronotopographical analysis of the human evoked potential in relation to

the visual field (data from normal individuals and hemianopic patients). Annals of the New

York Academy Of Sciences, 388, 156–182.

Levy-Tzedek, S., Hanassy, S., Abboud, S., Maidenbaum, S., & Amedi, A. (2012). Fast, accurate

reaching movements with a visual-to-auditory sensory substitution device. Restorative

Neurology and Neuroscience, 30(4), 313–23. doi:10.3233/RNN-2012-110219

Ma-Wyatt, A., & McKee, S. P. (2007). Visual information throughout a reach determines

endpoint precision. Experimental Brain Research, 179(1), 55–64. doi:10.1007/s00221-006-

0767-1

McGurk, H., MacDonald, J. (1976). Hearing lips and seeing voices. Nature, 264, 746–748.

Meredith, M. A., Nemitz, W. J., & Stein, B. E. (1987). Determinants of multisensory integration

neurons. I. temporal factors in superior colliculus, 7(10), 3215–3229

Proteau, L., Roujoula, A., & Messier, J. (2009). Evidence for continuous processing of visual

information in a manual video-aiming task. Journal of Motor Behavior, 41(3), 219–31.

doi:10.3200/JMBR.41.3.219-231

Page 81: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

75

Rock, I., Victor, J. (1964). Vision and touch  : An experimentally created conflict between the

two senses. Science, 143(3606), 594–596.

Russo, G. S., & Bruce, C. J. (1994). Frontal eye field activity preceding aurally guided saccades.

Journal of Neurophysiology, 71(3), 1250–1253.

Shams, L., Kamitani, Y., & Shimojo, S. (2000). What you see is what you hear. Nature, 408,

788.

Shams, L., Kamitani, Y., & Shimojo, S. (2002). Visual illusion induced by sound. Cognitive

Brain Research, 14(1), 147–152.

Sober, S. J., & Sabes, P. N. (2005). Flexible strategies for sensory integration during motor

planning. Nature Neuroscience, 8(4), 490–497. doi:10.1038/nn1427

Spence, Charles, Shore, I., D., & Klein, M, R. (2001). Multisensory prior entry. Journal of

Experimental Psychology: General, 130(4), 799–832.

Stein, B. E., Meredith, M. A., Huneycutt, W. S., & McDade, L. (1989). Behavioral indices of

multisensory integration: Orientation to visual cues is affected by auditory stimuli. Journal

of Cognitive Neuroscience, 1(1), 12–24. doi:10.1162/jocn.1989.1.1.12

Tremblay, L., & Nguyen, T. (2010). Real-time decreased sensitivity to an audio-visual illusion

during goal-directed reaching. PloS One, 5(1), e8952. doi:10.1371/journal.pone.0008952

van Atteveldt, N., Murray, M. M., Thut, G., & Schroeder, C. E. (2014). Multisensory integration:

Flexible use of general operations. Neuron, 81(6), 1240–53.

doi:10.1016/j.neuron.2014.02.044

Page 82: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

76

van Beers, R. J., Sittig, A. C., & Gon, J. J. D. Van Der. (1999). Integration of proprioceptive and

visual position-information: An experimentally supported model. Journal of

Neurophysiology, 81, 1355–1364.

van Beers, R. J., Wolpert, D. M., & Haggard, P. (2002). When feeling is more important than

seeing in sensorimotor adaptation. Current Biology, 12(10), 834–837.

Warren, J. E., Wise, R. J. S., & Warren, J. D. (2005). Sounds do-able: Auditory-motor

transformations and the posterior temporal plane. Trends in Neurosciences, 28(12), 636–

643. doi:10.1016/j.tins.2005.09.010

Woodworth, R. S. (1899). The accuracy of voluntary movement. Psychological Review, 3, 1–

119.

Zampini, M., Shore, D. I., & Spence, C. (2003). Audiovisual temporal order judgments.

Experimental Brain Research, 152(2), 198–210. doi:10.1007/s00221-003-1536-z

Zelaznik, H. N., Hawkins, B., & Kisselburgh, L. (1983). Rapid visual feedback processing in

single-aiming movements. Journal of Motor Behavior, 15(3), 217–236.

doi:10.1080/00222895.1983.10735298

Page 83: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

77

Appendices 7

Page 84: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

78

Appendix A – Block Analysis of Experiment 1 7.1

The additional analysis reported below tested how participants utilized feedback

throughout the experimental block. As such, the first third (i.e. trials 1-10), middle third (i.e.,

trials 11-20), and final third (i.e., trials 21-30) trials of the block were analyzed using repeated

measures ANOVA, across the movement axes. The analysis of variable error revealed that

participants did not significantly improve their performance across the block of trials in any

feedback condition (i.e., all p ≥ .27). Specifically, performance along the primary, F (2, 24) =

1.4, p = .27, 𝜂!! = .16, secondary, F (2, 24) = 1.41, p = .27, 𝜂!! = .11, and tertiary, F (2, 24) = 1.2,

p = .32,  𝜂!! = .09, axes when visual feedback was provided showed participants did not

significantly reduce their variability over time. This pattern of results was mirrored along the

primary, F (2, 24) = .47, p = .63, 𝜂!! = .04, secondary, F (2, 24) = .65, p = .53, 𝜂!! = .05, and

tertiary, F (2, 24) = .65, p = .53, 𝜂!! = .04, axes when audiovisual feedback was provided. Lastly,

no significant reduction in variability of displacement traveled when peak velocity was reached

when visual, F (2, 24) = 1.75, p = .12, 𝜂!! = .01, or audiovisual, F (2, 24) = .38, p = .69,  𝜂!! = .03,

feedback was provided. The means and standard deviations associated with these results are

shown in Table A.

Page 85: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

79

Table A

Variable Error (VE) Across Feedback Block for the Three Movement Axes (i.e.,

Y-X-Z) and Resultant Displacement (R).

Proportion of Experimental Block

VE-X (mm)

VE-Z (mm)

VE-R (mm)

Sensory Condition

First Third M (SD)

Middle Third M (SD)

Final Third M (SD)

Total M (SD)

Auditory VE-Y VE-X VE-Z VE-R

28.6 (14) 12.2 (3.6) 20.9 (11.5) 32.4 (17.6)

27.4 (8.1) 9.1 (2.8) 19.3 (4.5) 30.8 (8.8)

25.6 (11.4) 9.4 (4.2) 19.9 (5.6) 29.1 (10.2)

27.7 (11.5) 10.3 (3.8) 20.1 (8.4) 30.8 (14.8)

Visual VE-Y VE-X VE-Z VE-R

34.2 (13.4) 12.2 (4.3) 24.4 (5.4) 42.2 (13.1)

39.2 (23.1) 11.3 (5.5) 26.2 (21.7) 45.7 (31.1)

37.9 (28.8) 12.6 (11) 24.9 (20.4) 41.7 (36.8)

37.1 (18.7) 12.1 (8.2) 25.2 (16.3) 43.2 (25.5)

Audiovisual VE-Y VE-X VE-Z VE-R

36.7 (14.4) 15.7 (9.6) 22.8 (9.7) 43.3 (18.3)

37.3 (26.5) 13.2 (5.9) 22.2 (13.3) 40.8 (29.7)

40.8 (16.1) 13.6 (9) 21.2 (9.7) 39.2 (20.1)

38.3 (14.4) 14.1 (7.6) 22.1 (7.8) 41.1 (17.2)

Page 86: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

80

Appendix B – Block Analysis of Experiment 2 7.2

Of particular interest in Experiment 2 was the optimal integration of audiovisual cues

specifically when the limb was travelling at peak limb velocity (i.e., as measured by wrist

velocity). In order to determine whether accuracy in this task changed throughout the

experimental block, a repeated measures ANOVA was conducted contrasting across the block of

trials in all sensory conditions (i.e., auditory, visual, and audiovisual). Specifically, trials were

grouped into the first third (i.e., trials 1-20), middle third (i.e., trials 21-40), and final third (i.e.,

trials 41-60). When analyzing response accuracy across the auditory block, no significant

differences were found when judgments were made at rest, F (2, 24) = 2.6, p = .10, 𝜂!! = .12, or

following presentation at peak limb velocity, F (2, 24) = 1.4, p = .28, 𝜂!! = .10. In the visual

condition, no significant differences were found across the block at rest, F (2, 24) = .58, p = .57,

𝜂!! = .05, or at peak limb velocity, F (2, 24) = .67, p = .52, 𝜂!! = .05. Lastly, in the audiovisual

condition, no significant differences were found when at rest, F (2, 24) = .16, p = .85, 𝜂!! = .01,

or at peak limb velocity, F (2, 24) = 2.7, p = .08, 𝜂!! = .18. Overall, this pattern of results

suggests that performance in this task was not influenced by time or presentation condition.

Page 87: Multisensory Integration at Peak Limb Velocity · 2015. 12. 2. · localization when the limb is traveling at peak velocity. 1.1 Overview of the Current Thesis The overarching goal

81

Table B

Means and Between-subject Standard Deviations for Accuracy in the Temporal

Order Judgment Task Performed Following Stimuli Presentation At Rest and At

Peak Limb Velocity Across the First Third, Middle, and Final Third Number of

Trials.

Number of Correct Responses

Condition First Third

M (SD) Middle Third

M (SD) Final Third

M (SD) Total M (SD)

Auditory At Rest At PLV

14.6 (3.2) 12.4 (2.9)

14.4 (2.6) 13.4 (2.7)

12.6 (4.5) 12.9 (3.2)

41.6 (8.7) 38.7 (8.1)

Visual At Rest At PLV

11.6 (2.7) 10.8 (2.6)

11.3 (2.9) 10.8 (2.2)

10.8 (2.6) 11.8 (3.4)

33.7 (5.9) 33.4 (5.4)

Audiovisual At Rest At PLV

13.1 (2.5) 12.2 (2.8)

12.7 (3.7) 11.4 (2.2)

13.3 (3.1) 10 (3.7)

39.1 (7.3) 33.6 (6.1)