Next Article in Journal
Distinct Effects of Stimulus Repetition on Various Temporal Stages of Subject’s Own Name Processing
Previous Article in Journal
The Influence of the Ventricular-Lumbar Gradient on Cerebrospinal Fluid Analysis in Serial Samples
Previous Article in Special Issue
Pleasantness Ratings of Musical Dyads in Cochlear Implant Users
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

It Takes Two: Interpersonal Neural Synchrony Is Increased after Musical Interaction

by
Alexander Khalil
1,2,*,
Gabriella Musacchia
3 and
John Rehner Iversen
2,4
1
Department of Music, University College Cork, T23 X6Y0 Cork, Ireland
2
Institute for Neural Computation, University of California, San Diego, CA 92093, USA
3
Department of Speech Pathology and Audiology, University of the Pacific, San Francisco, CA 94103, USA
4
Swartz Center for Computational Neuroscience, University of California, San Diego, CA 92093, USA
*
Author to whom correspondence should be addressed.
Brain Sci. 2022, 12(3), 409; https://doi.org/10.3390/brainsci12030409
Submission received: 7 October 2021 / Revised: 11 March 2022 / Accepted: 13 March 2022 / Published: 20 March 2022
(This article belongs to the Special Issue Advances in Music-Related Neuroplasticity: Mechanisms and Medicine)

Abstract

:
Music’s deeply interpersonal nature suggests that music-derived neuroplasticity relates to interpersonal temporal dynamics, or synchrony. Interpersonal neural synchrony (INS) has been found to correlate with increased behavioral synchrony during social interactions and may represent mechanisms that support them. As social interactions often do not have clearly delineated boundaries, and many start and stop intermittently, we hypothesize that a neural signature of INS may be detectable following an interaction. The present study aimed to investigate this hypothesis using a pre-post paradigm, measuring interbrain phase coherence before and after a cooperative dyadic musical interaction. Ten dyads underwent synchronous electroencephalographic (EEG) recording during silent, non-interactive periods before and after a musical interaction in the form of a cooperative tapping game. Significant post-interaction increases in delta band INS were found in the post-condition and were positively correlated with the duration of the preceding interaction. These findings suggest a mechanism by which social interaction may be efficiently continued after interruption and hold the potential for measuring neuroplastic adaption in longitudinal studies. These findings also support the idea that INS during social interaction represents active mechanisms for maintaining synchrony rather than mere parallel processing of stimuli and motor activity.

1. Introduction

All interpersonal interaction requires coordination of perception and action. This coordination is observable as synchrony both at the behavioral and neural level between individuals in interaction. Synchrony of gesture, speech, breath, gaze, and gait have been found to occur spontaneously during social interaction [1,2,3] with a variety of positive effects [4]. Increased interpersonal synchrony during interaction is associated with improved communication [3,5], improved self-esteem [6], attentive behaviour [7], and increased prosocial behaviour [8,9,10]. Activities that explicitly involve interpersonal synchrony are associated with feelings of affiliation, social bonding, and a lessening of self-other distinction [11,12,13]. Decreased interpersonal synchrony is associated with moments of disagreement [14] and decreased ability to synchronize as seen in a number of disorders including Schizophrenia [15], Autism Spectrum Disorder [16], and Attention Deficit Hyperactivity Disorder [17].
Interpersonal synchrony is an explicit component and core feature of music and extends beyond playing music or singing to include many related activities such as dancing, marching, rehearsing, and listening. Multiple studies have investigated the significance and dynamics of musical synchrony. Shared movement and visual cues during music coordinate interaction and improve accuracy [18,19,20]. Musical synchrony increases a sense of shared intentionality and decreases the experience of self-other distinction [21,22,23,24], and can relate to a sense of communal identity [25,26,27]. Both a form of communication and joint action, music is intrinsically interpersonal [28,29,30]. Music is known to be related to neuroplasticity throughout life [31,32,33,34]. Neuroplasticity achieved through music, then, must be driven to some extent at the interpersonal level. The development of interpersonal measures of change at the neural level following musical interactions may identify the dynamics that drive neuroplasticity.
Studies relating the brain activity of two or more individuals (a technique that is often referred to as “hyperscanning”, for reviews see [35,36,37,38,39,40,41]) have found that the temporal dynamics of neural activity between people are more related during an interaction, a phenomenon called interpersonal neural synchrony (INS). Increased INS predicts dynamics of interaction [42], has been observed during face-to-face interactions [43,44,45] and is strongly associated with affiliation, cooperation, prosociality, and shared intention [46,47,48,49,50,51,52,53]. INS levels have also been found to predict cooperative success [54,55]. INS is increased during interactions between romantic partners [56], and induced by interpersonal touch [57]. Further highlighting connections between INS and relationships, nasal administration of oxytocin—a hormone associated with feelings of interpersonal connection—increased INS during interaction [58].
A significant number of INS studies have focused on music [38]. Due to its sustained and hierarchical temporal structure, emotional salience, and sustained synchrony, musical interaction has been posited as an ideal paradigm for interpersonal neuroscience [59]. Increased INS has been found between dyads singing and humming together, even when they were unable to see each other [60]. Increased INS between a violinist and audience members predicts the level of audience appreciation [61]. Suggesting a causal relationship, direct manipulation of INS using simultaneous, synchronized transcranial Alternating Current Stimulation (tACS) showed increased levels of behavioral synchrony in teacher/learner dyads and led to improved song learning outcomes [62].
EEG is particularly suited to music studies of interbrain synchrony due to its high temporal resolution, affording not only the ability to measure oscillatory activity at fine timescales but also precisely time-locked responses to musical events [59]. EEG studies investigating spectral changes in relation to musical events found that intrabrain alpha desynchronization musicians occur in parallel at points of reported increased emotion during performance [63,64] and that intrabrain event-related desynchronization (ERD) occurs in parallel between musicians when they intentionally coordinate with each other [65]. A series of EEG studies investigated interbrain coherence in dyads playing guitar, finding increased coherence in delta and theta band activity between participants [66,67,68]. This work, all conducted by the same team, posits that low-frequency interbrain coherence represents or is part of a mechanism that enables interpersonal temporal coordination [68]. These findings have also been replicated and extended by other researchers investigating interbrain delta and theta coherence in other types of interactions [46,69,70].
Such studies are corroborated by work that suggests brain activity in delta and theta band comprises an interface between exogenously driven and endogenously generated oscillations. A seminal study in this area showed that the rhythm of entrained oscillations is under endogenous attentional control in macaques, and can be switched to follow a specific rhythm in a set of competing rhythms [71]. In a human analogue, perceptual accuracy increases and auditory processing is more efficient when people attend to sounds that occur in a regular delta pattern, relative to an irregular one [72]. Theta-band oscillations in the auditory cortex can entrain the syllabic rate of speech [73,74,75] and help distinguish speech in background noise [76]. While more recent work in individuals has revealed a top-down dynamic to entrainment, in which persistent oscillations are sustained only at beta or gamma frequency bands, potentially driving the observed changes into lower frequency bands [77,78], interpersonal neural coherence during interaction has only been observed in lower frequencies.
Further highlighting the relevance of music and low-frequency activity is the fact that the hierarchical structure of music and brain rhythm is quite similar [79]. Both emphasize the delta band (0.5–4 Hz/30–240 bpm), as spontaneous delta band activity occurs at higher amplitudes than other frequency bands in the brain, and the beat—a key structure to which people synchronize in music—occurs almost entirely in the delta band.
Behavioral synchrony in music and rhythmic activities has been shown to continue briefly after the loss of sensory contact between people and it has been proposed that this phenomenon supports human cooperation or joint action which naturally features brief and intermittent loss of sensory contact [80]. Whether INS also persists beyond the interaction itself has remained unknown. INS beyond the period of interaction would support the idea that INS is perhaps an integral part of the mechanism by which people achieve, maintain, or recover behavioral synchrony and so represents more than parallel processing of stimuli or parallel motor activity.
The present study aimed to address this question by testing whether INS would be increased in spontaneous EEG recorded in silence after a musical interaction. A pre-post paradigm was developed in which participants grouped in dyad pairs, played a musical tapping game. Spontaneous EEG was synchronously recorded in silent, non-interactive periods immediately preceding and immediately following the musical interaction. Through comparison of identical silent, non-interactive periods that only differ by whether they precede or follow a musical interaction, we intended to avoid much of the confound of induced synchrony from simultaneous stimulation and activity.
We hypothesized that INS would be increased in the post-condition. We further hypothesized, given previous findings, that persistent effects would be found in the delta or theta band as increased wavelet power, interbrain phase coherence, or both.

2. Materials and Methods

2.1. Participants and Setup

Ten (10) participants were recruited through word of mouth at the department of Audiology at the University of the Pacific. Six graduate students, three faculty, and one child participated, forming ten (10) dyads, each participant undergoing the experiment twice, each time with a different, randomly-selected partner. All participants had normal hearing thresholds (>20 dB for 0.5, 1, 2, and 4 kHz) and no history of neurological disorders. Participants were between 8 and 50 years old with 3 males and 7 females. Participants were familiar with each other to varying degrees and while familiarity is known to modulate INS [81], we did not consider this to be problematic when solely establishing an effect in a pre/post paradigm. Further, not all participants were naïve to the purpose of the study. It would have been difficult for this knowledge to affect the brain-based pre/post paradigm. Participants had highly varied and diverse musical backgrounds and levels of experience. Figure 1 shows the equipment setup and participant placement. Both individuals in the dyad were seated in a sound isolated chamber facing each other. Between them, at lap height, was placed an Alesis SamplePad 4 MIDI drum pad, connected to a JBL 308P 8” studio monitor under a table directly beneath it. The drum pad was positioned so that each participant could easily reach it and tap on it by leaning forward and partly extending an arm. Participants were asked to tap with their dominant hand (all participants were right-handed). Another JBL studio monitor, connected to a computer, was placed 3′ away from the participants on a 4′ stand (Figure 1).
Each participant wore a low density, 4000 Hz sampling rate dry electrode headset made by Cognionics [82]. The design used for the present study featured 5 electrode placements: Fp1, Fp2, Cz, O1, and O2 with a forehead ground and left mastoid reference. As these headsets are not adjustable, the locations of all sites are approximate: head circumference and height will affect specific locations on each participant. Although all participant head circumferences fell within the manufacturer specified range of 52–62 cm, a custom, specially-made smaller version of the headset was used for participants with smaller heads, to align at the vertex. Two types of electrodes were used: smooth and tined. The smooth electrodes feature a soft, pliable plastic skin and are made for sitting against the bare skin of the forehead. The tined electrodes feature flexible plastic tines arranged in a star-like configuration that comb through hair to reach the scalp (Figure 2). These dry electrodes do not need any gel or any skin preparation other than swabbing bare skin with alcohol swabs. These headsets also feature two analog inputs that are recorded in sync with an EEG. These analog channels allow synchronization of EEG data with external stimuli and thus enable synchronization of recordings between headsets. For the present study, square pulses were sent from the experimental computer to both headsets. Aligning these square pulse “triggers” in the data allowed data from different headsets to be synchronized to within one sample (0.25 ms).

2.2. Task

The present study employs a cooperative musical activity: an anti-phase tapping game. The object of the game is to maintain anti-phase synchrony while tapping at progressively faster tempi. Anti-phase synchrony was chosen over in-phase synchrony for this study for several reasons. Trying to maintain anti-phase synchrony is a more cooperative and musical activity than in-phase, or unison, synchrony. Whether participants succeed at achieving and maintaining it, anti-phase tapping requires them to hold in mind the composite sound and focus on it rather than only their own tapping. Tapping in anti-phase synchrony has also been found to involve more significant changes in brain dynamics compared with in-phase synchronized tapping, placing significantly more demands on motor coordination than in-phase tapping [83,84,85]. Further, tapping in anti-phase synchrony also closely resembles the type of interpersonal attention and exchange required when playing music collaboratively, yet is simple and does not require pre-existing knowledge of specific musical material or performance techniques. Participants were not expected to successfully maintain anti-phase synchrony throughout their trials. Rather, attempting to do so in a game-like context was intended to be an engaging and musical form of joint action.
An induction sequence of alternating “high” (329.6 Hz) and “low” (220 Hz) guide tones, corresponding to the musical notes “E4” and “A3”, respectively, were played through a speaker connected to a computer controlled by a researcher outside the booth. Participants were assigned at random to tap with either the higher or lower guide tone and each participant’s MIDI drum pad triggered a high or low drum sound (Garritan Timpani KS patch) corresponding to their assignment of a higher or lower guide tone. The guide tones were played for 8 s and participants were instructed to tap with their target pitch and then to maintain the composite interlocking rhythm for as long as possible after the guide tones had stopped, constituting a simple, collaborative musical game (Figure 3). In order to maximally challenge each dyad, the guide tone tempo increased across trials, and participants were instructed to move up to the most challenging pace they could perform in a sustainable manner. The initial tempo had an Interstimulus Interval (ISI) of 800 ms, decreasing 100 ms each trial until an ISI of 200 ms was reached and then further decreasing by 25 ms to 175 ms and 150 ms. Since each participant was tapping to every other stimulus, the fastest any participant was required to tap was at an interval of 300 ms. In the case that a level was so easy as to be indefinitely sustainable, participants had the option to stop when they wished and move on to the next, more challenging one. Participants were not required to complete all 9 levels but rather, were instructed to collaboratively “find the most challenging, yet sustainable tempo.” They were instructed that they were free to repeat levels they enjoyed or wanted to practice. When the dyad reached an unsustainable tempo, the game would be considered complete. The time each dyad spent tapping varied from 4.6 to 9.9 min with a mean of 6.43 and standard deviation of 1.93 min.

2.3. Procedure

Participants received instructions while EEG headsets were applied and impedances checked. All impedances were brought below 300 kΩ and signal quality was visually checked and monitored throughout the recordings. The instructions did not include practice trials since doing so could possibly induce the brain dynamics that the post-task EEG was intended to measure. Rather, participants received verbal instructions and the first few trials were set to a pace that would facilitate learning and familiarity with the task. The EEG recording was signaled with a chime: participants would sit in silence, turn 90 degrees to their left (leaning slightly forward so that they could not see each other) for a period of 2 min. Participants were observed through a window to verify that they were in compliance with these conditions. When this was complete, a second chime would cue them to face each other and begin the task (described above). Following this, the participants would begin the game. When the game was complete, signaled by participants by again turning 90 degrees to their left, such that they could not see each other, a second EEG recording would take place as they sat in silence with their eyes open for 2 min. A final chime would indicate that the study was complete (Figure 3). While an EEG was recorded throughout the game, here we focus on the spontaneous EEG prior to and following the tapping game, to examine if any changes are present between these two conditions.

3. Data Analysis

3.1. EEG Processing

EEG analysis included three pre-processing steps: (1) temporal alignment of the two EEG files from a participant dyad using recorded triggers, (2) data filtering to remove low-frequency drift and dc offset and emphasize cortical activity, and (3) selection of artifact-free sections of data in the pre- and post-activity sections from each participant. To align the dyads’ files in real-time, time zero was adjusted to the initial trigger (T0) that was simultaneously sent to each recording at the beginning of each experiment. Continuous data from both participants in a dyad were overlaid to verify that T0 and subsequent triggers were aligned across the entire recording session. Data files were then filtered with a Finite Impulse Response (FIR) band-pass from 0.5 to 60 Hz. Finally, overlapping sections of EEG data free from eyeblink and EMG (i.e., clean in both participants at the same time in the recording) were identified and the first 15 s doubly-clean (artifact-free in both participants) segment was selected for analysis from pre- (PRE) and post-activity (POST) sections. The 15 s duration was selected in order to systematize section durations and maximize available data for spectral analysis as concomitantly clean durations of EEG beyond 15 s were not regularly observed across dyad data and several dyads did not reach 15 s. This resulted in a mean duration of 14.77 s (SD = 0.48) in the PRE condition and a mean of 14.78 s (SD = 1.31) in the POST condition. Clean sections began on average about one minute into the PRE section (M = 0.97 min, SD = 0.59 min), and on average were within the first 30 s following the end of tapping in the POST section (M = 0.45 min, SD = 0.56 min). There were no significant differences between clean section duration (t(9) = 0.20, p = 0.984) or time of occurrence (t(9) = 1.983, p = 0.079) across conditions.
While Independent Component Analysis (ICA) would seem an effective method for cleaning data, the mobile EEG headsets used in the present study had only five channels, a number we deemed insufficient for reliable ICA results. While an obvious disadvantage for ICA, the choice to use such low-density headsets is directed towards scaling up numbers of participants in future studies.
A combination of wavelet analysis and cross-correlation was used to obtain measures of INS. Wavelet analysis, such as Fourier decomposition, expands a function in terms of scaling properties that are localized in both time and frequency [86]. Here, we used a Morlet wavelet function [87], similar to a modulated Gaussian waveform, with a mother wavelet size of six cycles (ῳ0 = 6) in frequency bands between 1 and 100 Hz. Power and instantaneous phase were then extracted by taking the absolute and angle values of the wavelet output for each time point and frequency.
Wavelet power was averaged within two frequency bands: delta (defined here as 0.5–3 Hz to avoid overlap) and theta (4–8 Hz). To determine the degree of INS in a dyad, we cross-correlated the instantaneous phase vector for each frequency bin within delta and theta bands. For each frequency bin, the maximum cross-correlation r within a window of −5 and 5 ms lag around 0 ms was extracted in order to assess nearly-simultaneous inter-brain phase correlation, rather than inter-brain phase correlation that was offset by some time factor. By taking the maximum r-value over this time period, we also avoided wrap-around phase effects and antiphase correlations. While one might expect antiphase rhythms would be induced by an antiphase game, we focus here on the composite rhythm as attention to it is key to a sustained interaction. The maximum correlation across the frequency bins within each band was then selected. The result of this computation was a matrix of values that described how well the phase of neuronal oscillations in the delta and theta frequencies correlated between dyad participants in PRE and POST conditions.

3.2. Statistical Analysis

The two dependent variables in this dataset were phase correlation (r) and power (pwr). The predictor was condition (PRE, POST) and data were collected from a sample of ten dyads. Brain response measurements were taken from five channel locations (Fp1, Fp2, Cz, O1, O2) in two frequency bands (delta, theta). In the case of phase correlation, we have one measurement for each dyad (i.e., the correlation of phase between the two subjects), giving 200 observations and a balanced design. We used a linear mixed-effects model to determine statistical effects. Dyad was specified as the subject variable. Condition, channel, and frequency band were specified as repeated variables. The subject variable was modeled as a random effect and the repeated measures variables were modeled as fixed effects.
For power, we had two measurements per dyad (i.e., the power for each subject) and the same repeated variables, giving 400 observations and a balanced design. We again used a linear mixed-effects model to determine effects. Dyad and subject were specified as the subject variables. As in the phase correlation analysis, condition, channel, and frequency band were specified as repeated variables. Again, the subject variables were modeled random effects and the repeated measures variables were modeled as fixed effects.
In both analyses, Type III tests showed which fixed effects were related to the dependent variable of brain response measurements. Post-hoc paired sample t-tests were conducted to clarify differences when significant main effects of condition (i.e., PRE vs. POST) were observed.
The total duration of tapping for each dyad across all trials was in addition correlated with power magnitude and phase correlation values in delta and theta bands to determine if the strength of INS varied with the length of time playing together.

4. Results

4.1. Wavelet Power

To determine whether power differed in the PRE vs. POST condition, a mixed-effects model was conducted with one dependent variable (power), 20 levels of repeated effects (two conditions*five channels*two frequency bands), and 12 levels of random effects (two subjects + ten dyads). Results showed significant main effects for the channel (F(4/70.2) = 28.188, p < 0.001) and frequency band (F(1/138.59) = 57.22, p < 0.001), but no effect of condition (F(1/138.59) = 0.820, p = 0.367) and no interaction between the three independent variables (F(4/138.59) = 0.432, p = 0.785). Examination of the data showed the expected topographical differences in the distribution of power across the scalp, which was greatest in frontal electrodes. In sum, wavelet power did not significantly change from PRE to POST conditions, nor were there any channel-specific interactions with the condition. Figure 4 shows the grand average wavelet power across frequency for all channels. Delta (δ) and theta (θ) frequency ranges are shown with horizontal lines. Frontal channels do appear to have a significant reduction in PRE to POST power around 4Hz at the lower theta range.

4.2. Interpersonal Neural Synchrony—Frequency Specific Phase Coupling

INS was measured as the maximal inter-individual phase correlation within the delta and theta bands. To determine whether phase correlation differed in the PRE vs. POST condition, a mixed effects model was conducted with one dependent variable (r), 20 levels of repeated effects (two conditions*five channels*two frequency bands) and 10 levels of random effects (10 dyads). Results showed significant main effects for channel (F(4/59.28) = 10.906, p < 0.001), frequency band (F(1/107.77) = 297.347, p < 0.001) and condition (F(1/107.77) = 8.497, p = 0.004) as well as interaction between the three independent variables (F(4/59.28) = 3.358, p = 0.015).
Post-hoc, paired samples t-test analysis showed that phase correlation in the delta band was greater in the POST condition, compared to PRE, at midline (Cz) (t(9) = 3.130, p = 0.012) and right frontal (Fp2) (t(9) = 2.289, p = 0.048) channels. Taken together, this means that delta frequency INS is greater between dyad participants after musical interaction, and especially at electrodes over the central and right frontal brain areas. Figure 5A shows the mean and standard error of the phase correlation values in the delta band, illustrating the findings. No effect of condition (i.e., PRE vs. POST) was observed in the theta band data, suggesting that theta-band INS is similar before and after musical interaction.

4.3. Relationships between Interpersonal Neural Synchrony and Duration of Musical Interaction

To determine putative brain-behavior relationships, Pearson’s correlations were computed between the duration of musical interaction and delta-mediated INS in both the PRE and POST conditions. Results (Figure 5B) showed a significant correlation between delta-mediated INS and tapping duration at Cz (r = 0.879, p = 0.001) only for the POST condition, such that longer tapping duration was associated with greater post-tapping INS, but did not relate to pre-tapping INS. There was a similar trend for frontal electrodes though it did not reach significance. This result suggests a positive relationship between the behavior of musical interaction and the degree of interpersonal brain coherence after the behavior. The lack of significant behavioral correlations with PRE-tapping INS suggests that pre-interaction INS does not relate to the duration of the interaction, and argues against the POST finding being due to simple systematic bias. No significant correlations were observed between theta-band INS and musical interaction duration.

4.4. Tapping Characterization

Each participant’s taps were recorded by three-axis accelerometers and analog audio channels on each headset. The high level of freedom in the musical interaction resulted in significant variability in terms of the number of trials, tempi achieved, and stability of anti-phase synchrony. Relative phase calculations for each dyad show that all dyads were able to maintain anti-phase synchrony as the tempo increased across trials, with relative phase variability increasing in some dyads towards the final trials, which featured higher tempi (Figure 6). From this, it can be inferred that in no case were PRE/POST changes in INS the result of anti-phase tapping collapsing to in-phase tapping, a phenomenon that could be expected to occur frequently when tempi become challenging [88].

5. Discussion

This study explored the possibility of persistent oscillatory changes in interpersonal brain activity following a dyadic musical interaction. The main finding, a significant increase in delta band INS after a musical interaction, is novel to the authors’ knowledge and holds significant implications for the investigation of music-related neuroplasticity at the interindividual level. Further, that the effect was seen only in the Delta band may highlight the importance of the musical beat, which corresponds closely to this frequency band in practically all cultures [89]. This result complements the work of Nguyen et al., which found that the synchrony of cortical hemodynamic activation between mother-child subject pairs appeared to continue into rest periods between cooperative interaction [54]. While this could be interpreted as a post-interaction continuation of INS, the study design did not include a pre-task resting period and so the degree to which INS during between-task resting periods may differ from baseline mother-child INS was left unknown [54]. This result also complements the work of Gugnowska et al., who found amplitude-based INS increased in gamma band between two pianists during pauses in play [90]. Future studies will be needed to investigate the relationship between the temporal dynamics of these amplitude changes and the lower frequency phase coupling investigated in the present paper. The findings of the present study also complement past work, which focused on findings of increased INS during an interaction. A persistent issue with INS measured during interaction is that it may simply result because individuals in interaction are receiving simultaneous multisensory input and also engaging motor pathways in time with the simultaneous behavioral interaction [91,92,93,94]. While it is difficult to disambiguate more mundane or trivial forms of INS during interaction in the form of parallel processing from potential mechanisms of maintaining synchrony, the present finding of increased INS during a POST interaction period suggests the latter. We provide some preliminary interpretations of this surprising finding below, focusing on neuroplasticity behavioral significance, and neural dynamics, with suggestions for follow-on research to more fully understand this post-interaction INS phenomenon.
Li et al. [51] found that basketball players—whose sport requires very close synchrony between players—achieved INS more quickly and at higher levels than age-matched controls during a cooperative drawing task, which they also completed more quickly than controls. This suggests a long-term cumulative effect stemming from neuroplasticity: people who frequently engage in joint action, such as music or team sports, may adapt for those activities at the neural level, possibly becoming more efficient at them. In the present study, INS was found increased relating to the length of time each dyad played together, suggesting that there may also be a more immediate cumulative effect. At the same time, it is possible that dyads that performed well, or perhaps were just more comfortable with the game due to greater capacity to synchronize, naturally played for longer durations, accounting for this effect. In future studies, we intend to investigate these issues as part of an ongoing longitudinal study, facilitating tracking effect size with the magnitude of observed INS.
Koban et al. [95] propose that behavioral and neural synchrony can be explained from the perspective of optimization theory: being in synchrony allows conservation of computational resources and also causes alignment—and overlap—in self-other representations. Music, then, could facilitate further optimization as it comprises a temporal filter for behavior, organizing interaction into easily predictable rhythms. In line with this, we propose a corollary that it may also be more efficient to maintain these rhythms after an interaction in order, for example, to restart the interaction more easily. Supporting this, many interactions do not have well-defined endings. For example, a recent study found that 70% of adults were not able to determine when a conversation had ended [96]. Therefore a potential benefit exists for maintaining some kind of latent representation of the interaction dynamics (e.g., the rhythms of INS) until some new stimulus or situation occurs. Our findings would be consistent with this idea.
Beyond optimizing energetic and computational resources while maintaining a sense of affiliation, there would be a strong goal-oriented benefit in maintaining the rhythms of interaction. Many group behaviors from hunting to basketball feature moments in which perceptual contact between group members is lost [80]. Maintaining the group rhythm (a form of covert synchrony) across these moments would be crucial for success. This phenomenon has been observed at the behavioral level [80]. We propose that the persistent neural synchrony observed in the present study may represent a neural correlate, an interbrain dynamic that is related to this ability.
Perhaps the most well-known—and highly-debated—potential mechanism related to our findings would be neural entrainment. Sustained phasic oscillations persisting well after periodic stimulation has ceased would be compelling evidence for the classic entrainment models posited by theories such as the Dynamic Attending Theory (DAT) [97]. However, even at the individual level, the persistence of endogenous oscillations phase-locked to a low-frequency periodic stimulus has so far been reported for only a few cycles [98,99,100]. In this light, the present finding of what could be interpreted as evidence of entrained low-frequency rhythm persisting many seconds after an interaction between two brains is startling indeed. That said, some studies have found direct evidence for phase-locked entrainment at higher frequencies [77,101], supporting an alternative “top-down” phase reset model which could help account for such a finding [78].
Further complicating simple entrainment by phase reset interpretation would be that, as described above, virtually every trial of the anti-phase tapping game ended in the collapse of the anti-phase rhythm the participants were attempting to maintain. In each trial, participants’ tapping would fall out of synchrony or occasionally collapse briefly into in-phase synchrony. If the observed increases in INS represent continued synchronous oscillations resulting from a phase reset process, then the observed INS would more likely be related not to the periods of anti-phase tapping synchrony that invariably occurred earlier in each trial but rather to the chaotic or in-phase tapping synchrony that occurred at the end of each trial. The question of whether anti-phase tapping synchrony or in-phase tapping synchrony would be causally related to the observed effect can only be addressed through future studies that directly compare pre/post interpersonal phase coherence between anti-phase and in-phase tapping. Further, now that the existence of increased post-interaction INS has been established in a musical context, pre/post INS must be compared between tapping and control activities featuring other forms of interaction, such as conversation.
An alternative possibility to passive entrainment caused solely by phase reset of neural firing patterns at low frequencies would be an active attention/cooperation-driven form of entrainment. Participants might covertly or subconsciously replay their experience. This would achieve all the behavioral level benefits described above: optimization of energetic and computational resources, the scaffolding of potential continuance of interaction, and be evident seconds after discontinuation of activity. Such “covert replay” could result in the observed effect. Indeed, covert motor simulation has been posited repeatedly as an important behavioral and perceptual scaffold (reviewed in [102,103]) and it is known that internal rehearsal or imagery can impact brain oscillations [104]. Further supporting this idea, INS has been linked to explicit action plans during pauses in joint activity [90].
These ideas are eminently testable and invite future research. For example, the hypothesis that post-interaction INS relates to some advantage for restarting the interaction could be directly tested by examining how effectively interaction behaviors restart after a POST break. We would predict that dyads with higher POST INS would more effectively re-start the interaction. It would also be possible, using a simple interview format, to determine which participants experienced or engaged in mental imagery or active rehearsal of dynamics during the break. We would predict that dyads who were aware of such processes would more effectively re-start the interaction.
The important brain-behavior finding, that duration of interaction was significantly correlated with INS, could index the inherent ability of one or both of the participants. Better synchronizers would likely have a larger effect and be able to synchronize for more sustained periods. A very good synchronizer might also be able to adapt rhythmically to a less able one and so may still be able to carry the interaction for long periods, possibly also driving INS. We plan in future studies to characterize or compare the musical interaction (e.g., Did participants feel successful? Were they able to maintain synchrony at a high level before giving up? Did they enjoy the interaction?) and how such characterizations relate to observed effect sizes.
The present study did not compare brain dynamics during interaction with pre- or post- conditions. This is because during interaction overt movement (and possible movement artifacts) are synchronized, potentially inflating the measured INS and making the comparison difficult. In future work, we plan to develop methods that would facilitate such comparison. It may be that the antiphase condition will mitigate this to some extent since behavior will be out of phase, although the dyad still receives the same auditory inputs.
While the participants were somewhat familiar with each other, and it is known that familiarity modulates INS [81], this familiarity cannot explain the observed effect in the POST condition: the key variable between PRE and POST was the tapping game. However, familiarity and prosociality will need to be investigated in future work in this context.
In closing, we wish to emphasize that the observed post-interaction changes in brain dynamics only become apparent by taking the dyad as the fundamental unit of analysis. It may be that post-interaction EEG changes in stereotypical ways, making it nearly impossible to detect sustained changes in a single brain. While further study will clearly be needed to investigate this phenomenon, it seems apparent that an advantageous and perhaps necessary approach would involve multi-person EEG recording. It may, in fact, take two.

Author Contributions

Conceptualization, A.K. and G.M.; methodology, A.K. and G.M.; software, A.K. and G.M.; validation, G.M. and J.R.I.; formal analysis, G.M.; data curation, G.M.; writing—original draft preparation, A.K. and G.M.; writing—review and editing, J.R.I.; visualization, G.M., A.K. and J.R.I.; project administration, G.M. and J.R.I.; funding acquisition, A.K., G.M. and J.R.I. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Science Foundation, grant number SMA 1540943 and Enterprise Ireland, grant number CS20192102.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Institutional Review Boards at UoP and UCSD, approval number 151204.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data used for this study can be found in Appendix B in Table A2 and Table A3.

Acknowledgments

We would like to thank the rest of the NSF Group Brian Dynamics in Learning Network team: Tzyy-Ping Jung; Ying Wu; and Layne Kalbfleisch for contributions to the development of the technology and methods used for this work. We would also like to thank Natalie Sienko, Patricia Tan, Christopher and Darlene Hondros, and Jayaganesh Swaminathan for their assistance in carrying out this research.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Marginal Means and Standard Errors of Wavelet Power in PRE and POST conditions.
Table A1. Marginal Means and Standard Errors of Wavelet Power in PRE and POST conditions.
ConditionChannelsDelta Power (SE)Theta Power (SE)
PRE (before rhythm activity)Fp114.3 (1.9)7.1 (1.0)
Fp213.3 (1.9)6.9 (1.0)
Cz5.7 (0.7)4.5 (0.5)
O17.9 (2.2)4.9 (0.8)
O210.4 (3.4)5.3 (0.8)
POST (after rhythm activity)Fp117.6 (2.5)7.0 (0.9)
Fp217.4 (2.4)6.9 (0.9)
Cz6.9 (0.9)4.5 (0.5)
O18.5 (3.5)4.6 (1.1)
O29.0 (2.4)4.6 (0.8)

Appendix B

Individual Dyad Inter-Brain Correlations Underlying Figure 4.
Table A2. Dyad INS in delta band.
Table A2. Dyad INS in delta band.
Maximum Inter-Brain Correlation, Delta Band
Condition        DyadR Deltamax
Fp1
R Deltamax
Fp2
R Deltamax CzR Deltamax O1R Deltamax
O2
PRE10.190.170.190.130.15
20.240.250.160.230.18
30.220.210.150.200.27
40.200.180.270.250.16
50.440.270.100.160.20
60.230.150.170.180.17
70.300.280.190.170.23
80.280.240.260.230.26
90.220.230.200.190.23
100.300.240.210.230.13
Mean0.26340.22250.18880.19740.1998
Std. Error of Mean0.023270.013480.015960.011690.01460
POST10.370.290.260.290.19
20.510.450.280.240.15
30.260.280.210.230.22
40.310.300.170.140.13
50.230.210.220.240.21
60.360.280.260.240.22
70.270.190.290.210.23
80.350.360.310.180.21
90.300.290.260.200.17
100.240.230.200.140.20
Mean0.31970.28830.24460.21030.1927
Std. Error of Mean0.025950.023520.014260.015100.01050
Table A3. Dyad INS summaries data for theta band.
Table A3. Dyad INS summaries data for theta band.
Maximum Inter-Brain Correlation, Theta Band
Condition        DyadR Thetamax Fp1R Thetamax
Fp2
R Thetamax CzR Thetamax O1R Thetamax O2
PRE10.160.140.150.110.08
20.160.130.110.080.08
30.190.170.140.090.08
40.200.200.160.120.12
50.140.100.110.090.05
60.100.120.120.060.07
70.140.080.130.100.08
80.100.110.090.150.15
90.150.170.140.100.10
100.130.120.090.080.14
Mean0.14670.13490.12560.09850.0955
Std. Error of Mean0.010440.011500.007590.007750.01020
POST10.160.130.130.100.07
20.150.190.120.070.09
30.150.120.160.180.10
40.060.080.090.110.12
50.140.160.190.120.14
60.100.090.080.090.18
70.130.080.100.120.10
80.060.070.130.150.08
90.170.180.140.150.17
100.130.110.110.090.13
Mean0.12560.12080.12450.11870.1171
Std. Error of Mean0.012050.013620.010380.010400.01211
Table A4. Dyad-specific duration of the interaction.
Table A4. Dyad-specific duration of the interaction.
DyadDuration (Seconds)
1594
2534
3486
4378
5366
6318
7312
8300
9294
10276

References

  1. Delaherche, E.; Boucenna, S.; Karp, K.; Michelet, S.; Achard, C.; Chetouani, M. Social Coordination Assessment: Distinguishing between Shape and Timing. In Proceedings of the IAPR Workshop on Multimodal Pattern Recognition of Social Signals in Human-Computer Interaction; Springer: Berlin/Heidelberg, Germany, 2012; pp. 9–18. [Google Scholar]
  2. Kinsbourne, M.; Helt, M. Entrainment, Mimicry, and Interpersonal Synchrony. In The Neuropsychology of Autism; Oxford University Press: Oxford, UK, 2011; pp. 339–365. [Google Scholar]
  3. Chetouani, M.; Delaherche, E.; Dumas, G.; Cohen, D. 15 Interpersonal Synchrony: From Social Perception to Social Interaction. In Social Signal Processing; Cambridge University Press: Cambridge, UK, 2017; p. 202. [Google Scholar]
  4. Mogan, R.; Fischer, R.; Bulbulia, J.A. To Be in Synchrony or Not? A Meta-Analysis of Synchrony’s Effects on Behavior, Perception, Cognition and Affect. J. Exp. Soc. Psychol. 2017, 72, 13–20. [Google Scholar] [CrossRef]
  5. Beňuš, Š. Social Aspects of Entrainment in Spoken Interaction. Cogn. Comput. 2014, 6, 802–813. [Google Scholar] [CrossRef]
  6. Lumsden, J.; Miles, L.K.; Macrae, C.N. Sync or Sink? Interpersonal Synchrony Impacts Self-Esteem. Front. Psychol. 2014, 5, 1064. [Google Scholar] [CrossRef]
  7. Khalil, A.K.; Minces, V.; McLoughlin, G.; Chiba, A. Group Rhythmic Synchrony and Attention in Children. Front. Psychol. 2013, 4, 564. [Google Scholar] [CrossRef] [Green Version]
  8. Cirelli, L.K.; Einarson, K.M.; Trainor, L.J. Interpersonal Synchrony Increases Prosocial Behavior in Infants. Dev. Sci. 2014, 17, 1003–1011. [Google Scholar] [CrossRef] [PubMed]
  9. Rennung, M.; Göritz, A.S. Prosocial Consequences of Interpersonal Synchrony. Z. Psychol. 2016, 224, 168–189. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Tunçgenç, B.; Cohen, E. Interpersonal Movement Synchrony Facilitates Pro-Social Behavior in Children’s Peer-Play. Dev. Sci. 2018, 21, e12505. [Google Scholar] [CrossRef] [PubMed]
  11. Cross, L.; Turgeon, M.; Atherton, G. How Moving Together Binds Us Together: The Social Consequences of Interpersonal Entrainment and Group Processes. Open Psychol. 2019, 1, 273–302. [Google Scholar] [CrossRef]
  12. Fessler, D.M.; Holbrook, C. Synchronized Behavior Increases Assessments of the Formidability and Cohesion of Coalitions. Evol. Hum. Behav. 2016, 37, 502–509. [Google Scholar] [CrossRef] [Green Version]
  13. Hove, M.J.; Risen, J.L. It’s All in the Timing: Interpersonal Synchrony Increases Affiliation. Soc. Cogn. 2009, 27, 949–960. [Google Scholar] [CrossRef] [Green Version]
  14. Paxton, A.; Dale, R. Argument Disrupts Interpersonal Synchrony; SAGE Publications Sage UK: London, UK, 2013. [Google Scholar]
  15. Kupper, Z.; Ramseyer, F.; Hoffmann, H.; Tschacher, W. Nonverbal Synchrony in Social Interactions of Patients with Schizophrenia Indicates Socio-Communicative Deficits. PLoS ONE 2015, 10, e0145882. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Kaur, M.; Srinivasan, S.M.; Bhat, A.N. Comparing Motor Performance, Praxis, Coordination, and Interpersonal Synchrony between Children with and without Autism Spectrum Disorder (ASD). Res. Dev. Disabil. 2018, 72, 79–95. [Google Scholar] [CrossRef] [PubMed]
  17. Toplak, M.E.; Dockstader, C.; Tannock, R. Temporal Information Processing in ADHD: Findings to Date and New Methods. J. Neurosci. Methods 2006, 151, 15–29. [Google Scholar] [CrossRef] [PubMed]
  18. Chang, A.; Livingstone, S.R.; Bosnyak, D.J.; Trainor, L.J. Body Sway Reflects Leadership in Joint Music Performance. Proc. Natl. Acad. Sci. USA 2017, 114, E4134–E4141. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Colley, I.; Varlet, M.; MacRitchie, J.; Keller, P.E. The Influence of a Conductor and Co-Performer on Auditory-Motor Synchronisation, Temporal Prediction, and Ancillary Entrainment in a Musical Drumming Task. Hum. Mov. Sci. 2020, 72, 102653. [Google Scholar] [CrossRef] [PubMed]
  20. Eerola, T.; Jakubowski, K.; Moran, N.; Keller, P.E.; Clayton, M. Shared Periodic Performer Movements Coordinate Interactions in Duo Improvisations. R. Soc. Open Sci. 2018, 5, 171520. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Ellamil, M.; Berson, J.; Wong, J.; Buckley, L.; Margulies, D.S. One in the Dance: Musical Correlates of Group Synchrony in a Real-World Club Environment. PLoS ONE 2016, 11, e0164783. [Google Scholar] [CrossRef]
  22. Himberg, T.; Laroche, J.; Bigé, R.; Buchkowski, M.; Bachrach, A. Coordinated Interpersonal Behaviour in Collective Dance Improvisation: The Aesthetics of Kinaesthetic Togetherness. Behav. Sci. 2018, 8, 23. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Keller, P. Ensemble Performance: Interpersonal Alignment of Musical Expression; Oxford University Press: Oxford, UK, 2014. [Google Scholar]
  24. Reddish, P.; Bulbulia, J.; Fischer, R. Does Synchrony Promote Generalized Prosociality? Relig. Brain Behav. 2014, 4, 3–19. [Google Scholar] [CrossRef]
  25. Lucas, G.; Clayton, M.; Leante, L. Inter-Group Entrainment in Afro-Brazilian Congado Ritual. Empir. Musicol. Rev. 2011, 6, 75–102. [Google Scholar] [CrossRef] [Green Version]
  26. Turino, T. Music as Social Life: The Politics of Participation; University of Chicago Press: Chicago, IL, USA, 2008. [Google Scholar]
  27. Witek, M.A. Filling in: Syncopation, Pleasure and Distributed Embodiment in Groove. Music. Anal. 2017, 36, 138–160. [Google Scholar] [CrossRef]
  28. Higgins, K.M. The Music between Us; University of Chicago Press: Chicago, IL, USA, 2012. [Google Scholar]
  29. Mithen, S. The Music Instinct. Ann. N. Y. Acad. Sci. 2009, 1169, 3–12. [Google Scholar] [CrossRef] [PubMed]
  30. Trevarthen, C. Communicative Musicality: The Human Impulse to Create and Share Music. In Musical Imaginations: Multidisciplinary Perspectives on Creativity, Performance, and Perception; Oxford University Press: Oxford, UK, 2012; pp. 259–284. [Google Scholar]
  31. Herholz, S.C.; Zatorre, R.J. Musical Training as a Framework for Brain Plasticity: Behavior, Function, and Structure. Neuron 2012, 76, 486–502. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Merrett, D.L.; Peretz, I.; Wilson, S.J. Moderating Variables of Music Training-Induced Neuroplasticity: A Review and Discussion. Front. Psychol. 2013, 4, 606. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Kraus, N.; Slater, J.; Thompson, E.C.; Hornickel, J.; Strait, D.L.; Nicol, T.; White-Schwoch, T. Music Enrichment Programs Improve the Neural Encoding of Speech in At-Risk Children. J. Neurosci. 2014, 34, 11913–11918. [Google Scholar] [CrossRef] [PubMed]
  34. Reybrouck, M.; Brattico, E. Neuroplasticity beyond Sounds: Neural Adaptations Following Long-Term Musical Aesthetic Experiences. Brain Sci. 2015, 5, 69–91. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Babiloni, F.; Astolfi, L. Social Neuroscience and Hyperscanning Techniques: Past, Present and Future. Neurosci. Biobehav. Rev. 2014, 44, 76–93. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Czeszumski, A.; Eustergerling, S.; Lang, A.; Menrath, D.; Gerstenberger, M.; Schuberth, S.; Schreiber, F.; Rendon, Z.Z.; König, P. Hyperscanning: A Valid Method to Study Neural Inter-Brain Underpinnings of Social Interaction. Front. Hum. Neurosci. 2020, 14, 39. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Koike, T.; Tanabe, H.C.; Sadato, N. Hyperscanning Neuroimaging Technique to Reveal the “Two-in-One” System in Social Interactions. Neurosci. Res. 2015, 90, 25–32. [Google Scholar] [CrossRef] [Green Version]
  38. Balconi, M.; Vanutelli, M.E. Cooperation and Competition with Hyperscanning Methods: Review and Future Application to Emotion Domain. Front. Comput. Neurosci. 2017, 11, 86. [Google Scholar] [CrossRef] [PubMed]
  39. Mu, Y.; Cerritos, C.; Khan, F. Neural Mechanisms Underlying Interpersonal Coordination: A Review of Hyperscanning Research. Soc. Personal. Psychol. Compass 2018, 12, e12421. [Google Scholar] [CrossRef]
  40. Liu, D.; Liu, S.; Liu, X.; Zhang, C.; Li, A.; Jin, C.; Chen, Y.; Wang, H.; Zhang, X. Interactive Brain Activity: Review and Progress on EEG-Based Hyperscanning in Social Interactions. Front. Psychol. 2018, 9, 1862. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Wang, M.-Y.; Luan, P.; Zhang, J.; Xiang, Y.-T.; Niu, H.; Yuan, Z. Concurrent Mapping of Brain Activation from Multiple Subjects during Social Interaction by Hyperscanning: A Mini-Review. Quant. Imaging Med. Surg. 2018, 8, 819. [Google Scholar] [CrossRef] [PubMed]
  42. Dikker, S.; Wan, L.; Davidesco, I.; Kaggen, L.; Oostrik, M.; McClintock, J.; Rowland, J.; Michalareas, G.; Van Bavel, J.J.; Ding, M.; et al. Brain-to-Brain Synchrony Tracks Real-World Dynamic Group Interactions in the Classroom. Curr. Biol. 2017, 27, 1375–1380. [Google Scholar] [CrossRef]
  43. Hasegawa, C.; Ikeda, T.; Yoshimura, Y.; Hiraishi, H.; Takahashi, T.; Furutani, N.; Hayashi, N.; Minabe, Y.; Hirata, M.; Asada, M.; et al. Mu Rhythm Suppression Reflects Mother-Child Face-to-Face Interactions: A Pilot Study with Simultaneous MEG Recording. Sci. Rep. 2016, 6, 34977. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Jiang, J.; Dai, B.; Peng, D.; Zhu, C.; Liu, L.; Lu, C. Neural Synchronization during Face-to-Face Communication. J. Neurosci. 2012, 32, 16064–16069. [Google Scholar] [CrossRef] [PubMed]
  45. Liu, N.; Mok, C.; Witt, E.E.; Pradhan, A.H.; Chen, J.E.; Reiss, A.L. NIRS-Based Hyperscanning Reveals Inter-Brain Neural Synchronization during Cooperative Jenga Game with Face-to-Face Communication. Front. Hum. Neurosci. 2016, 10, 82. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  46. Balconi, M.; Vanutelli, M.E. EEG Hyperscanning and Behavioral Synchronization during a Joint Actions. Neuropsychol. Trends. 2018, 23–47. [Google Scholar] [CrossRef] [Green Version]
  47. Cui, X.; Bryant, D.M.; Reiss, A.L. NIRS-Based Hyperscanning Reveals Increased Interpersonal Coherence in Superior Frontal Cortex during Cooperation. Neuroimage 2012, 59, 2430–2437. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  48. Feng, X.; Sun, B.; Chen, C.; Li, W.; Wang, Y.; Zhang, W.; Xiao, W.; Shao, Y. Self–Other Overlap and Interpersonal Neural Synchronization Serially Mediate the Effect of Behavioral Synchronization on Prosociality. Soc. Cogn. Affect. Neurosci. 2020, 15, 203–214. [Google Scholar] [CrossRef]
  49. Fishburn, F.A.; Murty, V.P.; Hlutkowsky, C.O.; MacGillivray, C.E.; Bemis, L.M.; Murphy, M.E.; Huppert, T.J.; Perlman, S.B. Putting Our Heads Together: Interpersonal Neural Synchronization as a Biological Mechanism for Shared Intentionality. Soc. Cogn. Affect. Neurosci. 2018, 13, 841–849. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. Hu, X.; Han, Z.R.; Wang, H.; Hu, Y.; Wang, Q.; Feng, S.; Yi, L. The Relation of Parental Emotion Regulation to Child Autism Spectrum Disorder Core Symptoms: The Moderating Role of Child Cardiac Vagal Activity. Front. Psychol. 2018, 9, 2480. [Google Scholar] [CrossRef] [PubMed]
  51. Li, L.; Wang, H.; Luo, H.; Zhang, X.; Zhang, R.; Li, X. Interpersonal Neural Synchronization During Cooperative Behavior of Basketball Players: A FNIRS-Based Hyperscanning Study. Front. Hum. Neurosci. 2020, 14, 169. [Google Scholar] [CrossRef] [PubMed]
  52. Toppi, J.; Borghini, G.; Petti, M.; He, E.J.; De Giusti, V.; He, B.; Astolfi, L.; Babiloni, F. Investigating Cooperative Behavior in Ecological Settings: An EEG Hyperscanning Study. PLoS ONE 2016, 11, e0154236. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  53. Zhang, Y.; Meng, T.; Hou, Y.; Pan, Y.; Hu, Y. Interpersonal Brain Synchronization Associated with Working Alliance during Psychological Counseling. Psychiatry Res. Neuroimaging 2018, 282, 103–109. [Google Scholar] [CrossRef] [PubMed]
  54. Nguyen, T.; Schleihauf, H.; Kayhan, E.; Matthes, D.; Vrtička, P.; Hoehl, S. The Effects of Interaction Quality on Neural Synchrony during Mother-Child Problem Solving. Cortex 2020, 124, 235–249. [Google Scholar] [CrossRef] [PubMed]
  55. Reindl, V.; Gerloff, C.; Scharke, W.; Konrad, K. Brain-to-Brain Synchrony in Parent-Child Dyads and the Relationship with Emotion Regulation Revealed by FNIRS-Based Hyperscanning. NeuroImage 2018, 178, 493–502. [Google Scholar] [CrossRef] [PubMed]
  56. Kinreich, S.; Djalovski, A.; Kraus, L.; Louzoun, Y.; Feldman, R. Brain-to-Brain Synchrony during Naturalistic Social Interactions. Sci. Rep. 2017, 7, 1–12. [Google Scholar]
  57. Long, Y.; Zheng, L.; Zhao, H.; Zhou, S.; Zhai, Y.; Lu, C. Interpersonal Neural Synchronization during Interpersonal Touch Underlies Affiliative Pair Bonding between Romantic Couples. Cereb. Cortex 2021, 31, 1647–1659. [Google Scholar] [CrossRef] [PubMed]
  58. Mu, Y.; Guo, C.; Han, S. Oxytocin Enhances Inter-Brain Synchrony during Social Coordination in Male Adults. Soc. Cogn. Affect. Neurosci. 2016, 11, 1882–1893. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  59. Acquadro, M.A.; Congedo, M.; De Riddeer, D. Music Performance as an Experimental Approach to Hyperscanning Studies. Front. Hum. Neurosci. 2016, 10, 242. [Google Scholar] [CrossRef] [PubMed]
  60. Osaka, N.; Minamoto, T.; Yaoi, K.; Azuma, M.; Shimada, Y.M.; Osaka, M. How Two Brains Make One Synchronized Mind in the Inferior Frontal Cortex: FNIRS-Based Hyperscanning during Cooperative Singing. Front. Psychol. 2015, 6, 1811. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  61. Hou, Y.; Song, B.; Hu, Y.; Pan, Y.; Hu, Y. The Averaged Inter-Brain Coherence between the Audience and a Violinist Predicts the Popularity of Violin Performance. Neuroimage 2020, 211, 116655. [Google Scholar] [CrossRef]
  62. Pan, Y.; Novembre, G.; Song, B.; Zhu, Y.; Hu, Y. Dual Brain Stimulation Enhances Interpersonal Learning through Spontaneous Movement Synchrony. Soc. Cogn. Affect. Neurosci. 2021, 16, 210–221. [Google Scholar] [CrossRef] [PubMed]
  63. Babiloni, C.; Vecchio, F.; Infarinato, F.; Buffo, P.; Marzano, N.; Spada, D.; Rossi, S.; Bruni, I.; Rossini, P.M.; Perani, D. Simultaneous Recording of Electroencephalographic Data in Musicians Playing in Ensemble. Cortex 2011, 47, 1082–1090. [Google Scholar] [CrossRef] [PubMed]
  64. Babiloni, C.; Buffo, P.; Vecchio, F.; Marzano, N.; Del Percio, C.; Spada, D.; Rossi, S.; Bruni, I.; Rossini, P.M.; Perani, D. Brains “in Concert”: Frontal Oscillatory Alpha Rhythms and Empathy in Professional Musicians. Neuroimage 2012, 60, 105–116. [Google Scholar] [CrossRef] [PubMed]
  65. Washburn, A.; Román, I.; Huberth, M.; Gang, N.; Dauer, T.; Reid, W.; Nanou, C.; Wright, M.; Fujioka, T. Musical Role Asymmetries in Piano Duet Performance Influence Alpha-Band Neural Oscillation and Behavioral Synchronization. Front. Neurosci. 2019, 13, 1088. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  66. Lindenberger, U.; Li, S.-C.; Gruber, W.; Müller, V. Brains Swinging in Concert: Cortical Phase Synchronization while Playing Guitar. BMC Neurosci. 2009, 10, 1–12. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  67. Sänger, J.; Müller, V.; Lindenberger, U. Intra-and Interbrain Synchronization and Network Properties When Playing Guitar in Duets. Front. Hum. Neurosci. 2012, 6, 312. [Google Scholar] [CrossRef] [Green Version]
  68. Müller, V.; Sänger, J.; Lindenberger, U. Intra-and Inter-Brain Synchronization during Musical Improvisation on the Guitar. PLoS ONE 2013, 8, e73852. [Google Scholar] [CrossRef] [Green Version]
  69. Pérez, A.; Carreiras, M.; Duñabeitia, J.A. Brain-to-Brain Entrainment: EEG Interbrain Synchronization While Speaking and Listening. Sci. Rep. 2017, 7, 1–12. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  70. Szymanski, C.; Pesquita, A.; Brennan, A.A.; Perdikis, D.; Enns, J.T.; Brick, T.R.; Müller, V.; Lindenberger, U. Teams on the Same Wavelength Perform Better: Inter-Brain Phase Synchronization Constitutes a Neural Substrate for Social Facilitation. Neuroimage 2017, 152, 425–436. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  71. Lakatos, P.; Shah, A.S.; Knuth, K.H.; Ulbert, I.; Karmos, G.; Schroeder, C.E. An Oscillatory Hierarchy Controlling Neuronal Excitability and Stimulus Processing in the Auditory Cortex. J. Neurophysiol. 2005, 94, 1904–1911. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  72. van Atteveldt, N.; Musacchia, G.; Zion-Golumbic, E.; Sehatpour, P.; Javitt, D.C.; Schroeder, C. Complementary FMRI and EEG Evidence for More Efficient Neural Processing of Rhythmic vs. Unpredictably Timed Sounds. Front. Psychol. 2015, 6, 1663. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  73. Abrams, D.A.; Nicol, T.; Zecker, S.; Kraus, N. Right-Hemisphere Auditory Cortex Is Dominant for Coding Syllable Patterns in Speech. J. Neurosci. 2008, 28, 3958–3965. [Google Scholar] [PubMed]
  74. Giraud, A.-L.; Poeppel, D. Cortical Oscillations and Speech Processing: Emerging Computational Principles and Operations. Nat. Neurosci. 2012, 15, 511. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  75. Luo, H.; Poeppel, D. Cortical Oscillations in Auditory Perception and Speech: Evidence for Two Temporal Windows in Human Auditory Cortex. Front. Psychol. 2012, 3, 170. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  76. Golumbic, E.M.Z.; Poeppel, D.; Schroeder, C.E. Temporal Context in Speech Processing and Attentional Stream Selection: A Behavioral and Neural Perspective. Brain Lang. 2012, 122, 151–161. [Google Scholar] [CrossRef] [Green Version]
  77. Lerousseau, J.P.; Trebuchon, A.; Morillon, B.; Schön, D. Frequency Selectivity of Persistent Cortical Oscillatory Responses to Auditory Rhythmic Stimulation. Journal of Neuroscience 2021, 41, 7991–8006. [Google Scholar]
  78. Rimmele, J.M.; Morillon, B.; Poeppel, D.; Arnal, L.H. Proactive Sensing of Periodic and Aperiodic Auditory Patterns. Trends Cogn. Sci. 2018, 22, 870–882. [Google Scholar] [CrossRef]
  79. Lloyd, D. The Musical Structure of Time in the Brain: Repetition, Rhythm, and Harmony in FMRI During Rest and Passive Movie Viewing. Front. Comput. Neurosci. 2020, 13, 98. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  80. Bardy, B.G.; Calabrese, C.; De Lellis, P.; Bourgeaud, S.; Colomer, C.; Pla, S.; di Bernardo, M. Moving in Unison after Perceptual Interruption. Sci. Rep. 2020, 10, 1–13. [Google Scholar] [CrossRef] [PubMed]
  81. Djalovski, A.; Dumas, G.; Kinreich, S.; Feldman, R. Human Attachments Shape Interbrain Synchrony toward Efficient Performance of Social Goals. Neuroimage 2021, 226, 117600. [Google Scholar] [CrossRef]
  82. Chi, Y.M.; Wang, Y.; Wang, Y.-T.; Jung, T.-P.; Kerth, T.; Cao, Y. A Practical Mobile Dry EEG System for Human Computer Interfaces. In Proceedings of the International Conference on Augmented Cognition; Springer: Berlin/Heidelberg, Germany, 2013; pp. 649–655. [Google Scholar]
  83. Oullier, O.; De Guzman, G.C.; Jantzen, K.J.; Lagarde, J.; Scott Kelso, J.A. Social Coordination Dynamics: Measuring Human Bonding. Soc. Neurosci. 2008, 3, 178–192. [Google Scholar] [CrossRef] [Green Version]
  84. Repp, B.H.; Su, Y.-H. Sensorimotor Synchronization: A Review of Recent Research (2006–2012). Psychon. Bull. Rev. 2013, 20, 403–452. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  85. Thaut, M.H.; Demartin, M.; Sanes, J.N. Brain Networks for Integrative Rhythm Formation. PLoS ONE 2008, 3, e2312. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  86. Strang, G. Wavelet Transforms versus Fourier Transforms. Bull. Am. Math. Soc. 1993, 28, 288–305. [Google Scholar] [CrossRef] [Green Version]
  87. Holschneider, M.; Kronland-Martinet, R.; Morlet, J.; Tchamitchian, P. A Real-Time Algorithm for Signal Analysis with the Help of the Wavelet Transform. In Wavelets; Springer: Berlin/Heidelberg, Germany, 1990; pp. 286–297. [Google Scholar]
  88. Schmidt, R.C.; Carello, C.; Turvey, M.T. Phase Transitions and Critical Fluctuations in the Visual Coordination of Rhythmic Movements between People. J. Exp. Psychol. Hum. Percept. Perform. 1990, 16, 227. [Google Scholar] [CrossRef] [PubMed]
  89. London, J. Hearing in Time: Psychological Aspects of Musical Meter; Oxford University Press: Oxford, UK, 2012. [Google Scholar]
  90. Gugnowska, K.; Novembre, G.; Kohler, N.; Villringer, A.; Keller, P.E.; Sammler, D. Endogenous Sources of Interbrain Synchrony in Duetting Pianists. Cereb. Cortex 2022, bhab469. [Google Scholar] [CrossRef]
  91. Burgess, A.P. On the Interpretation of Synchronization in EEG Hyperscanning Studies: A Cautionary Note. Front. Hum. Neurosci. 2013, 7, 881. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  92. Lu, K.; Hao, N. When Do We Fall in Neural Synchrony with Others? Soc. Cogn. Affect. Neurosci. 2019, 14, 253–261. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  93. Novembre, G.; Iannetti, G.D. Hyperscanning Alone Cannot Prove Causality. Multibrain Stimulation Can. Trends Cogn. Sci. 2021, 25, 96–99. [Google Scholar] [CrossRef] [PubMed]
  94. Hamilton, A.F.d.C. Hyperscanning: Beyond the Hype. Neuron 2021, 109, 404–407. [Google Scholar] [CrossRef] [PubMed]
  95. Koban, L.; Ramamoorthy, A.; Konvalinka, I. Why Do We Fall into Sync with Others? Interpersonal Synchronization and the Brain’s Optimization Principle. Soc. Neurosci. 2019, 14, 1–9. [Google Scholar] [CrossRef]
  96. Mastroianni, A.M.; Gilbert, D.T.; Cooney, G.; Wilson, T.D. Do Conversations End When People Want Them To? Proc. Natl. Acad. Sci. USA 2021, 118, e2011809118. [Google Scholar] [CrossRef] [PubMed]
  97. Large, E.W.; Jones, M.R. The Dynamics of Attending: How People Track Time-Varying Events. Psychol. Rev. 1999, 106, 119. [Google Scholar] [CrossRef]
  98. De Graaf, T.A.; Gross, J.; Paterson, G.; Rusch, T.; Sack, A.T.; Thut, G. Alpha-Band Rhythms in Visual Task Performance: Phase-Locking by Rhythmic Sensory Stimulation. PLoS ONE 2013, 8, e60035. [Google Scholar] [CrossRef] [PubMed]
  99. Mathewson, K.E.; Prudhomme, C.; Fabiani, M.; Beck, D.M.; Lleras, A.; Gratton, G. Making Waves in the Stream of Consciousness: Entraining Oscillations in EEG Alpha and Fluctuations in Visual Awareness with Rhythmic Visual Stimulation. J. Cogn. Neurosci. 2012, 24, 2321–2333. [Google Scholar] [CrossRef]
  100. Spaak, E.; de Lange, F.P.; Jensen, O. Local Entrainment of Alpha Oscillations by Visual Stimuli Causes Cyclic Modulation of Perception. J. Neurosci. 2014, 34, 3536–3544. [Google Scholar] [CrossRef]
  101. Stupacher, J.; Witte, M.; Hove, M.J.; Wood, G. Neural Entrainment in Drum Rhythms with Silent Breaks: Evidence from Steady-State Evoked and Event-Related Potentials. J. Cogn. Neurosci. 2016, 28, 1865–1877. [Google Scholar] [CrossRef]
  102. Ross, J.M.; Iversen, J.R.; Balasubramaniam, R. Motor Simulation Theories of Musical Beat Perception. Neurocase 2016, 22, 558–565. [Google Scholar] [CrossRef] [PubMed]
  103. Iversen, J.R.; Balasubramaniam, R. Synchronization and Temporal Processing. Curr. Opin. Behav. Sci. 2016, 8, 175–180. [Google Scholar] [CrossRef] [Green Version]
  104. Iversen, J.; Repp, B.; Patel, A. Top-down Control of Rhythm Perception Modulates Early Auditory Responses. Ann. N. Y. Acad. Sci. 2009, 1169, 58–73. [Google Scholar] [CrossRef] [PubMed]
Figure 1. An illustration of the experimental setup during the tapping game. Participants wearing EEG caps (A,B) faced each other during the tapping game, tapping on a synth drum pad (1), and turned 90 degrees to their left during pre/post silent periods. Guide tones were played through a speaker (2) from a computer (C) in an adjacent room, which was also used for data collection.
Figure 1. An illustration of the experimental setup during the tapping game. Participants wearing EEG caps (A,B) faced each other during the tapping game, tapping on a synth drum pad (1), and turned 90 degrees to their left during pre/post silent periods. Guide tones were played through a speaker (2) from a computer (C) in an adjacent room, which was also used for data collection.
Brainsci 12 00409 g001
Figure 2. Left: Cognionics headset used for the present study (viewed from the left); right top: a research assistant tests the headset and drum machine; right bottom: tined and smooth electrodes for placement in hair and on the skin, respectively.
Figure 2. Left: Cognionics headset used for the present study (viewed from the left); right top: a research assistant tests the headset and drum machine; right bottom: tined and smooth electrodes for placement in hair and on the skin, respectively.
Brainsci 12 00409 g002
Figure 3. (Top) Representation of one instance of the anti-phase tapping game. The sequence above was repeated multiple times at increasing rates (tempi) selected by participants. (Bottom) Representation of the entire study procedure each dyad underwent.
Figure 3. (Top) Representation of one instance of the anti-phase tapping game. The sequence above was repeated multiple times at increasing rates (tempi) selected by participants. (Bottom) Representation of the entire study procedure each dyad underwent.
Brainsci 12 00409 g003
Figure 4. Grand average wavelet power at each of the 5 electrode sites. PRE (black) POST (red). Shading shows standard error and horizontal bars show delta and theta band analysis regions.
Figure 4. Grand average wavelet power at each of the 5 electrode sites. PRE (black) POST (red). Shading shows standard error and horizontal bars show delta and theta band analysis regions.
Brainsci 12 00409 g004
Figure 5. (A) Bar graphs representing mean phase correlation in PRE (grey) and POST (red) conditions at each of the five electrode sites used for the study. Statistical significance is indicated by the asterisks (*). (B) Correlations between duration of musical interaction and POST condition delta band INS.
Figure 5. (A) Bar graphs representing mean phase correlation in PRE (grey) and POST (red) conditions at each of the five electrode sites used for the study. Statistical significance is indicated by the asterisks (*). (B) Correlations between duration of musical interaction and POST condition delta band INS.
Brainsci 12 00409 g005
Figure 6. Relative phase per trial for participant dyads. Each dyad’s trials comprise blue dots connected by a blue line. The final trial of each dyad is marked in bold.
Figure 6. Relative phase per trial for participant dyads. Each dyad’s trials comprise blue dots connected by a blue line. The final trial of each dyad is marked in bold.
Brainsci 12 00409 g006
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Khalil, A.; Musacchia, G.; Iversen, J.R. It Takes Two: Interpersonal Neural Synchrony Is Increased after Musical Interaction. Brain Sci. 2022, 12, 409. https://doi.org/10.3390/brainsci12030409

AMA Style

Khalil A, Musacchia G, Iversen JR. It Takes Two: Interpersonal Neural Synchrony Is Increased after Musical Interaction. Brain Sciences. 2022; 12(3):409. https://doi.org/10.3390/brainsci12030409

Chicago/Turabian Style

Khalil, Alexander, Gabriella Musacchia, and John Rehner Iversen. 2022. "It Takes Two: Interpersonal Neural Synchrony Is Increased after Musical Interaction" Brain Sciences 12, no. 3: 409. https://doi.org/10.3390/brainsci12030409

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop