Next Article in Journal
Intraoperative Brain Mapping in Multilingual Patients: What Do We Know and Where Are We Going?
Next Article in Special Issue
DIANA, a Process-Oriented Model of Human Auditory Word Recognition
Previous Article in Journal
Cerebral Blood Flow in Healthy Subjects with Different Hypnotizability Scores
Previous Article in Special Issue
Neural–Behavioral Relation in Phonetic Discrimination Modulated by Language Background
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Learning to Perceive Non-Native Tones via Distributional Training: Effects of Task and Acoustic Cue Weighting

1
The MARCS Institute for Brain, Behaviour and Development, Western Sydney University, Westmead, NSW 2145, Australia
2
School of Psychology, Western Sydney University, Westmead, NSW 2145, Australia
3
Center of Multilingualism across the Lifespan, University of Oslo, 0316 Oslo, Norway
4
Australian Research Council Centre of Excellence for the Dynamics of Language, Canberra, ACT 2601, Australia
5
School of Foreign Studies, Hefei University of Technology, Hefei 230009, China
6
School of Psychology and Clinical Language Sciences, University of Reading, Reading RG6 6AH, UK
7
Department of Psychology, Bilkent University, Ankara 06800, Turkey
*
Author to whom correspondence should be addressed.
Brain Sci. 2022, 12(5), 559; https://doi.org/10.3390/brainsci12050559
Submission received: 5 March 2022 / Revised: 7 April 2022 / Accepted: 20 April 2022 / Published: 27 April 2022
(This article belongs to the Special Issue Auditory and Phonetic Processes in Speech Perception)

Abstract

:
As many distributional learning (DL) studies have shown, adult listeners can achieve discrimination of a difficult non-native contrast after a short repetitive exposure to tokens falling at the extremes of that contrast. Such studies have shown using behavioural methods that a short distributional training can induce perceptual learning of vowel and consonant contrasts. However, much less is known about the neurological correlates of DL, and few studies have examined non-native lexical tone contrasts. Here, Australian-English speakers underwent DL training on a Mandarin tone contrast using behavioural (discrimination, identification) and neural (oddball-EEG) tasks, with listeners hearing either a bimodal or a unimodal distribution. Behavioural results show that listeners learned to discriminate tones after both unimodal and bimodal training; while EEG responses revealed more learning for listeners exposed to the bimodal distribution. Thus, perceptual learning through exposure to brief sound distributions (a) extends to non-native tonal contrasts, and (b) is sensitive to task, phonetic distance, and acoustic cue-weighting. Our findings have implications for models of how auditory and phonetic constraints influence speech learning.

1. Introduction

Listening to speech involves identifying linguistic structures in a fast, continuous utterance stream and processing their relative ordering to retrieve an intended meaning. In the native language, multiple sources of relevant information are drawn upon to accomplish this task, including knowledge of the vocabulary and the relative likelihood of different sequential patterns at lower and higher levels of linguistic structure, as well as the rules governing sequential processes at the phonetic level. Non-native languages can differ from a listener’s native languages on all these dimensions. We here address the question of whether listeners faced with the task of discriminating a novel non-native contrast show preferences as to which dimension of linguistic information they attend to (or attend to most). To better understand the source of any such preference, we compare behavioural against neurophysiological measures.
There is a considerable literature concerning the perception and learning of non-native contrasts, and it naturally shares ground with the very extensive literature on native contrast learning, with the differences in large part arising from learner constraints, given that in the native situation, the learners are still in infancy. Interestingly, overlap between the native and non-native literatures has been increased by the growing attention paid to statistical accounts of learning processes, due to the extensive evidence showing that such information is not restricted to mature brains, but is processed even by the youngest of learners.
Moreover, statistical learning is not specific to the language domain. It is observed for visual objects [1], spatial structures [2], tactile sequences [3] and music perception [4]. Crucially, two-month-olds exhibit neural sensitivity to statistical properties of non-native speech sounds during sleep [5], indicating that statistical learning is a formative mechanism that may have considerable explanatory power for our understanding of perceptual learning processes.
The language acquisition process can draw on statistical regularities in the linguistic environment ranging from simple frequency counts to complex conditional probabilities, including transitional probabilities [6], non-adjacent dependencies of word occurrences [7] and distributional patterns of speech sounds [8]. Listeners’ most basic task, namely segmenting continuous speech streams into words and speech sounds, can also be assisted by attention to the distributional properties of the input [9]. Even infants in their first year attend to such information. Maye and colleagues [8] showed that infants as young as 6 months of age use distributional properties to learn phonetic categories. They exposed infants to manipulated frequency distributions of speech sounds that differed in their number of peaks, producing typically unimodal (one-peak) versus bimodal (two-peak) distributions, with the former thought to support a single, large category, while the latter (a distribution with two peaks) promotes the assumption of two categories.
A major factor affecting distributional learning is age, with learning appearing to be more effective for infants than for adults [10] and for younger than older infants [11]. Another factor is attention, with better results for learners whose task requires them to attend to the stimuli, as opposed to those who do not need to payattention to the stimuli’s properties [12]. A third factor is the amount of exposure provided to the learner, with more extended distributional exposure yielding better discrimination results [13]. Fourth, experimental design also plays a role: A learning effect is more likely to appear with a habituation-dishabituation procedure than with familiarization-alternation paradigms [14]. Fifth and not least, the learning target itself can directly impact learning outcomes [15] due to its acoustic properties (a non-native contrast similar to the native inventory is more rapidly acquired through DL than a dissimilar contrast [16]) or its perceptual salience (consonants are relatively less salient acoustically than vowels and are thus less rapidly learned distributionally [17,18]). The last factor further leads to a series of cue-weighting studies demonstrating that statistical information in the ambient environment is not the only cue listeners adopt. For example, when learning a speech sound contrast, learners may focus more on the phonetic/acoustic properties of stimuli than statistical regularities constraining the contrast [19,20,21,22].
Most studies in this literature so far have focused on consonants and vowels. But speech has further dimensions, including suprasegmental structure which is manifested on many levels, including the lexical level. Around 60–70% of the world’s languages are tonal, i.e., they use pitch to distinguish word meanings [23]. Tone language speakers show categorical perception for tones, but non-tone language speakers report them as psycho-acoustic rather than linguistic [24]. However, acoustic interactions of tonal and segmental information affect simple discrimination task performance equivalently for native listeners of a tone language and non-native listeners without tonal experience [25]. Specifically, both Cantonese and Dutch listeners take longer and are less accurate when judging syllables that differ in tones compared to segments, suggesting a slower processing of tonal than segmental distinctions, at least in speeded-response tasks.
Non-tonal language speakers have substantial difficulty discriminating and learning tone categories [26], and studies have shown mixed results after distributional training [27]. For infants, bimodal distributional training enhances Mandarin tone perception for Dutch 11-month-olds, but not for 5- or 14-month-olds [28]. For adults, Ong and colleagues [12] found that Australian-English speakers exposed to a bimodal distribution of a Thai tone contrast showed no automatic learning effect; but when the training involved active listening (through a request for acknowledging the heard stimuli), their sensitivity to tone improved. Interestingly, exposure to a bimodal distribution of Thai tones resulted in enhanced perception of both linguistic and musical tones for Australian-English speakers, demonstrating cross-domain transfer [4,29]. Note that linguistic and musical tone perception also tend to be correlated in non-tone language adults’ performance [26,30] and in psycho-acoustic perception [31].
All the above results stem from behavioural research because statistical learning research so far has made little use of neurophysiological methods, despite evidence that neural responses can provide processing information at both pre-attentive and cortical levels [32,33]. Compared to behavioural measures, however, neural techniques such as electroencephalography (EEG) have the benefit of requiring no overt attention or decision processes. They are typically sensitive to early pre-attentive responses, reflecting the neural basis of acoustic-phonetic processing [34,35] and providing another measure of implicit discrimination.
Importantly, evidence from non-native speech perception studies shows that non-native listeners exhibit mismatch negativity (MMN) responses for contrasts they do not discriminate in behavioural tasks [36,37]. The MMN response has been used extensively in speech perception studies to examine how the perceptual system indexes incoming acoustic stimuli pre-attentively, i.e., in a manner not requiring overt attention. Based on the formation of memory traces, the MMN response is a negative-going waveform that is typically observed in the frontal electrodes. It signals detection of change in a stream of auditory stimuli, and specifically indicates differences between stimuli with different acoustics. It is obtained by computing the difference in the event-related potential (ERP) response to an infrequent (termed deviant) stimulus versus a frequent (termed standard) one. This response typically occurs 150 to 250 ms post onset of the stimulus switch [38,39].
Despite the scarcity of neurally-based research on distributional learning, some evidence is available on listeners’ perceptual flexibility for the tonal dimension of speech. In tone perception, pitch height and pitch direction are the two main acoustic cues [40]. Nixon and colleagues [41] explored German listeners’ neural discrimination of a Cantonese high-mid pitch height contrast, by exposing them to a bimodal distribution of a 13-step tone continuum. Although a prediction of enhanced MMN responses at the two Gaussian peaks was not supported, the listeners’ perception of pitch height improved across all steps along the continuum. A follow-up study of listeners’ neural sensitivity to cross-boundary differences again showed enhanced sensitivity to overall pitch differences over the course of the tone exposure [42]. In other words, acuity with respect to acoustic pitch differences was increased during distributional learning not only between but also within categories. In contrast to the findings of previous studies testing segmental features, these results indicate that exposure to a bimodal distribution may not necessarily lead to the enhanced discrimination of specific steps or categories along the pitch continuum, but may rather alter listeners’ overall sensitivity to tonal changes. A caveat in this interpretation is that these studies did not include a unimodal distribution. Taken together, when EEG studies are adopted to examine tone perception and learning, results often illustrate robust sensitivity not merely restricted to the patterns predicted by frequency distributions as shown in behavioural studies.
On the same note, a recent study examining 5–6-month-old infants’ neural sensitivity to an 8-step contrast of flat versus falling pitch (a Mandarin tone contrast) found a surprising enhancement effect after exposure to a unimodal but not a bimodal distribution [43]. This finding was explained in relation to listeners’ acoustic sensitivity to frequently heard tokens at peak locations along the tone continuum. The high-frequency tokens had smaller differences in the unimodal (steps 4–5) than in the bimodal (steps 2–7) distributions. The authors of [43] argued that frequent exposure to greater acoustic distance may lead to reduced neural sensitivity to a smaller acoustic distance (steps 3–6). This interpretation highlights the role of the magnitude of the acoustic distinctions in the stimuli when prior training and exposure is insufficient to establish phonetic categories, which can be explained by models of non-native perception that focus on acoustic cue weighting and salience (see for instance, [44]).
In the present study, we examined tone perception and learning for non-tonal language speakers, collecting and directly comparing behavioural and neural responses. Australian English listeners with no prior knowledge of any tone language heard a Mandarin tone contrast. Tone perception ability has long been known to vary as a function of the acoustic properties of the tonal input [45]. We varied tone features and the nature of the input distribution (comparing uni- versus bimodality). Experiment 1 tested listeners’ ability to discern the level (T1)–falling (T4) Mandarin tone contrast before and after distributional training, using discrimination and identification tasks, respectively. Experiment 2 then examined listeners’ neural sensitivity to the same contrast, in a standard MMN paradigm, assessing amplitude, latency, anteriority, laterality and topographic differences between the two modalities. Previous work has shown differences in factors such as anteriority and laterality depending on stimulus characteristics and participant language background [46,47,48] while other work has not [33,49]. We included them here to ensure that we captured the most comprehensive set of results for examining potential effects of distributional learning at the neural level.

2. Experiment 1: Mandarin Tone Discrimination and Identification by Australian Listeners before and after Distributional Learning

2.1. Methods

2.1.1. Participants

Forty-eight native Australian English speakers naïve to tone languages took part (36 females; Mage = 23.06, SDage = 5.57). Nine participants reported having musical training (ranging from 1 to 10 years), though only one continued to practise music at the time of testing. All participants reported normal speech and hearing, provided written informed consent prior to testing, and received course credit or a small monetary compensation.

2.1.2. Stimuli

This study focused on the Mandarin Chinese level [T1] versus falling [T4] tonal contrast. A female Mandarin speaker produced natural tokens of /taT1/ (‘take’) and /taT4/ (‘big’) in a soundproof booth. Recording used the computer program Audacity and a Genelec 1029A microphone, with a 16-bit sampling depth and a sampling rate of 44.1 kHz. To create the 8-step continuum, equidistant stimulus steps differing in pitch contour were constructed from /taT1/ to /taT4/ (Figure 1) using the following procedure in Praat [50]: First, four interpolation points along the pitch contours (at 0%, 33%, 67% and 100%) were marked (Supplementary Material A). Next, the distances (in Hz) between the corresponding points were divided into seven equal spaces, generating six new layers. New pitch tokens were then created by connecting the four corresponding intermediate points on the same layer. This formed a continuum of eight steps (including the endpoint contours) from /taT1/ (step 1) to /taT4/ (step 8). Stimulus intensity was set to 65 dB SPL and duration was set to approximately 400 ms. Five native Mandarin speakers listened to the stimuli and confirmed that they were acceptable tokens of these Mandarin syllables. The contrast (with the same stimuli) has been used in previous studies [51,52,53,54].

2.1.3. Procedure

The experiment consisted of three phases in the following order: pre-test, distributional learning and post-test. Task programming, presentation of stimuli and response recording were conducted using E-Prime (Psychology Software Tools Inc., Sharpsburg, PA, USA) on a Dell Latitude E5550 laptop. Auditory stimuli were presented at 65 dB SPL via Sennheiser HD 280 Pro headphones and were ordered randomly. No corrective feedback was given. The experiment took approximately 40 min to complete.
In the distributional learning phase, participants were randomly assigned to one of the two conditions: unimodal or bimodal distribution (Figure 2). The two conditions had different distributional peaks (a single central category vs. two separate categories) but were equal in terms of the total amount of distributional learning tonal input (256 tokens) and duration (360 s). In the bimodal condition, stimuli from the peripheral positions of the continuum were presented with higher frequency. In other words, participants heard tokens of steps 2 and 7 most frequently. In the unimodal condition, stimuli near the central positions, namely tokens of steps 4 and 5, were presented most frequently. Crucially, stimulus steps 3 and 6 were presented an equal number of times in both conditions.
At pre- and at post-test, participants were first presented with a discrimination task in which they indicated via keypresses whether paired lexical tone stimuli steps were the same or different. Trials included same (e.g., steps 1–1, 2–2) and different pairs (e.g., steps 2–4, 3–6) along the continuum, with each pair presented 10 times. Trials not involving the target contrast functioned as controls. The inter-stimulus interval between tokens was 1000 ms.
Participants then completed an identification task in which they indicated (again via keypresses) whether the tone of each continuum step (e.g., step 3) was a flat tone (indicated by a flat arrow) or a falling tone (indicated by a falling arrow), with each tone presented 6 times. For each task, four auditory examples were played as practice trials prior to testing. Trials were self-paced and presented in random order.
Analyses for the discrimination task targeted the perception of steps 3 to 6, and those for the identification task focused on step 3 and step 6. As an additional control, the Pitch-Contour Perception Test (PCPT; [56,57]) was included in the post-training phase, to examine participants’ pitch perceptual abilities (Supplementary Material B). This test required indication of whether isolated tone tokens had a flat, rising or falling contour. The PCPT allows allocation -of participants to high and low aptitude groups, to examine whether ability to perceive pitch affects identification and discrimination responses. No differences either in the identification or in the discrimination tasks were observed between listeners with high versus low aptitude in the present study.

2.2. Results

We first compared listeners’ percentage of accurate choices for the target contrast (steps 3–6) in the discrimination task before and after distributional learning to chance (50%) with a one-sample t-test (Table 1). Neither condition showed discrimination above chance before distributional learning, while after training, listeners in both conditions were able to discriminate the contrast. We then conducted a Repeated Measures Analysis of Variance (RM ANOVA) with condition (2-level, unimodal vs. bimodal) as the between-subjects factor and accuracy (2-level, pre- and post-training phases) as the within-subjects variable (Figure 3). The main effect of training was significant (F(1, 46) = 4.731, p = 0.035, ηg2 = 0.093), indicating a difference in accuracy before and after distributional learning. The interaction between condition and test phase was not significant (F(1, 46) = 0.526, p = 0.472, ηg2 = 0.011), suggesting no difference between unimodal and bimodal exposure. In other words, listeners’ tone discrimination improved after exposure to either distribution.
We then examined listeners’ choices in the identification task before and after learning (Table 2). Across participants, Step 6 was more often identified as falling than step 3. We then conducted a repeated measures ANOVA with condition (2-level, unimodal vs. bimodal) as the between-subjects factor, and with percentage of falling choices across phase (2-level, pre- and post-training) and step (2-level, steps 3 & 6) as the within-subjects variables (Figure 4). The only significant factor was step (F(1, 46) = 57.343, p < 0.001, ηg2 = 0.555). No other factors or interactions were significant (Fs < 1.936, ps > 0.170, ηg2 < 0.040). In contrast to the discrimination outcomes, no trace of improvement was observed in listeners’ tone identification after either distributional condition.

2.3. Discussion

Experiment 1 used behavioural measures to investigate how distributional learning of a Mandarin tone contrast affects listeners’ tone discrimination and identification. Listeners showed improved discrimination abilities after exposure to either unimodal or bimodal distributions, and no change was observed in their identification patterns.
In the discrimination task, successful distributional learning would predict enhanced discrimination of steps 3–6 after the bimodal condition, and/or reduced discrimination of this step after the unimodal experience. However, our results showed that listeners’ tonal sensitivity was enhanced after distributional learning irrespective of the embedded statistical information. Statistical exposure appears to benefit participants more in acoustic than in statistical cues. This interpretation resembles reported EEG studies with a Cantonese tone contrast [41,42], and agrees with findings showing that listeners attend more to prosodic than statistical cues to segment speech streams [20]. It is worth mentioning that only a single (bimodal) distribution was tested in these cited studies.
In identification, listeners’ difficulty in anchoring non-native tones to given categories plausibly reflects their lack of tonal categories in the first place (N.B. performance differences in identification versus discrimination have been attested before.) Non-tone language speakers often find it hard to make recognition responses to tones. Their much better tone discrimination ability, in contrast, reflects sensitivity to general pitch information (note that pitch perception in language and music correlate [26,30]).
To further examine the relationship between tone processing, distributional learning, and the cues listeners pay attention to in these processes, Experiment 2 examined listeners’ neural changes induced by distributional learning.

3. Experiment 2: Australian Listeners’ Neural Sensitivity to Tones before and after Distributional Learning

3.1. Methods

3.1.1. Participants

A new sample of 32 Australian English speakers naïve to tone languages participated in Experiment 2 (22 females; Mage = 22.9, SDage = 7.60). Eight participants reported having musical training (ranging from 1 to 7 years), though only one continued to practise music at time of test. Participants provided written informed consent prior to the experiment and received course credit or a small reimbursement for taking part.

3.1.2. Stimuli

The same stimuli were used as in Experiment 1, except that the duration of all stimulus tokens was reduced to 100 ms to accommodate to the EEG paradigm.

3.1.3. Procedure

As in Experiment 1, there were three phases: pre-test, distributional learning, and post-test. In the distributional learning phase, participants were randomly assigned to either the unimodal or the bimodal condition, with equal numbers of bilingual and monolingual participants in each condition. The two conditions did not differ in the total number of exposure trials (256 tokens) or duration (360 s) but varied in the frequency distribution (one vs. two Gaussian peaks) along the phonetic continuum only. Stimuli near the central positions were presented most frequently in the unimodal condition, whereas those from the peripheral sides of the continuum were presented with the highest frequency in the bimodal condition. Importantly, the frequency of occurrence of tokens from steps 3 and 6 was again identical across both conditions.
An EEG passive oddball paradigm was used for pre- and post-test, in which two separate blocks were presented. Step 3 was standard and step 6 was deviant in one block, and this was reversed in the other. The standard-deviant probability ratio was 80–20%. Since steps 3 and 6 were presented the same number of times in each condition, any potential differences observed in the post-test should be attributed to the condition. The sequence of the blocks was counterbalanced. No fewer than three and no more than eight standard stimuli occurred between deviant stimuli. Each block started with 20 standards and contained 500 trials in total. The inter-stimulus interval was randomly varied between 600 and 700 ms. Following the presentation phase, participants heard as a control (lasting approximately 1 min) 100 instances of the deviant stimuli they had heard in the previous oddball presentation. This design allowed a response comparison of the same number of deviant stimuli in the oddball presentation (N = 100) and the control presentation (N = 100).
Participants were tested in a single session in a sound-attenuated booth at the MARCS Institute for Brain, Behaviour and Development at Western Sydney University. They watched a self-selected silent movie with subtitles during the experiment and were instructed to avoid excessive motor movement and to disregard the auditory stimuli. The stimuli were presented binaurally via Etymotic earphones with the intensity set at 70 dB SPL in Praat [50] and the volume level was set at a comfortable listening level consistent across participants as a result of piloting that showed MMN elicitation. The duration of the EEG experiment was approximately 45 min.

3.1.4. EEG Data Recording & Analysis

EEG data were recorded from a 64-channel active BioSemi system with Ag/AgCl electrodes placed according to the international 10/20 system fitted to the participant’s head. Six external electrodes were used: four to record eye movements (above and below the right, on the left and right temple), and two for offline referencing (left and right mastoid). Data were recorded at a 512 Hz sampling rate and we made sure the electrode offset was kept below 50 mV.
Data pre-processing and analysis used EEGLAB [58] and ERPLAB [59]. First, data points were re-referenced to the average of the right and left mastoids. They were then bandpass-filtered with half power cut-offs at 0.1 and 30 Hz at 12 dB/octave. Time windows from 100 to 600 ms post stimulus onset were extracted (“epoched”) from the EEG signal and baseline-corrected by subtracting the mean voltage in the 100 ms pre-stimulus interval from each sample in the window. Independent component analyses were conducted to identify and remove noisy EEG channels. Eye-movement components based on the activity power spectrum, scalp topography, and activity over trials were also removed. Noisy EEG channels that were removed were interpolated using spherical spline interpolation. Artefacts above 70 mV were rejected automatically for all channels. Participants with more than 40% of artefact-contaminated epochs were excluded from further analyses (n = 6). The epochs were averaged separately for standards (excluding the first 20 standards and the standards immediately following a deviant stimulus), for each deviant token, and for each control block.
Two difference waves were calculated by subtracting the mean ERP response to each control stimulus from the mean ERP response to its deviant counterpart. These difference waves were then grand-averaged across participants. In the grand-averaged waveform, we sought a negative peak 100 to 250 ms after consonant production (taking the 20 ms consonant portion of the stimulus into consideration) to ensure that we were measuring the neural response to the tone. This resulted in measuring the 120 to 270 ms time window post-stimulus onset. We then centred a 40 ms time window at the identified peak and measured the mean amplitude in that window per individual participant (cf. [47]). These mean individual amplitudes were our measure of MMN amplitude in further statistical analyses. Within the same 40 ms time window, latency was measured by establishing the most negative peak for each participant. These mean individual latencies then became the measure of MMN latency in the further analyses.

3.2. Results

Following previous studies (e.g., [47,60]), MMN amplitudes and latencies were measured at nine channels (Fz, FCz, Cz, F3, FC3, C3, F4, FC4, C4) and were analysed in two separate mixed analyses of variance (ANOVAs) with a between-subject factor of condition (unimodal vs. bimodal) and within-subject factors of phase (pre- vs. post-training), anteriority (Frontal (F) vs. frontocentral (FC) vs. central I), and laterality (left, middle, right). Based on activating different neural populations, the dependent variables including mean amplitude and peak latency may reflect different processing mechanisms [61]: the former may show the robustness of listeners’ discrimination as well as the acoustic/phonetic difference between the standard and the deviant stimuli, and the latter may reflect the required time to process the difference between the stimuli. Both variables have been used as auditory perceptual processing measures at early pre-attentive levels for native and non-native speech [36,62]. MMN responses tend to occur at frontal (F) and frontocentral (FC) sites. If distributional training has an effect on tone perception, we predicted an increase in MMN amplitude at those sites between pre- and post-test, as evidence of the auditory change in stimuli initiating an involuntary attentional switch [39,63].
MMN mean amplitude. Figure 5 shows the grand-averaged MMN component in response to the contrast at pre- and post-training. There was a main effect of phase (F(1, 30) = 4.37, p = 0.045, ηg2 = 0.046). Specifically, the MMN amplitude at pre-test (M = −1.67, SD = 3.03) was larger than that at post-test (M = −0.80, SD = 2.56). No other effects or interactions were significant.
As condition was our variable of interest, the MMN amplitude response at the Fz electrode site in each phase was compared against zero (Figure 6) following previous literature [49,60,61,64,65]. Participants in the unimodal group exhibited significant MMN amplitudes at pre-test (t(15) = −4.55, p < 0.001, d = −1.14) and at post-test (t(15) = −2.76, p = 0.015, d = −0.69), whereas participants in the bimodal group exhibited a significant MMN amplitude at pre-test (t(15) = −2.90, p = 0.011, d = −0.72) but not at post-test (t(15) = −0.51, p = 0.616, d = −0.13). Paired t-tests comparing the MMN amplitude between pre- and post-test for each condition revealed no difference for the unimodal group (t(15) = −1.01, p = 0.329, d = −0.23) whereas there was a marginal difference for the bimodal group (t(15) = −2.12, p = 0.051, d = −0.35), indicating statistical-learning-induced changes.
MMN peak latency. A mixed ANOVA with condition (2-level, unimodal vs. bimodal) as the between-subjects factor, and within-subject factors of test (pre- vs. post-test), anteriority (F vs. FC vs. C), and laterality (left vs. middle vs. right) was conducted on the mean MMN peak latency. Main effects of phase (F(1, 30) = 240.35, p < 0.001, ηg2 = 0.729) and condition (F(1, 30) = 24.57, p < 0.001, ηg2 = 0.179) were found, which were qualified by a significant phase × condition interaction (F(1, 30) = 19.68, p < 0.001, ηg2 = 0.177). Post hoc Tukey tests revealed that there were no group differences in latency at pre-test (Mdiff = 0.05, p = 0.986) but at post-test, participants in the bimodal group had significantly earlier MMN peaks than those in the unimodal group (Mdiff = 20.39, p < 0.001). No other effects nor interactions were significant.

3.3. Discussion

Experiment 2 showed learning-induced neurophysiological changes in these listeners’ tone perception at pre- and post-test. Reduction in tonal sensitivity was significant in the bimodal condition. In addition, the latency results showed an earlier peak in the post- than in the pre-test, with a larger impact again in the bimodal condition. Although behavioural results indicated limited discrimination prior to training, neural sensitivity was consistent with discrimination. Note that the pre-test sensitivity is not unexpected given listeners’ psycho-acoustic sensitivity to non-native tones across ages [51], with sensitivity modulated by tone salience [66].
At post-test, although behavioural outcomes indicated improved discrimination, listeners’ neural processing was generally weakened in both conditions. The results from the bimodal distribution may suggest that the increased exposure and familiarity with tones in the neural experiment hindered distributional learning. Alternatively, although the acoustic experience had given listeners more familiarity with tones, the information they received was insufficient to establish tonal categories from the frequency distribution.
Indeed, in the bimodal condition where MMN responses were diminished, frequency peaks in the distribution were near the two ends of the continuum (Figure 2, steps 2–7), whereas the peaks in the frequency distribution were at the midpoint (Figure 2, steps 4–5) in the unimodal condition. To process stimuli efficiently, listeners may focus on the most frequently presented (hence, most salient) stimuli in each condition. In the bimodal condition, steps 2 and 7 were highlighted when played alongside stimuli close to the discrimination boundary (steps 3 and 6), and listeners may expect a similar (or larger) difference to detect deviation compared with post-test. Contrasts with a smaller acoustic difference (i.e., steps 3–6) may then be harder to detect. On the other hand, those who were exposed to the unimodal condition, where steps 4 and 5 were the most frequent, would show neural responses to a contrast of larger acoustic difference (i.e., steps 3–6). In other words, the most frequent and prominent steps, namely the peaks of each distribution, would impact subsequent perception (To explore our proposed hypothesis on the impact of frequency peak, an additional behavioural test was conducted measuring Australian listeners’ (N = 12) sensitivity to the designated tonal contrasts with manipulations of the acoustic distance between the tokens. While discrimination of steps 2–7, the acoustically distant contrast mostly presented in the bimodal condition, reached 82% accuracy, discrimination of steps 3–6 was at a chance (p = 0.511 against 0.5) and accuracy in the discrimination of steps 4–5, the acoustically close contrast presented most frequently in the unimodal condition, was 20%. These results reflect contrast salience). The overall pattern suggests that listeners’ sensitivity is more auditory or psychophysical than phonetic or phonological at this stage: they can discriminate, but fail to establish categories.
With either an explanation in terms of acoustic salience, or an explanation in terms of perceptual assimilation, certain interactions between acoustic and statistical cues in listeners’ neural processing will be assumed. Previous research has not only shown humans’ ability to track input frequency distributions from the ambient environment, and their ability to abstract and retain the memory of non-native pitch directional cues, but also clear ability to shift their weighting of acoustic/phonetic cues and to reconfigure their learning strategies [19,20]. In a speech segmentation task when both statistical and prosodic cues are presented, listeners attend more to the latter to acquire speech information [21]. Indeed, when various types of cues (e.g., acoustic feature, frequency distribution) are presented to listeners, the weighting of these cues may be dynamic and change in real-time during experimental training [67,68]. Last but not least, the outcomes also confirm that EEG is more sensitive than behavioral measures in revealing listeners’ responses in the course of speech perception [34,35,36,37].

4. General Discussion

This study tested the learning of a non-native tone contrast by non-tonal language speakers. Data were collected with both behavioural and neural test methods, statistical distributions were varied (along a single continuum) in the input, and both identification and discrimination data were collected and analysed. Non-native listeners’ tone discrimination improved after both unimodal and bimodal exposure, and there was a reduction in sensitivity after training (observed with the more sensitive measures, to wit, MMN amplitude and latency, which were recorded using EEG). In the EEG data, perceptual differences emerged between the two exposure conditions. The bimodal condition appears to have a more negative impact than the unimodal condition after training. The reduced amplitude and early latency in the bimodal condition may reflect inability to establish categories from the frequency distribution, although listeners had become more familiar with tones after training.
Our neurophysiological finding contrasts with results in prior distributional learning studies (see Figure 2), where a bimodal distribution leads listeners to display enhanced distinction of steps midway along the distribution, while theeffect is reversed with the unimodal distribution, where steps within one peak become less distinct perceptually. We proposed that the lack of sensitivity at post-test after bimodal training is due to the exposure to a salient contrast during training that participants then expected to find again at test. In the absence of distributional learning, there was exploitation of salient acoustic cues instead. This interpretation is in line with the phonetic-magnitude hypothesis [44], which holds that the size of acoustic features (or phonetic distinctions, articulatory correlates) plays a more central role than the extent of native language experience, an interpretation consistent with the Second-Language Linguistic Perception model (L2LP; [69,70,71,72]), which highlights the interaction between listeners’ native phonology and the magnitude of the phonetic difference in auditory dimensions. In a way, the adult listeners in the current study behaved like infants without prior knowledge of lexical tones and like adult L2 learners without prior knowledge of vowel duration contrasts. They concentrated on the most salient phonetic cues, while ignoring or lowering the weighting of other cues.
We further argue that compared to segmental features, pitch is a particularly salient feature. Pitch contrasts, regardless of height [41,42] or direction (as illustrated in the current study), may be more prone to be weighted acoustically compared to statistically, especially among non-tone language speakers who predominantly use pitch in pragmatic but not lexical functions. Note that this finding was only seen neurally and not behaviourally, which speaks to the fact that this might only be detected through the deployment of sensitive measures.
When native listeners process lexical tones, they attend to linguistic features such as pitch, intensity, and duration, based on their existing knowledge of the categorical structure and the phonotactics of their language [73]. When non-native listeners are presented with the same input, they have no relevant categorical or phonotactic knowledge to call upon. But their auditory abilities may be assumed to parallel those of the native listener, so that it is not surprising that simple discrimination elicits a similar pattern of performance for native and non-native listeners [25], even when any task involving recourse to frequency distributions leads to very different results from these two listener groups.
Results conform to a heuristic approach to processing: when hearing a subtle non-native tone contrast, listeners’ neural sensitivity may be insufficient to perceive the fine-grained tone steps and turn more towards acoustic information, and especially the most frequently presented (i.e., most salient) stimuli within each type of distribution. The general pattern somewhat conforms to previous findings showing that a bimodal distribution does not necessarily promote distinct discrimination between the two peaks [13,38,39], and a unimodal distribution does not always hinder it [43]. Factors such as acoustic properties or perceptual salience between tokens may play a role.
Furthermore, we argue that listeners who were trained on a bimodal distribution in which the peak contrast (steps 2–7) contains a large, discriminable acoustic distance may exhibit hampered discrimination of contrasts that rest on smaller differences. In consequence, smaller differences along the continuum may be disregarded. On the other hand, training on a unimodal distribution in which the peak contrast (steps 4–5) is extremely difficult to discriminate may ease the processing of contrasts with a larger acoustic difference (i.e., steps 3 and 6). This explanation does not come out of the blue, as similar ideas have been proposed for studies showing that infants are sensitive to acoustically subtle phonetic contrasts [74] and develop their word learning abilities of very small differences in vowels in a speedy fashion [75]. This interpretation assumes that listener sensitivity to ambient acoustic and statistical information can pave the way for perception and learning of new contrasts in a second language. However, our results should not be construed as evidence that native English speakers cannot achieve a level of tone categorization matching that of native tone speakers [76]. Many factors may play a role. For example, past research has shown that although statistical information can prompt the formation of distinct phonemic categories within three minutes for some non-native contrasts, longer duration of exposure may be required to trigger learning in a bimodal condition [13]. In a previous study examining listeners’ non-native tone discrimination ability, an overall effect of learning surfaced after around ten minutes of exposure [52]. In the present experiment, embedded statistical information altered perception within six minutes of exposure. The take-home message for this study is that for non-native contrasts presented without context, the acoustic salience of the sounds may initially matter more than the statistical distribution of the sounds. At least in the initial stages of learning, distributional evidence may play only a minor role. Future studies can investigate whether increased input might introduce robust effects of exposure to a bimodal distribution.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/brainsci12050559/s1, Supplementary Material A: F0 values at 4 points along the /taT1/-/taT4/ continuum; Supplementary Material B: Pitch-Contour Perception Test results; Supplementary Material C: Discrimination task results across all contrasts; Supplementary Material D: Identification task results across all steps.

Author Contributions

Team ACOLYTE proved an irresistible nickname for our collaboration of M.A., A.C., J.H.O., L.L., C.Y., A.T. and P.E.. The specific contributions of each team member however are: Conceptualization, L.L. and P.E.; methodology, J.H.O. and A.T.; data collection, C.Y. and J.H.O.; data curation, L.L.; data analyses, L.L. and J.H.O.; supervision: A.C. and M.A.; writing—original draft preparation, L.L.; writing—review and editing, all authors. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Australian Research Council grants, principally the Centre of Excellence for the Dynamics of Language (CE140100041: A.C., L.L. & P.E.), plus DP130102181 (P.E.), DP140104389 (A.C.) and DP190103067 (M.A. & A.C.). P.E. was further supported by an ARC Future Fellowship (FT160100514), C.Y. by China Scholarship Council Grant 201706695047, and L.L. by a European Union Horizon 2020 research and innovation programme under Marie Skłodowska-Curie grant agreement 798658, hosted by the University of Oslo Center for Multilingualism across the Lifespan, financed by the Research Council of Norway through Center of Excellence funding grant No. 223265.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by Human Ethics Committee of Western Sydney University (protocol code H11383 in 2015).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Anonymized data is available upon request to the first author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kirkham, N.Z.; Slemmer, J.A.; Johnson, S.P. Visual statistical learning in infancy: Evidence for a domain general learning mechanism. Cognition 2002, 83, B35–B42. [Google Scholar] [CrossRef]
  2. Fiser, J.; Aslin, R.N. Unsupervised statistical learning of higher-order spatial structures from visual scenes. Psychol. Sci. 2001, 12, 499–504. [Google Scholar] [CrossRef]
  3. Conway, C.M.; Christiansen, M.H. Modality-constrained statistical learning of tactile, visual, and auditory sequences. J. Exp. Psychol. Learn. Mem. Cogn. 2005, 31, 24. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Ong, J.H.; Burnham, D.; Stevens, C.J. Learning novel musical pitch via distributional learning. J. Exp. Psychol. Learn. Mem. Cogn. 2017, 43, 150. [Google Scholar] [CrossRef]
  5. Wanrooij, K.; Boersma, P.; Van Zuijen, T.L. Fast phonetic learning occurs already in 2-to-3-month old infants: An ERP study. Front. Psychol. 2014, 5, 77. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Saffran, J.R.; Aslin, R.N.; Newport, E.L. Statistical learning by 8-month-old infants. Science 1996, 274, 1926–1928. [Google Scholar] [CrossRef] [Green Version]
  7. Newport, E.L.; Aslin, R.N. Learning at a distance I. Statistical learning of non-adjacent dependencies. Cogn. Psychol. 2004, 48, 127–162. [Google Scholar] [CrossRef]
  8. Maye, J.; Werker, J.F.; Gerken, L. Infant sensitivity to distributional information can affect phonetic discrimination. Cognition 2002, 82, B101–B111. [Google Scholar] [CrossRef]
  9. Saffran, J.R.; Johnson, E.K.; Aslin, R.N.; Newport, E.L. Statistical learning of tone sequences by human infants and adults. Cognition 1999, 70, 27–52. [Google Scholar] [CrossRef]
  10. Wanrooij, K.; Boersma, P.; van Zuijen, T.L. Distributional vowel training is less effective for adults than for infants. A study using the mismatch response. PLoS ONE 2014, 9, e109806. [Google Scholar] [CrossRef]
  11. Reh, R.K.; Hensch, T.K.; Werker, J.F. Distributional learning of speech sound categories is gated by sensitive periods. Cognition 2021, 213, 104653. [Google Scholar] [CrossRef] [PubMed]
  12. Ong, J.H.; Burnham, D.; Escudero, P. Distributional learning of lexical tones: A comparison of attended vs. unattended listening. PLoS ONE 2015, 10, e0133446. [Google Scholar] [CrossRef] [Green Version]
  13. Yoshida, K.A.; Pons, F.; Maye, J.; Werker, J.F. Distributional phonetic learning at 10 months of age. Infancy 2010, 15, 420–433. [Google Scholar] [CrossRef]
  14. Cristia, A. Can infants learn phonology in the lab? A meta-analytic answer. Cognition 2018, 170, 312–327. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Escudero, P.; Benders, T.; Wanrooij, K. Enhanced bimodal distributions facilitate the learning of second language vowels. J. Acoust. Soc. Am. 2011, 130, EL206–EL212. [Google Scholar] [CrossRef] [PubMed]
  16. Chládková, K.; Boersma, P.; Escudero, P. Unattended distributional training can shift phoneme boundaries. Biling. Lang. Cogn. 2022, 1–14. [Google Scholar] [CrossRef]
  17. Pons, F.; Sabourin, L.; Cady, J.C.; Werker, J.F. Distributional learning in vowel distinctions by 8-month-old English infants. In Proceedings of the 28th Annual Conference of the Cognitive Science Society, Vancouver, BC, Canada, 26–29 July 2006. [Google Scholar]
  18. Antoniou, M.; Wong, P.C.M. Varying irrelevant phonetic features hinders learning of the feature being trained. J. Acoust. Soc. Am. 2016, 139, 271–278. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Escudero, P.; Benders, T.; Lipski, S.C. Native, non-native and L2 perceptual cue weighting for Dutch vowels: The case of Dutch, German, and Spanish listeners. J. Phon. 2009, 37, 452–465. [Google Scholar] [CrossRef]
  20. Lany, J.; Saffran, J.R. Statistical learning mechanisms in infancy. Compr. Dev. Neurosci. Neural Circuit Dev. Funct. Brain 2013, 3, 231–248. [Google Scholar]
  21. Marimon Tarter, M. Word Segmentation in German-Learning Infants and German-Speaking Adults: Prosodic and Statistical Cues. Ph.D. Thesis, University of Potsdam, Potsdam, Germany, 2019. Available online: https://publishup.uni-potsdam.de/frontdoor/index/docId/43740 (accessed on 1 January 2020).
  22. Tuninetti, A.; Warren, T.; Tokowicz, N. Cue strength in second-language processing: An eye-tracking study. Q. J. Exp. Psychol. 2015, 68, 568–584. [Google Scholar] [CrossRef] [Green Version]
  23. Yip, M. Tone; Cambridge University Press: Cambridge, UK, 2002. [Google Scholar]
  24. Kaan, E.; Barkley, C.M.; Bao, M.; Wayland, R. Thai lexical tone perception in native speakers of Thai, English and Mandarin Chinese: An event-related potentials training study. BMC Neurosci. 2008, 9, 53. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Cutler, A.; Chen, H.C. Lexical tone in Cantonese spoken-word processing. Percept. Psychophys. 1997, 59, 165–179. [Google Scholar] [CrossRef] [Green Version]
  26. Chen, A.; Liu, L.; Kager, R. Cross-domain correlation in pitch perception: The influence of native language. Lang. Cogn. Neurosci. 2016, 31, 751–760. [Google Scholar] [CrossRef] [Green Version]
  27. Antoniou, M.; Chin, J.L.L. What can lexical tone training studies in adults tell us about tone processing in children? Front. Psychol. 2018, 9, 1. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Liu, L.; Kager, R. Statistical learning of speech sounds is most robust during the period of perceptual attunement. J. Exp. Child Psychol. 2017, 164, 192–208. [Google Scholar] [CrossRef]
  29. Ong, J.H.; Burnham, D.; Stevens, C.J.; Escudero, P. Naïve learners show cross-domain transfer after distributional learning: The case of lexical and musical pitch. Front. Psychol. 2016, 7, 1189. [Google Scholar] [CrossRef] [Green Version]
  30. Liu, L.; Chen, A.; Kager, R. Perception of tones in Mandarin and Dutch adult listeners. Lang. Linguist. 2017, 18, 622–646. [Google Scholar] [CrossRef]
  31. Hallé, P.A.; Chang, Y.C.; Best, C.T. Identification and discrimination of Mandarin Chinese tones by Mandarin Chinese vs. French listeners. J. Phon. 2004, 32, 395–421. [Google Scholar] [CrossRef] [Green Version]
  32. Xu, Y.; Krishnan, A.; Gandour, J.T. Specificity of experience-dependent pitch representation in the brainstem. Neuroreport 2006, 17, 1601–1605. [Google Scholar] [CrossRef]
  33. Chandrasekaran, B.; Krishnan, A.; Gandour, J.T. Mismatch negativity to pitch contours is influenced by language experience. Brain Res. 2007, 1128, 148–156. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Näätänen, R.; Winkler, I. The concept of auditory stimulus representation in cognitive neuroscience. Psychol. Bull. 1999, 125, 826. [Google Scholar] [CrossRef] [PubMed]
  35. Sams, M.; Alho, K.; Näätänen, R. Short-term habituation and dishabituation of the mismatch negativity of the ERP. Psychophysiology 1984, 21, 434–441. [Google Scholar] [CrossRef] [PubMed]
  36. Kraus, N.; McGee, T.; Carrell, T.D.; King, C.; Tremblay, K.; Nicol, T. Central auditory system plasticity associated with speech discrimination training. J. Cogn. Neurosci. 1995, 7, 25–32. [Google Scholar] [CrossRef] [PubMed]
  37. Lipski, S.C.; Escudero, P.; Benders, T. Language experience modulates weighting of acoustic cues for vowel perception: An event-related potential study. Psychophysiology 2012, 49, 638–650. [Google Scholar] [CrossRef] [PubMed]
  38. Näätänen, R. The perception of speech sounds by the human brain as reflected by the mismatch negativity (MMN) and its magnetic equivalent (MMNm). Psychophysiology 2001, 38, 1–21. [Google Scholar] [CrossRef]
  39. Näätänen, R.; Paavilainen, P.; Rinne, T.; Alho, K. The mismatch negativity (MMN) in basic research of central auditory processing: A review. Clin. Neurophysiol. 2007, 118, 2544–2590. [Google Scholar] [CrossRef] [PubMed]
  40. Gandour, J. Tone perception in Far Eastern languages. J. Phon. 1983, 11, 149–175. [Google Scholar] [CrossRef]
  41. Nixon, J.S.; Boll-Avetisyan, N.; Lentz, T.O.; van Ommen, S.; Keij, B.; Cöltekin, C.; Liu, L.; van Rij, J. Short-term exposure enhances perception of both between-and within-category acoustic information. In Proceedings of the 9th International Conference on Speech Prosody, Poznan, Poland, 13–16 June 2018; pp. 114–118. [Google Scholar]
  42. Boll-Avetisyan, N.; Nixon, J.S.; Lentz, T.O.; Liu, L.; van Ommen, S.; Çöltekin, Ç.; van Rij, J. Neural Response Development During Distributional Learning. In Proceedings of the Interspeech, Hyderabad, India, 2–6 September 2018; pp. 1432–1436. [Google Scholar]
  43. Liu, L.; Ong, J.H.; Peter, V.; Escudero, P. Revisiting infant distributional learning using event-related potentials: Does unimodal always inhibit and bimodal always facilitate? In Proceedings of the 10th International Conference on Speech Prosody, Online, 25 May–31 August 2020. [Google Scholar]
  44. Escudero, P.; Best, C.T.; Kitamura, C.; Mulak, K.E. Magnitude of phonetic distinction predicts success at early word learning in native and non-native accents. Front. Psychol. 2014, 5, 1059. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Burnham, D.; Francis, E.; Webster, D.; Luksaneeyanawin, S.; Attapaiboon, C.; Lacerda, F.; Keller, P. Perception of lexical tone across languages: Evidence for a linguistic mode of processing. In Proceeding of the Fourth International Conference on Spoken Language Processing (ICSLP 1996), Philadelphia, PA, USA, 3–6 October 1996; Volume 4, pp. 2514–2517. [Google Scholar]
  46. Gu, F.; Zhang, C.; Hu, A.; Zhao, G. Left hemisphere lateralization for lexical and acoustic pitch processing in Cantonese speakers as revealed by mismatch negativity. Neuroimage 2013, 83, 637–645. [Google Scholar] [CrossRef] [PubMed]
  47. Tuninetti, A.; Chládková, K.; Peter, V.; Schiller, N.O.; Escudero, P. When speaker identity is unavoidable: Neural processing of speaker identity cues in natural speech. Brain Lang. 2017, 174, 42–49. [Google Scholar] [CrossRef] [PubMed]
  48. Tuninetti, A.; Tokowicz, N. The influence of a first language: Training nonnative listeners on voicing contrasts. Lang. Cogn. Neurosci. 2018, 33, 750–768. [Google Scholar] [CrossRef]
  49. Chen, A.; Peter, V.; Wijnen, F.; Schnack, H.; Burnham, D. Are lexical tones musical? Native language’s influence on neural response to pitch in different domains. Brain Lang. 2018, 180, 31–41. [Google Scholar] [CrossRef] [PubMed]
  50. Boersma, P.; Weenink, D. Praat: Doing Phonetics by Computer (Version 5.1. 05) [Computer Program]. Available online: https://www.fon.hum.uva.nl/praat/ (accessed on 1 May 2009).
  51. Huang, T.; Johnson, K. Language specificity in speech perception: Perception of Mandarin tones by native and nonnative listeners. Phonetica 2010, 67, 243–267. [Google Scholar] [CrossRef]
  52. Liu, L.; Ong, J.H.; Tuninetti, A.; Escudero, P. One way or another: Evidence for perceptual asymmetry in pre-attentive learning of non-native contrasts. Front. Psychol. 2018, 9, 162. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  53. Liu, L.; Kager, R. Perception of tones by infants learning a non-tone language. Cognition 2014, 133, 385–394. [Google Scholar] [CrossRef] [PubMed]
  54. Liu, L.; Kager, R. Perception of tones by bilingual infants learning non-tone languages. Biling. Lang. Cogn. 2017, 20, 561–575. [Google Scholar] [CrossRef]
  55. Ong, J.H.; Burnham, D.; Escudero, P.; Stevens, C.J. Effect of linguistic and musical experience on distributional learning of nonnative lexical tones. J. Speech Lang. Hear. Res. 2017, 60, 2769–2780. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  56. Wong, P.C.; Perrachione, T.K. Learning pitch patterns in lexical identification by native English-speaking adults. Appl. Psycholinguist. 2007, 28, 565–585. [Google Scholar] [CrossRef] [Green Version]
  57. Perrachione, T.K.; Lee, J.; Ha, L.Y.; Wong, P.C. Learning a novel phonological contrast depends on interactions between individual differences and training paradigm design. J. Acoust. Soc. Am. 2011, 130, 461–472. [Google Scholar] [CrossRef] [Green Version]
  58. Delorme, A.; Makeig, S. EEGLAB: An open source toolbox for analysis of single-trial EEG dynamics including independent component analysis. J. Neurosci. Methods 2004, 134, 9–21. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  59. Lopez-Calderon, J.; Luck, S.J. ERPLAB: An open-source toolbox for the analysis of event-related potentials. Front. Hum. Neurosci. 2014, 8, 213. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  60. Colin, C.; Hoonhorst, I.; Markessis, E.; Radeau, M.; De Tourtchaninoff, M.; Foucher, A.; Collet, G.; Deltenre, P. Mismatch negativity (MMN) evoked by sound duration contrasts: An unexpected major effect of deviance direction on amplitudes. Clin. Neurophysiol. 2009, 120, 51–59. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  61. Horváth, J.; Czigler, I.; Jacobsen, T.; Maess, B.; Schröger, E.; Winkler, I. MMN or no MMN: No magnitude of deviance effect on the MMN amplitude. Psychophysiology 2008, 45, 60–69. [Google Scholar] [CrossRef] [PubMed]
  62. Cheour, M.; Shestakova, A.; Alku, P.; Ceponiene, R.; Näätänen, R. Mismatch negativity shows that 3–6-year-old children can learn to discriminate non-native speech sounds within two months. Neurosci. Lett. 2002, 325, 187–190. [Google Scholar] [CrossRef]
  63. Escera, C.; Alho, K.; Winkler, I.; Näätänen, R. Neural mechanisms of involuntary attention to acoustic novelty and change. J. Cogn. Neurosci. 1998, 10, 590–604. [Google Scholar] [CrossRef] [PubMed]
  64. Liu, R.; Holt, L.L. Neural changes associated with nonspeech auditory category learning parallel those of speech category acquisition. J. Cogn. Neurosci. 2011, 23, 683–698. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  65. Näätänen, R.; Lehtokoski, A.; Lennes, M.; Cheour, M.; Huotilainen, M.; Iivonen, A.; Vainio, M.; Alku, P.; Ilmoniemi, R.J.; Alho, K.; et al. Language-specific phoneme representations revealed by electric and magnetic brain responses. Nature 1997, 385, 432–434. [Google Scholar] [CrossRef]
  66. Burnham, D.K.; Singh, L. Coupling tonetics and perceptual attunement: The psychophysics of lexical tone contrast salience. J. Acoust. Soc. Am. 2018, 144, 1716. [Google Scholar] [CrossRef]
  67. Thiessen, E.D.; Saffran, J.R. When cues collide: Use of stress and statistical cues to word boundaries by 7-to 9-month-old infants. Dev. Psychol. 2003, 39, 706. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  68. Origlia, A.; Cutugno, F.; Galatà, V. Continuous emotion recognition with phonetic syllables. Speech Commun. 2014, 57, 155–169. [Google Scholar] [CrossRef]
  69. Escudero, P.; Boersma, P. Bridging the gap between L2 speech perception research and phonological theory. Stud. Second. Lang. Acquis. 2004, 26, 551–585. [Google Scholar] [CrossRef] [Green Version]
  70. Escudero, P. Linguistic Perception and Second Language Acquisition: Explaining the Attainment of Optimal Phonological Categorization; Netherlands Graduate School of Linguistics: Amsterdam, The Netherlands, 2005. [Google Scholar]
  71. Escudero, P. The linguistic perception of similar L2 sounds. In Phonology in Perception; Boersma, P., Hamann, S., Eds.; Mouton de Gruyter: Berlin, Germany, 2009; pp. 152–190. [Google Scholar]
  72. Yazawa, K.; Whang, J.; Kondo, M.; Escudero, P. Language-dependent cue weighting: An investigation of perception modes in L2 learning. Second. Lang. Res. 2020, 36, 557–581. [Google Scholar] [CrossRef]
  73. Zeng, Z.; Mattock, K.; Liu, L.; Peter, V.; Tuninetti, A.; Tsao, F.-M. Mandarin and English adults’ cue-weighting of lexical stress. In Proceedings of the 21st Annual Conference of the International Speech Communication Association, Shanghai, China, 25–29 October 2020. [Google Scholar]
  74. Sundara, M.; Ngon, C.; Skoruppa, K.; Feldman, N.H.; Onario, G.M.; Morgan, J.L.; Peperkamp, S. Young infants’ discrimination of subtle phonetic contrasts. Cognition 2018, 178, 57–66. [Google Scholar] [CrossRef] [PubMed]
  75. Escudero, P.; Mulak, K.E.; Elvin, J.; Traynor, N.M. “Mummy, keep it steady”: Phonetic variation shapes word learning at 15 and 17 months. Dev. Sci. 2018, 21, e12640. [Google Scholar] [CrossRef]
  76. Xi, J.; Zhang, L.; Shu, H.; Zhang, Y.; Li, P. Categorical perception of lexical tones in Chinese revealed by mismatch negativity. Neuroscience 2010, 170, 223–231. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Pitch contours along a /taT1/–/taT4/ continuum (Stimuli and figure from [28]).
Figure 1. Pitch contours along a /taT1/–/taT4/ continuum (Stimuli and figure from [28]).
Brainsci 12 00559 g001
Figure 2. Frequency of occurrence for each training token encountered by listeners in the unimodal (grey line) and bimodal (black line) conditions. Figure from [55].
Figure 2. Frequency of occurrence for each training token encountered by listeners in the unimodal (grey line) and bimodal (black line) conditions. Figure from [55].
Brainsci 12 00559 g002
Figure 3. Mean accuracy percentage before and after distributional learning (Error bars = ±1 standard error). The horizontal line indicates chance level (50%) performance.
Figure 3. Mean accuracy percentage before and after distributional learning (Error bars = ±1 standard error). The horizontal line indicates chance level (50%) performance.
Brainsci 12 00559 g003
Figure 4. Mean percentage of “falling” classifications for steps 3 and 6 before and after bimodal (left) and unimodal (right) distributional learning (Error bars = ±1 standard error) White bars indicate performance at pretest, and black bars posttest. The horizontal line indicates chance (50%) performance.
Figure 4. Mean percentage of “falling” classifications for steps 3 and 6 before and after bimodal (left) and unimodal (right) distributional learning (Error bars = ±1 standard error) White bars indicate performance at pretest, and black bars posttest. The horizontal line indicates chance (50%) performance.
Brainsci 12 00559 g004
Figure 5. Grand-averaged MMN component by unimodal (left) and bimodal (right) condition. Dotted lines show the MMN component at pre-training and solid lines represent the MMN component at post-training. The red boxes highlight the time window in which the MMN amplitude peaks were measured (i.e., 120–270 ms post-stimulus onset to account for consonant production).
Figure 5. Grand-averaged MMN component by unimodal (left) and bimodal (right) condition. Dotted lines show the MMN component at pre-training and solid lines represent the MMN component at post-training. The red boxes highlight the time window in which the MMN amplitude peaks were measured (i.e., 120–270 ms post-stimulus onset to account for consonant production).
Brainsci 12 00559 g005
Figure 6. Mean MMN amplitude (large dots) for the two conditions at each test phase. Small dots represent individual data. Error bars represent one standard error. Asterisks represent significant MMN amplitude.
Figure 6. Mean MMN amplitude (large dots) for the two conditions at each test phase. Small dots represent individual data. Error bars represent one standard error. Asterisks represent significant MMN amplitude.
Brainsci 12 00559 g006
Table 1. Mean (SD) accuracy percentage and corresponding t and p values in one-sample t-test against the chance level in bimodal and unimodal before and after distributional learning in the discrimination task. (See Supplementary Material C for descriptive statistics of other contrasts.).
Table 1. Mean (SD) accuracy percentage and corresponding t and p values in one-sample t-test against the chance level in bimodal and unimodal before and after distributional learning in the discrimination task. (See Supplementary Material C for descriptive statistics of other contrasts.).
MeanSDTp
BimodalPre60.83%35.62%1.4900.150
Post67.50%33.26%2.5770.017
UnimodalPre51.67%30.59%0.2670.792
Post65.00%30.21%2.4320.023
Table 2. Mean (SD) percentage of choosing falling over flat tones in bimodal and unimodal before and after distributional learning in the identification task. (See Supplementary Material D for descriptive statistics of other contrasts.).
Table 2. Mean (SD) percentage of choosing falling over flat tones in bimodal and unimodal before and after distributional learning in the identification task. (See Supplementary Material D for descriptive statistics of other contrasts.).
Step 3Step 6
MeanSDtpMeanSDtp
BimodalPre34.75%28.62%−2.6100.01667.29%31.61%2.6800.013
Post25.08%33.30%−3.6650.00165.87%35.28%2.2040.038
UnimodalPre45.83%28.30%−0.7210.47874.29%29.79%3.9940.001
Post42.25%28.98%−1.3100.20369.38%29.81%3.1830.004
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, L.; Yuan, C.; Ong, J.H.; Tuninetti, A.; Antoniou, M.; Cutler, A.; Escudero, P. Learning to Perceive Non-Native Tones via Distributional Training: Effects of Task and Acoustic Cue Weighting. Brain Sci. 2022, 12, 559. https://doi.org/10.3390/brainsci12050559

AMA Style

Liu L, Yuan C, Ong JH, Tuninetti A, Antoniou M, Cutler A, Escudero P. Learning to Perceive Non-Native Tones via Distributional Training: Effects of Task and Acoustic Cue Weighting. Brain Sciences. 2022; 12(5):559. https://doi.org/10.3390/brainsci12050559

Chicago/Turabian Style

Liu, Liquan, Chi Yuan, Jia Hoong Ong, Alba Tuninetti, Mark Antoniou, Anne Cutler, and Paola Escudero. 2022. "Learning to Perceive Non-Native Tones via Distributional Training: Effects of Task and Acoustic Cue Weighting" Brain Sciences 12, no. 5: 559. https://doi.org/10.3390/brainsci12050559

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop