50
Accepted manuscripts are peer-reviewed but have not been through the copyediting, formatting, or proofreading process. Copyright © 2017 the authors This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version. Research Articles: Behavioral/Cognitive Sleep disrupts high-level speech parsing despite significant basic auditory processing Shiri Makov 1,3 , Omer Sharon 1 , Nai Ding 2 , Michal Ben-Shachar 3,4 , Yuval Nir 1,5,6,† and Elana Zion Golumbic 3,† 1 Sagol School of Neuroscience, Tel Aviv University, Tel Aviv, 69978, Israel. 2 College of Biomedical Engineering and Instrument Sciences, Zhejiang University, Hangzhou, 310027, China. 3 The Gonda Center for Multidisciplinary Brain Research, Bar Ilan University, Ramat Gan, 5290002, Israel. 4 Department of English Literature and Linguistics, Bar Ilan University, Ramat Gan, 5290002, Israel. 5 Department of Physiology and Pharmacology, Sackler School of Medicine, Tel Aviv University, Tel Aviv, 69978, Israel. 6 Functional Neurophysiology and Sleep Research Lab, Tel-Aviv Sourasky Medical Center, 64239, Tel-Aviv, Israel. DOI: 10.1523/JNEUROSCI.0168-17.2017 Received: 19 January 2017 Revised: 27 April 2017 Accepted: 3 May 2017 Published: 16 June 2017 Author contributions: S.M., E.Z.G. and Y.N. designed research; S.M. performed research; S.M., N.D. and M.B.S. developed experimental materials; S.M. and O.S. contributed the experimental setup and analytic tools; S.M analyzed data; S.M., E.Z.G. and Y.N. wrote the paper. Conflict of Interest: The authors declare no conflict of interest. This work was supported by the I-CORE Program of the Planning and Budgeting Committee and the Israel Science Foundation (YN and EZG), the FP7 Marie Curie Career Integration Grant (YN and EZG), ISF grant 1326/15 (YN), BSF grant 2015385 (EZG) and the Adelis Foundation (YN). The authors declare no competing financial interests. We thank Talma Hendler for continuing support at the Tel Aviv Sourasky Medical Center; Shani Shalgi for help setting up the EEG-sleep lab; Shlomit Beker for her assistance in setting up the experiment; Noam Amir for advising on acoustic aspects of stimulus preparation; Netta Neeman for assistance with data acquisition; Noa Bar-Ilan Regev for administrative help; Yaniv Sela and lab members for suggestions. Equal senior authorship Corresponding authors: Elana Zion-Golumbic ([email protected]) ; The Gonda Center for Multidisciplinary Brain Research, Bar Ilan University, Ramat Gan, 5290002, Israel) and Yuval Nir ([email protected]) ; Sagol School of Neuroscience, Tel Aviv University, Tel Aviv, 69978, Israel). Cite as: J. Neurosci ; 10.1523/JNEUROSCI.0168-17.2017 Alerts: Sign up at www.jneurosci.org/cgi/alerts to receive customized email alerts when the fully formatted version of this article is published.

Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

  • Upload
    others

  • View
    0

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

Accepted manuscripts are peer-reviewed but have not been through the copyediting, formatting, or proofreadingprocess.

Copyright © 2017 the authors

This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version.

Research Articles: Behavioral/Cognitive

Sleep disrupts high-level speech parsing despite significant basic auditoryprocessing

Shiri Makov1,3, Omer Sharon1, Nai Ding2, Michal Ben-Shachar3,4, Yuval Nir1,5,6,† and Elana Zion

Golumbic3,†

1Sagol School of Neuroscience, Tel Aviv University, Tel Aviv, 69978, Israel.2College of Biomedical Engineering and Instrument Sciences, Zhejiang University, Hangzhou, 310027, China.3The Gonda Center for Multidisciplinary Brain Research, Bar Ilan University, Ramat Gan, 5290002, Israel.4Department of English Literature and Linguistics, Bar Ilan University, Ramat Gan, 5290002, Israel.5Department of Physiology and Pharmacology, Sackler School of Medicine, Tel Aviv University, Tel Aviv, 69978,Israel.6Functional Neurophysiology and Sleep Research Lab, Tel-Aviv Sourasky Medical Center, 64239, Tel-Aviv,Israel.

DOI: 10.1523/JNEUROSCI.0168-17.2017

Received: 19 January 2017

Revised: 27 April 2017

Accepted: 3 May 2017

Published: 16 June 2017

Author contributions: S.M., E.Z.G. and Y.N. designed research; S.M. performed research; S.M., N.D. andM.B.S. developed experimental materials; S.M. and O.S. contributed the experimental setup and analytic tools;S.M analyzed data; S.M., E.Z.G. and Y.N. wrote the paper.

Conflict of Interest: The authors declare no conflict of interest.

This work was supported by the I-CORE Program of the Planning and Budgeting Committee and the IsraelScience Foundation (YN and EZG), the FP7 Marie Curie Career Integration Grant (YN and EZG), ISFgrant 1326/15 (YN), BSF grant 2015385 (EZG) and the Adelis Foundation (YN). The authors declare nocompeting financial interests. We thank Talma Hendler for continuing support at the Tel Aviv Sourasky MedicalCenter; Shani Shalgi for help setting up the EEG-sleep lab; Shlomit Beker for her assistance in setting up theexperiment; Noam Amir for advising on acoustic aspects of stimulus preparation; Netta Neeman for assistancewith data acquisition; Noa Bar-Ilan Regev for administrative help; Yaniv Sela and lab members for suggestions.

†Equal senior authorship

Corresponding authors: Elana Zion-Golumbic ([email protected]) ; The Gonda Centerfor Multidisciplinary Brain Research, Bar Ilan University, Ramat Gan, 5290002, Israel) and Yuval Nir([email protected]) ; Sagol School of Neuroscience, Tel Aviv University, Tel Aviv, 69978, Israel).

Cite as: J. Neurosci ; 10.1523/JNEUROSCI.0168-17.2017

Alerts: Sign up at www.jneurosci.org/cgi/alerts to receive customized email alerts when the fully formattedversion of this article is published.

Page 2: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

1

Title: Sleep disrupts high-level speech parsing despite 1

significant basic auditory processing 2

Abbreviated title: Sleep disrupts high-level speech parsing 3

Authors and affiliations: 4

Shiri Makov1,3, Omer Sharon1, Nai Ding2, Michal Ben-Shachar3,4, Yuval Nir1,5,6*, Elana Zion 5

Golumbic3*. 6

* Equal senior authorship 7

8

1. Sagol School of Neuroscience, Tel Aviv University, Tel Aviv, 69978, Israel. 9

2. College of Biomedical Engineering and Instrument Sciences, Zhejiang University, 10

Hangzhou, 310027, China. 11

3. The Gonda Center for Multidisciplinary Brain Research, Bar Ilan University, Ramat Gan, 12

5290002, Israel. 13

4. Department of English Literature and Linguistics, Bar Ilan University, Ramat Gan, 14

5290002, Israel. 15

5. Department of Physiology and Pharmacology, Sackler School of Medicine, Tel Aviv 16

University, Tel Aviv, 69978, Israel. 17

6. Functional Neurophysiology and Sleep Research Lab, Tel-Aviv Sourasky Medical Center, 18

64239, Tel-Aviv, Israel. 19

Page 3: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

2

Corresponding authors: 20

Elana Zion-Golumbic ([email protected]) ; The Gonda Center for 21

Multidisciplinary Brain Research, Bar Ilan University, Ramat Gan, 5290002, Israel) and Yuval 22

Nir ([email protected]) ; Sagol School of Neuroscience, Tel Aviv University, Tel Aviv, 69978, 23

Israel). 24

25

Number of pages: 44 26

Number of figures: 4 27

Number of tables: 1 28

Number of words: Abstract: 244; Introduction: 650; Discussion: 1500 29

30

Author contributions: S.M., E.Z.G. and Y.N. designed research; S.M. performed research; 31

S.M., N.D. and M.B.S. developed experimental materials; S.M. and O.S. contributed the 32

experimental setup and analytic tools; S.M analyzed data; S.M., E.Z.G. and Y.N. wrote the 33

paper. 34

The authors declare no conflict of interest. 35

36

Acknowledgements: This work was supported by the I-CORE Program of the Planning and 37

Budgeting Committee and the Israel Science Foundation (YN and EZG), the FP7 Marie Curie 38

Career Integration Grant (YN and EZG), ISF grant 1326/15 (YN), BSF grant 2015385 (EZG) and 39

Page 4: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

3

the Adelis Foundation (YN). The authors declare no competing financial interests. We thank 40

Talma Hendler for continuing support at the Tel Aviv Sourasky Medical Center; Shani Shalgi 41

for help setting up the EEG-sleep lab; Shlomit Beker for her assistance in setting up the 42

experiment; Noam Amir for advising on acoustic aspects of stimulus preparation; Netta 43

Neeman for assistance with data acquisition; Noa Bar-Ilan Regev for administrative help; 44

Yaniv Sela and lab members for suggestions. 45

Page 5: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

4

Abstract 46

The extent to which the sleeping brain processes sensory information remains 47

unclear. This is particularly true for continuous and complex stimuli, such as 48

speech, in which information is organized into hierarchically embedded 49

structures. Recently, novel metrics for assessing the neural representation of 50

continuous speech have been developed, using non-invasive brain recordings. 51

Those were so far tested during wakefulness. Here we investigated, for the first 52

time, the sleeping brain’s capacity to process continuous speech at different 53

hierarchical levels, using a newly developed Concurrent Hierarchical Tracking 54

(CHT) approach that can online monitor the neural representation and 55

processing-depth of continuous speech. Speech sequences were compiled with 56

syllables, words, phrases and sentences occurring at fixed time intervals, such 57

that different linguistic levels correspond to distinct frequencies. This allowed us 58

to distinguish their neural signatures in brain activity. We compared the neural 59

tracking of intelligible vs. unintelligible (scrambled and foreign) speech across 60

states of wakefulness and sleep, using high-density electroencephalogram (EEG) 61

in humans. We found that neural tracking of stimulus acoustics was comparable 62

across wakefulness and sleep, and similar across all conditions regardless of 63

speech intelligibility. In contrast, neural tracking of higher-order linguistic 64

constructs (words, phrases and sentences) was only observed for intelligible 65

Page 6: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

5

speech during wakefulness, and could not be detected at all during NREM or REM 66

sleep. These results suggest that while low-level auditory processing is relatively 67

preserved during sleep, higher-level hierarchical linguistic parsing is severely 68

disrupted, thereby revealing the capacity and limits of language processing 69

during sleep. 70

Page 7: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

6

Significance statement 71

Despite the persistence of some sensory processing during sleep, it is unclear whether 72

high-level cognitive processes, such as speech parsing, are also preserved. We used a 73

novel approach for studying the depth of speech processing across wakefulness and 74

sleep, while tracking neuronal activity with EEG. We found that responses to the 75

auditory sound-stream remain intact, however the sleeping brain does not show signs of 76

hierarchical parsing of the continuous stream of syllables into words, phrases, and 77

sentences. The results suggest that sleep imposes a functional barrier between basic 78

sensory processing and high-level cognitive processing. This paradigm also holds 79

promise for studying residual cognitive abilities in a wide array of unresponsive states. 80

Page 8: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

7

Introduction 81

Sleep is defined as a reversible state where external stimuli rarely affect perception or 82

elicit meaningful behavioral responses (Nir and Tononi, 2010). Despite such 83

disconnection, it is clear that some discriminative sensory processing persists during 84

sleep. Recent studies, particularly in the auditory domain, found preserved activation of 85

primary sensory cortices during sleep (Pena et al., 1999; Portas et al., 2000; Issa and 86

Wang, 2008; Andrillon et al., 2015). Accordingly, single-neuron responses to both simple 87

(click or tone) and complex (vocalization) stimuli are comparable to those in 88

wakefulness, including frequency tuning curves and stimulus-specific adaptation to 89

deviant sounds (Issa and Wang, 2008; Nir et al., 2015). Thus, external stimuli give rise to 90

robust sensory representations in the sleeping brain, yet the extent of their processing 91

remains unclear. 92

The uncertainty regarding the depth of processing is enhanced with respect to complex 93

stimuli such as speech, which require both sensory and high-order processing, engaging 94

multiple brain areas beyond auditory cortex (Peelle et al., 2010; Wylie and Regner, 95

2014; Friederici and Singer, 2015). To date, there is limited and inconsistent evidence 96

regarding the level of processing that speech undergoes during sleep. Several studies 97

reported that during sleep, activity in high-order language-related regions (such as the 98

superior temporal gyrus (STG), temporal-parietal junction (TPJ) and the inferior frontal 99

gyrus (IFG)), is robustly attenuated or absent altogether, and brain responses to regular 100

speech and to meaningless control conditions are similar (Portas et al., 2000; Dehaene-101

Page 9: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

8

Lambertz et al., 2002; Wilf et al., 2016). However, other evidence suggests that some 102

level of semantic information is extracted even during sleep. For example, behaviorally 103

relevant stimuli such as one’s name, lead to more frequent awakenings (Oswald et al., 104

1960; Langford et al., 1974; McDonald et al., 1975) and induce a spread of cortical 105

activation (Pratt et al., 1999; Portas et al., 2000; Blume et al., 2016). In addition, studies 106

measuring event-related potentials (ERPs) during sleep, have demonstrated an N400 107

response to lexical-level semantic violations (Brualla et al., 1998; Bastuji et al., 2002; 108

Ibáñez et al., 2006) and residual neural signatures indicating semantic categorization of 109

words (Kouider et al., 2014; Andrillon et al., 2016). Notably, semantic-level responses 110

during sleep are often considerably attenuated and delayed in time compared to those 111

in wakefulness (Brualla et al., 1998; Perrin et al., 2002; Andrillon et al., 2016), raising 112

important questions as to the nature of residual high-level language processing during 113

sleep. 114

A main challenge for addressing this question is how to assess speech processing depth 115

in-lieu of behavioral metrics. Previous studies have resorted to measuring ERPs. 116

However, during NREM sleep, repetitive presentation of brief isolated stimuli often 117

elicits a large stereotypical response, known as a ‘K-complex’ (Colrain, 2005; Halász, 118

2016), that masks the precise neuronal dynamics and limits data interpretation. Indeed, 119

ERPs recorded during NREM sleep often differ substantially in their time-course and 120

morphology from those observed during wakefulness (Colrain and Campbell, 2007), 121

making it difficult to assess their functional significance. 122

Page 10: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

9

Here, we utilize the newly developed Concurrent Hierarchical Tracking (CHT) approach 123

(Ding et al., 2015), to assess the depth of speech processing during sleep. In CHT, stimuli 124

are structured in a manner that allows distinguishing neural responses to different levels 125

of linguistic analysis of continuous speech. Specifically, speech sequences are compiled 126

such that different linguistic levels (syllables, words, phrases and sentences) correspond 127

to distinct frequencies, allowing us to distinguish their neural signatures in brain activity, 128

while refraining from presentation of abrupt stimuli. We employed this method to test 129

whether hierarchical parsing of intelligible speech is preserved or disrupted during 130

sleep. This approach allows examining the capacity and limits of processing during sleep, 131

and may afford insights regarding the depth of speech processing that occurs without 132

attention more generally. We hypothesized that during sleep, the neural representation 133

of higher linguistic levels will be greatly diminished, as has been shown during 134

wakefulness for unintelligible or unattended speech (Zion Golumbic et al., 2013; Ding et 135

al., 2015). 136

137

Materials and Methods 138

Participants. Full-night sleep recordings were performed in 29 native Hebrew speakers 139

(16 females, mean age 28.7±3.6 years, range 22–38) who reported to be healthy, 140

without any history of neuropsychiatric or sleep disorders. The study was approved by 141

the Medical Institutional Review Board at the Tel Aviv Sourasky Medical Center. All 142

participants provided their written consent for participation. Participants underwent an 143

Page 11: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

10

interview determining their sleep habits and their propensity to fall asleep in noisy 144

environments. Eight participants (6 females) were excluded from analysis for either 145

technical issues (n=3) or lack of sufficient data (n=5 participants who kept falling asleep 146

during the wake sessions or experienced difficulties sleeping during the night). Twenty-147

one participants were included in the wake analysis (10 females, mean age 28.2±4.0), of 148

them 17 were included in the NREM sleep analysis (6 females, mean age 28.8±4.2), and 149

13 were included in the REM sleep analysis (5 females, mean age 28.6± 4.3). 150

Hebrew Materials. A bank of individually-recorded Hebrew syllables was used to create 151

a set of intelligible speech stimuli, as well as unintelligible (Scrambled) control stimuli. 152

Single syllables were uttered in random order by a human male voice and recorded at 153

44,100 Hz. The sound intensity was manually normalized using the Audacity® software 154

(version 2.0.5). In order to prevent biasing of speech perception by prosody, prosodic 155

cues were removed via pitch normalization using Praat© (version 5.4.04) (Boersma and 156

Weenink, 2016). The length of individual syllables (original mean duration 243.6±64.3 157

ms, range 168-397 ms) was adjusted to precisely 250 ms by truncation or silence 158

padding at the end. In case of truncation, a fading out effect was applied on the last 25 159

ms. Finally, syllables were concatenated using custom-written scripts in MATLAB (The 160

MathWorks) to form 25 Intelligible Hebrew sentences and 25 unintelligible (Scrambled) 161

sequences. All intelligible sentences were constructed to form hierarchical linguistic 162

structures as follows: every 2 syllables formed a 500 ms-long word, every 2 words 163

formed a 1000 ms-long phrase and every 2 phrases formed a 2000 ms-long sentence 164

(Figure 1A; Press here to hear an example of CHT speech stimuli). Since syllables were 165

Page 12: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

11

hierarchically grouped into linguistic constituents, with no additional acoustic gaps 166

inserted between them (Figure 1B), the linguistic structures appeared at fixed 167

periodicities throughout the stimuli (syllables at 4 Hz; words at 2 Hz; phrases at 1 Hz and 168

sentences at 0.5 Hz). Sentences did not include rhymes, passive form of verbs or 169

arousing semantic content. The intelligibility of Hebrew materials was verified in a pilot 170

study, demonstrating that all Hebrew sentences could be fully repeated after a mean of 171

1.28 presentations, and most (78.7%±15.8) sentences could be fully repeated by 172

participants after a single presentation. The same Hebrew syllables were scrambled to 173

compose 25 unintelligible pseudo-sentences. Scrambling was performed by shuffling 174

syllables across sentences while maintaining their original position within the sequence. 175

Thus, syllables that tend to occur at the beginning/end of words in natural language, 176

retain this position in the control condition. We also verified that the scrambling 177

procedure did not produce any real Hebrew words by chance. 178

Chinese Materials. We used sentences in Chinese as an unintelligible-speech control 179

condition. The Chinese materials were constructed and used by Ding et al. (2015) in a 180

similar manner as the Hebrew materials, and were comprised of sequences of individual 181

syllables lasting 250 ms, uttered by a computerized male voice (see Stimuli I in Ding et 182

al., 2015). The Chinese materials, too, were compiled such that higher linguistic 183

structures appeared at fixed frequencies. However, since the Chinese sentences were 184

comprised of mono-syllabic words, the syllabic rate of 4 Hz also represented the word 185

rate. Consequently, only two additional linguistic levels were formed - phrases at 2 Hz 186

and sentences at 1 Hz. The fact that the Hebrew materials included an additional 187

Page 13: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

12

linguistic level should not have interfered with our experimental design, since none of 188

the participants understood Chinese, and it was only used as an unintelligible-speech 189

control condition. 190

Stimulus presentation. A trial lasted 12 seconds and was composed of six concatenated 191

2 sec-long sentences of the same speech type (Intelligible, Scrambled or Foreign speech; 192

Figure 1D). In the first 5 seconds of each trial, sound intensity increased from zero to full 193

intensity to prevent abrupt onsets. To assure that the sound intensity of the stimuli 194

fluctuated at the syllabic rate of 4 Hz, the power spectrum of the mean sound intensity 195

in each condition was calculated across 50 trials for each condition (Figure 1C). Inter-196

trial intervals (ITIs) were distributed pseudo-randomly between 1.5 and 4 seconds to 197

avoid expectation effects. In addition to the 3 speech conditions, we presented a 198

baseline Sham condition where no sound was presented. Typically, during a full 199

experiment we presented 200-500 trials (mean 406±87) for each condition (Intelligible, 200

Scrambled, Foreign and Sham). Trials were not presented in a discrete block design but 201

rather in a continuous manner throughout the experiment. Roughly 20% of the trials 202

were presented during wakefulness, 10% during REM sleep, 65% during NREM sleep and 203

additional 5% were presented during state transitions (excluded). Only the final 8 204

seconds of each trial were used for analysis, to further avoid onset responses (Figure 205

1D). 206

Experimental design. The experiment took place in an acoustically-attenuated and 207

electrically-shielded sleep lab. Auditory stimulation was delivered through loudspeakers, 208

Page 14: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

13

and sound intensity was adjusted to a convenient level that remained constant 209

throughout the night (see ‘Auditory stimulation’ below). Hebrew sentences were 210

confirmed to be intelligible prior to the commencement of the experiment (see 211

‘Intelligibility test’ below). Experiments included stimulation during wakefulness in the 212

evening, stimulation during overnight sleep, and another block of stimulation in the 213

morning after spontaneous awakening (Figure 3A). The room was dark throughout the 214

entire experiment. During the wakefulness sessions, participants sat on a chair rather 215

than lying in bed, to ensure they stay awake and if any sign of falling asleep occurred 216

(such as rolling eye movements, EEG slowing or appearance of sleep spindles or K-217

complexes), participants were woken up immediately (3.85 ± 3.21, range 0-11 218

awakenings per participant). In order to minimize differences between wake and sleep 219

sessions, the speakers were raised to ear-height keeping the same distance as when 220

lying in bed, and participants were instructed to keep their eyes closed and were not 221

required to perform any explicit task. After the evening wakefulness session, 222

participants were allowed to fall asleep at their own convenience. Auditory stimulation 223

was paused manually whenever awakening or movement was detected, and resumed 224

shortly after detection of unequivocal sleep activity in the EEG. 225

Auditory stimulation. Auditory stimulation was delivered through speakers situated on 226

both sides of the bed (during sleep) or on both sides of the chair (during wakefulness). 227

Sound intensity was individually adjusted to a relatively low level (range 42.1-45.9 dB 228

SPL) that allowed speech comprehension. Importantly, sound intensity was adjusted 229

Page 15: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

14

prior to the first experimental session and remained constant throughout all conditions 230

for each participant. 231

Intelligibility test. Participants performed an intelligibility-test to verify that they 232

understood the materials used in the Intelligible speech condition. All intelligible and 233

scrambled sentences were presented in random order and participants were asked to 234

report whether they contained meaningful speech or not and were then asked to repeat 235

them. 236

Data Acquisition. High-density EEG was recorded continuously using a 256-channel 237

hydrocel geodesic sensor net with passive electrodes (Electrical Geodesics, Inc. system 238

[EGI]). Each carbon-fiber electrode consists of a silver-chloride carbon fiber pellet, a lead 239

wire, a gold-plated pin, and was injected with conductive gel (Electro-Cap International 240

[ECI]). Signals were referenced to Cz, amplified via an AC-coupled high-input impedance 241

amplifier (NetAmps 300, EGI), and digitized at 1000 Hz. Electrode impedance in all 242

sensors was verified to be less than 50 kΩ before starting the recording. 243

Sleep scoring. Sleep scoring in 30s epochs was performed manually according to 244

established guidelines of the American Academy of Sleep Medicine (Iber et al., 2007) 245

based on EEG, EOG, EMG and video. To this end, EEG data from F3/F4, C3/C4 and O1/O2 246

were referenced to the contralateral mastoid and two EOG channels were referenced to 247

the other mastoid. Scoring channels were visualized along with synchronized EMG in 30s 248

epochs. Sleep scoring was further verified by inspecting the time-frequency 249

representation (spectrogram) of the Pz electrode (not involved in scoring process) 250

Page 16: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

15

superimposed with the hypnogram (as in Figure 3A,B). Each epoch was categorized as 251

N1/N2/N3/REM sleep or wakefulness. N1 sleep stage epochs (13.4±1.4% of sleep time, 252

range 7.6-30.1%) were excluded from further analysis to avoid uncertainty regarding 253

precise sleep onset. 254

EEG pre-processing. Preprocessing was performed in MATLAB (The MathWorks) using 255

the FieldTrip toolbox (Oostenveld et al., 2011) and custom-written scripts. Data were 256

segmented (-2 to +14 sec) around stimulus onset, down-sampled to 250 Hz, re-257

referenced to the average signal of the mastoids, and linearly de-trended. Bad 258

electrodes (<14% in all participants) were identified as those whose variance and 259

maximal absolute value constituted outliers relative to adjacent electrodes upon visual 260

inspection per participant and per state, and were replaced with the weighted average 261

of their neighbors using a linear, distance weighted interpolation. Outlier trials 262

(15.3±0.3%) were manually identified by visual inspection and were discarded from 263

subsequent analysis. In the N2 dataset, we additionally excluded trials containing K-264

complex events. K-complexes were detected automatically as those trials in which the 265

raw EEG amplitude was both higher than +40 μV and lower than -40 μV within a 2.4 sec 266

window, sliding at a resolution of 40 ms. Independent component analysis (ICA) was 267

used for removal of eye movement and heartbeat traces, separately for each state. 268

After cleaning, we randomly selected the same number of trials from all conditions 269

(Intelligible, Scrambled, Foreign and Sham) for each subject separately. Participants with 270

fewer than 30 clean trials in any single condition were discarded from further analysis. 271

Page 17: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

16

Data analysis during wakefulness. Data were analyzed by two complementary 272

approaches. First, Inter-Trial Phase-Coherence (ITPC) was calculated as follows: The Fast 273

Fourier Transform (FFT) was calculated separately for each trial with 0.125 Hz 274

resolution. Next, the phase component at each frequency was used to calculate the 275

ITPC, which is the sum (absolute value) of the phases across trials: 276

Note that ITPC represents phase consistency across trials, which is the inverse of 277

response variability across trials (Berens, 2009). 278

Second, evoked power spectrum analysis was performed as follows: We averaged the 279

clean preprocessed trials for each condition separately and computed the power 280

spectrum of the average using FFT with 0.125 Hz resolution. We normalized the power 281

at each frequency by subtracting the mean power level at adjacent frequencies within 282

±0.125 Hz (Nozaradan et al., 2011). 283

Comparison of sleep and wakefulness. To reduce the effects of widespread slow waves 284

during sleep (Figure 3C) that could preclude analysis of low frequencies of interest (0.5, 285

1 and 2 Hz), we applied a spatial current source density transformation (CSD, also 286

known as surface Laplacian) (Kayser and Tenke, 2015) to the preprocessed EEG data. 287

Indeed, EEG spectral power after CSD transformation revealed comparable energy at 288

slow (<4Hz) frequencies across wakefulness, N2 sleep, and REM sleep (Figure 3D) 289

attesting to the utility of this procedure in minimizing the potential effect of slow 290

Page 18: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

17

ongoing sleep activities. However, N3 sleep data were still dominated by robust slow 291

wave activity (Figure 3D) and therefore were excluded from analysis. Furthermore, for 292

comparing wakefulness and sleep states we focused on ITPC analysis, since phase 293

consistency is less affected than power by ongoing slow activities. Importantly, identical 294

procedures were applied across all states of sleep and wakefulness (Figure 4). Finally, 295

when comparing the results in each sleep stage to wakefulness, we used the same 296

participants for each comparison, and randomly selected an equal number of trials 297

across states to ensure similar statistical power. Topographical distribution of ITPC 298

values in Figure 4 was calculated for Intelligible speech after normalizing ITPC at each 299

frequency by subtracting the mean ITPC at adjacent frequencies within ±0.125 Hz. 300

SNR estimation. SNR (used as a covariate in subsequent ANCOVA tests) was quantified 301

in each state and subject separately, by calculating the ratio between the standard 302

deviation (SD) of the wide-band ERP evoked by intelligible speech and the SD of the 303

wide-band ERP of ongoing activity (sham condition), over the 8 seconds used for 304

analysis: 305

SNR = S of ERPSD of ERP

306

Statistical analysis. Statistical analysis was performed using custom-written Matlab 307

scripts (The Mathworks) and SPSS software (version 23.0) (IBM Corp., 2015). Statistical 308

analyses focused on the average response within a pre-defined mid-central region of 309

interest (ROI; see inset in Figure 3A), which is typical of auditory responses in EEG 310

(Picton et al., 1974). The ROI included 92 electrodes that lay within a 6 cm radius from 311

Page 19: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

18

Cz. Fisher’s z-transformation was applied to individual ITPC values before statistical 312

analysis. Hypothesis testing consisted of independent comparisons between each of the 313

speech conditions (Intelligible, Scrambled and Foreign language) and the Sham (no 314

stimulation) condition. Hypotheses were tested via paired t-tests for all comparisons, 315

after verifying normality via Kolmogorov-Smirnov tests. Otherwise, a Wilcoxon rank sum 316

test was used. To ensure that effects in wakefulness were exclusive for the a-priori 317

frequencies of interest (0.5, 1, 2 and 4 Hz) we statistically tested the initial ITPC results 318

on a wider range of 12 frequencies (every 0.5 Hz between 0.5-6 Hz). To account for 319

multiple comparisons (3 speech conditions vs. sham X 12 frequencies = 36 comparisons), 320

we controlled the false discovery rate (FDR) using a q value of 0.05 (Benjamini and 321

Yekutieli, 2001). After confirming that significant responses in wakefulness are only 322

observed at the a-priori frequencies of interest, subsequent comparisons of sleep and 323

wakefulness were restricted to those frequencies, and corrected for 12 comparisons (3 324

speech conditions vs. Sham X 4 frequencies) via FDR correction at q=0.05. 325

Given differences in ongoing spontaneous activity across states (which was substantially 326

reduced by the CSD transformation employed here, but not entirely equated; see Figure 327

3D), direct comparisons across arousal states were performed via two-way repeated 328

measures ANCOVA tests, while including the SNR at each state as a covariate. This 329

analysis focused on the Intelligible speech condition, comparing the response at each 330

frequency of interest (1,2,4 Hz) across states (wakefulness vs. REM / NREM). We also 331

tested whether the response at the acoustic/syllabic level (4Hz), differed across states 332

(wakefulness vs. sleep) and speech conditions (Intelligible, Scrambled, Foreign speech). 333

Page 20: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

19

Note that these ANCOVAs were performed separately when comparing wakefulness vs. 334

REM / NREM, given the lower number of participants with REM sleep. 335

336

Results 337

Neuronal speech tracking during wakefulness is evident in the entire acoustic-linguistic 338

hierarchy during intelligible speech 339

To validate the utility of the CHT paradigm with scalp EEG, we first assessed the ITPC of 340

cortical activity during wakefulness, across a range of 12 frequencies between 0.5 to 6 341

Hz (0.5 Hz intervals) within a mid-central ROI. ITPC at the acoustic/syllabic rate of 4 Hz 342

was significant in all speech conditions (all t-tests vs. sham condition p< ; remained 343

significant after FDR correction at q=0.05; See Table 1 for detailed p-values; Figure 2A). 344

ITPC at frequencies corresponding to higher linguistic structures (2, 1 and 0.5 Hz) was 345

significant only for the intelligible speech condition (word level: p<0.001; phrase level: 346

p=0.007; sentence level: p=0.001; remained significant after FDR correction; see Figure 347

2A). None of the unintelligible speech conditions elicited significant ITPC at any of these 348

frequencies (Table 1). Importantly, significant ITPC was only observed at the frequencies 349

of interest corresponding to pre-determined linguistic structures in the stimuli (p>0.06 350

at all other frequencies). 351

We repeated the same analysis for the EEG evoked power within the same central ROI 352

(Figure 2B). Similar to ITPC, normalized EEG power at the acoustic/syllabic rate of 4 Hz 353

4 -10

Page 21: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

20

was significant in all speech conditions (p<0.002; remained significant after FDR 354

correction), whereas significant power at frequencies representing linguistic parsing was 355

observed only for intelligible speech for most linguistic levels (sentence level: p=0.004; 356

phrase level: p=0.012; remained significant after FDR correction; word level: p=0.21, 357

n.s.). Corresponding scalp topographies were in line with those observed for ITPC 358

(Figure 2B, bottom). 359

Overall, our ITPC and power analyses during wakefulness converge with previous results 360

(Ding et al., 2015), indicating that, for intelligible speech, neural tracking is evident 361

throughout the linguistic hierarchy, whereas for unintelligible speech neural responses 362

can be attributed to the acoustic structure. 363

Sleep preserves auditory responses but disrupts high-order linguistic parsing 364

Continuous overnight recordings lasted 9 hours and 34 minutes on average (±75 min). 365

We verified that normal sleep was preserved in the presence of speech stimulation. 366

Figure 3 illustrates the sleep architecture and spectral content of EEG activity, 367

demonstrating all the established hallmarks of the different vigilance states, including 368

alpha (8-10 Hz) activity during quiet wakefulness, sleep spindle/sigma (10-15 Hz) and 369

slow wave (<4 Hz) activities during N2/N3 sleep, and diffuse theta (6-9 Hz) during REM 370

sleep. Awakenings associated with speech stimulation were rare (and even if such 371

events were present and not observed, these trials would be tagged as wake or N1 372

sleep, so any differences reported here would constitute a lower bound). Sleep 373

parameters (Figure 3E) at the group level were in accordance with typical values for 374

Page 22: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

21

healthy young adults (Carskadon and Dement, 2011). In the morning debriefing, all 375

participants reported being well-rested (slept very well, not tired). Most participants 376

vaguely recalled hearing the stimuli a few times after going to sleep, but all confirmed 377

that this did not interfere with sleep quality, as they returned back to sleep 378

immediately. Thus, intermittent speech stimulation did not exert significant effects on 379

sleep architecture or subjective measures. 380

We proceeded to re-analyze EEG activity as a function of sleep and wakefulness states, 381

after applying a CSD spatial filter to reduce the effects of slow-wave activity during sleep 382

(see Methods and Figure 4). We focused our analysis on comparing wakefulness with N2 383

sleep (n=17) and comparing wakefulness with REM sleep (n=13), equating the number 384

of trials and participants in each comparison to ensure identical statistical power. We 385

did not analyze responses during N1 sleep, as it was ambiguous and rare (13.4±1.4% of 386

sleep time) or during N3 sleep (where slow wave activity precluded analysis, see 387

Methods). 388

Our re-analysis of the wakefulness data (performed after CSD in the subset of 389

participants with adequate sleep data) confirmed that, even in these reduced datasets, 390

all speech conditions elicited a significant acoustic/syllabic-rate response at 4 Hz (p< 391

for both datasets). Moreover, intelligible speech ITPC remained significant at the word 392

and phrase levels in these reduced datasets (Figure 4 top; Table 1). However, the 393

sentential rate of 0.5 Hz was no longer significant in either subset, probably due to the 394

lower power of this analysis and/or the removal of low-frequencies by the CSD. 395

3 -10

Page 23: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

22

In the sleep conditions, significant ITPC at the acoustic/syllabic rate of 4 Hz was 396

observed in both sleep states for all stimuli (NREM: p< ; REM: p< ; remained 397

significant after FDR correction). Furthermore, we directly compared the 4Hz 398

(acoustic/syllabic) response for all audible conditions during wakefulness vs. each sleep 399

state, using repeated measures ANCOVA with SNR differences (sleep-wake) as a 400

covariate. This comparison did not reveal significant differences in the acoustic/syllabic 401

response across states [wakefulness vs. REM: F(2,10)=0.063, p=0.940, n.s.; wakefulness 402

vs. NREM: F(2,14)=0.624, p=0.550, n.s.], suggesting no significant change in the acoustic 403

response across states. 404

Importantly, despite robust activity at the acoustic level, none of the speech conditions 405

during sleep elicited significant ITPC response vs. Sham, at frequencies corresponding to 406

high-level linguistic structures (Table 1). In addition, we compared the responses to 407

Intelligible speech during wakefulness vs. each sleep state at frequencies of interest 408

(1,2,4Hz), using ANCOVA with SNR difference as a covariate. This analysis revealed 409

significant state X frequency interaction effects [wakefulness vs. REM: F(2,10)=5.21, 410

p=0.028; wakefulness vs. NREM: F(2,14)=4.35, p=0.034], in line with the comparisons of 411

each speech condition vs. the sham condition performed within each state (Table 1). 412

Altogether, our results suggest that while both NREM and REM sleep preserve acoustic 413

responses, neural tracking of higher-order linguistic levels within a speech-stream is not 414

evident. 415

416

5 -10 3 -10

Page 24: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

23

Discussion 417

The main novel result reported here is that during sleep, basic neural encoding of 418

acoustic features of speech persists, while parsing of higher-order linguistic structures, 419

evident for intelligible speech during wakefulness - is disrupted. These findings further 420

our understanding of the capacity and limits of cortical processing during sleep, and 421

demonstrate the effectiveness of the CHT approach for probing high-level cognitive 422

processes covertly in unresponsive states. 423

Studying linguistic processing during wakefulness using CHT 424

Our results in wakefulness are in line with Ding et al. (2015), showing that cortical 425

activity concurrently tracks the time-course of linguistic structures at multiple 426

hierarchical levels when speech is intelligible. For unintelligible speech, cortical 427

responses are only evident at the acoustic/syllabic rate. Our results extend the original 428

findings in several ways. First, using bi-syllabic words allowed distinguishing between 429

the syllabic (4 Hz) and word rate (2 Hz), affording further separation between the 430

acoustic and linguistic aspects of speech. Second, our design did not include an explicit 431

task, demonstrating that passive listening is sufficient for revealing hierarchical neural 432

tracking of linguistic structures. Third, our results demonstrate that one can effectively 433

quantify hierarchical linguistic parsing with non-invasive, inexpensive, portable and 434

readily available EEG. These extensions pave the way for using the CHT approach in a 435

Page 25: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

24

wide array of settings, ranging from language acquisition to studies of residual cognitive 436

processing in various clinical populations. 437

Depth of speech processing during sleep 438

The current findings make a substantial contribution to understanding the extent of 439

language processing during sleep. We find that the neural response at the syllabic rate, 440

which likely reflects basic auditory representation of the acoustic envelope, was 441

comparable across wakefulness, REM and NREM sleep. This is in line with previous 442

studies supporting preserved responses in low-level auditory cortex during sleep (Issa 443

and Wang, 2008; Nir et al., 2015). At the same time, parsing of higher-order linguistic 444

structures is disrupted during sleep, implying the existence of a functional ‘bottleneck’ 445

precluding full processing of continuous speech. 446

Before discussing broader implications, we address an important methodological 447

caveat: Is the lack of observable low-frequency peaks during sleep due to increased 448

background noise leading to poor sensitivity, rather than a lack of high-level speech 449

parsing? This potential criticism applies mainly to NREM sleep, as spontaneous activity 450

during REM sleep is wake-like. Although applying CSD transformation effectively 451

reduced slow widespread activity in NREM sleep (Figure 3D), it was nevertheless still 452

higher than in wakefulness. However, we believe that this does not account for our 453

results, for two reasons. First, our experimental design deliberately focused on testing 454

different speech conditions within each state, where SNR and other signal properties 455

are matched. Second, we conducted several analyses taking into account SNR 456

Page 26: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

25

differences across states. These showed that while the 4Hz acoustic responses did not 457

differ significantly across states, responses at the word and sentence level were found 458

for Intelligible speech only during wakefulness, as indicated by significant interactions 459

between state X frequency. Nevertheless, we acknowledge that given the dominance of 460

low-frequency background activity during NREM sleep (as in other non-responsive 461

states such as anesthesia and vegetative states), results should be interpreted with 462

great caution and appropriate controls applied in future studies. 463

What is the nature of the functional bottleneck observed during sleep between basic 464

auditory processing and high-level linguistic parsing? Previous ECoG results using the 465

CHT paradigm attribute phrasal and sentential responses to non-sensory areas, involved 466

in language processing, e.g. the left IFG and bilateral TPJ regions. In contrast, syllabic 467

rate responses were found primarily in auditory cortex (Ding et al., 2015). The lack of 468

evidence for neural tracking of high-level speech structures during sleep suggests that it 469

disrupts efficient signal propagation from auditory cortex to higher cortical regions. 470

Along these lines, inter-cortical connectivity during sleep is restricted upon brief 471

electromagnetic perturbation (Massimini et al., 2005, 2007; Pigorini et al., 2015). 472

Similarly, several functional magnetic resonance imaging (fMRI) studies demonstrated 473

robust attenuation of speech responses in IFG and frontal regions during sleep (Portas 474

et al., 2000; Dehaene-Lambertz et al., 2002; Wilf et al., 2016). Together with the current 475

findings, these studies imply that speech processing during sleep is limited to low-level 476

acoustic processing, with comparable neural responses for intelligible and unintelligible 477

speech. 478

Page 27: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

26

However, other studies report evidence of residual semantic processing during sleep 479

upon presentation of single words (Kouider et al., 2014; Andrillon et al., 2016), word-480

pairs (Brualla et al., 1998) and short sentences (Ibáñez et al., 2006; Daltrozzo et al., 481

2012), although such processing is generally weaker and with altered time dynamics 482

compared to wakefulness (Brualla et al., 1998; Perrin et al., 2002). In attempt to 483

reconcile these seemingly contradictory results, we suggest that the functional 484

bottleneck for speech processing during sleep is not semantic analysis per-se, but rather 485

the mediating process of segmentation. One key difference between this and previous 486

studies is the use of continuous speech, rather than single words or short sentences. 487

This not only allowed us to overcome serious methodological limitations – such as the 488

pervasiveness of K-complexes – but also to tap into the process of word segmentation, 489

which constitutes a prerequisite for continuous speech comprehension. Indeed, speech 490

processing relies critically on accurately parsing the ongoing acoustic stream input into 491

discrete and meaningful linguistic units (Hickok et al., 1993; Mattys, 1997; Giraud and 492

Poeppel, 2012; Doelling et al., 2014). Parsing itself is a hierarchical highly-demanding 493

cognitive process (Greenberg et al., 2004; Ghitza, 2012) that involves matching bottom-494

up cues with existing lexical, syntactic and semantic representations (Traxler, 2014), as 495

well as predictive and anticipatory processes (Arnal et al., 2011; DeLong et al., 2014). 496

The current findings suggest that, at minimum, it is the process of word segmentation 497

that is compromised during sleep, a process that requires ongoing analysis of the 498

continuous global stream of input (Strauss et al., 2015; Tononi et al., 2016). It thus 499

remains possible that when phrases or sentences are presented in a discrete format (as 500

Page 28: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

27

in most previous studies), individual words can be identified and processed fully up to a 501

semantic level, even during sleep. 502

Another critical aspect of continuous speech parsing is integrating information across 503

multiple time-scales (Rosen, 1992; Greenberg et al., 2003; Ghitza, 2012; Zion Golumbic 504

et al., 2013), allowing short-scale information (phonemes/syllables) to be combined to 505

create lexical units (e.g. words) and higher-order structures (phrases and sentences). 506

Such integration requires short-term memory buffers to sustain information for longer 507

durations, i.e. to handle larger 'temporal receptive windows' (Lerner et al., 2011; Luo 508

and Poeppel, 2012; Chait et al., 2015). Notably, previous studies have shown that neural 509

activity during sleep is restricted to short timescales in the range of hundreds of 510

milliseconds, preventing the emergence of long-lasting causal interactions (Pigorini et 511

al., 2015; Strauss et al., 2015). The proposed restriction on brain activity during sleep to 512

short time intervals may also contribute to the diminished hierarchical parsing of 513

continuous speech observed here. 514

One important limitation of the current study is that prosodic cues were purposefully 515

removed from the speech material, allowing us to probe grammatical speech parsing 516

irrespective of correlated acoustic variations. However, since syntactic parsing of natural 517

speech also benefits from prosody (Eckstein and Friederici, 2006), it remains to be 518

tested whether prosody can facilitate continuous speech parsing during sleep. 519

Sleep and Inattention 520

Page 29: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

28

Analogous results to those found here during sleep have recently been reported for 521

unattended speech, for which neural tracking is robust in auditory cortex, but 522

substantially reduced in higher-order language areas (Zion Golumbic et al., 2013; 523

Rimmele et al., 2015). Along this line, acoustic features of unattended speech affect 524

behavior more than high-level (semantic) features (Ellermeier et al., 2015; Wöstmann 525

and Obleser, 2016). Additional similarities between sleep and inattention exist at the 526

behavioral level, where unattended speech typically cannot be overtly recalled (Lachter 527

et al., 2004), although some personally relevant words occasionally capture attention 528

(Cherry, 1953; Moray, 1958, 1959; Bentin et al., 1995; Wood and Cowan, 1995), as also 529

found during sleep. These similarities between speech processing during sleep and 530

inattention may carry broader implications as to the nature of functional ‘bottlenecks’ in 531

speech processing (Broadbent, 1958; Treisman, 1969), constituting an intriguing topic 532

for future research. 533

534

Conclusions 535

To the best of our knowledge, this study is the first to investigate hierarchical parsing of 536

continuous speech during sleep. We found that bottom-up auditory processing is 537

preserved in sleep and comparable to that found in wakefulness. In sharp contrast, 538

neural tracking of high-order linguistic structures – words, phrases and sentences - is 539

disrupted in sleep, in a manner similar to unattended or unintelligible speech during 540

wakefulness. Current results suggest that parsing of continuous speech - which requires 541

integration across multiple time scales, matching of bottom-up input to stored linguistic 542

Page 30: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

29

representations, and top-down predictive coding - may not be possible without overt 543

attention and consciousness. Our results imply a functional barrier between auditory 544

sensation and linguistic processing, a barrier that may be essential to ensure 545

preservation of sleep in the face of external events and support its functions. This study 546

sets the ground towards studying residual speech processing across states of 547

consciousness, anesthesia, neurodegeneration, development and language disorders. 548

549

Page 31: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

30

References 550

Andrillon T, Nir Y, Cirelli C, Tononi G, Fried I (2015) Single-neuron activity and eye 551

movements during human REM sleep and awake vision. Nat Commun 6:7884 552

Available at: http://www.nature.com/doifinder/10.1038/ncomms8884. 553

Andrillon T, Poulsen AT, Hansen LK, Léger D, Kouider S (2016) Neural Markers of 554

Responsiveness to the Environment in Human Sleep. J Neurosci 36:6583–6596 555

Available at: http://www.ncbi.nlm.nih.gov/pubmed/27307244. 556

Arnal LH, Wyart V, Giraud A-L (2011) Transitions in neural oscillations reflect prediction 557

errors generated in audiovisual speech. Nat Neurosci 14:797–801. 558

Bastuji H, Perrin F, Garcia-Larrea L (2002) Semantic analysis of auditory input during 559

sleep: studies with event related potentials. Int J Psychophysiol 46:243–255. 560

Benjamini Y, Yekutieli D (2001) The control of the false discovery rate in multiple testing 561

under dependency. Ann Stat 29:1165–1188. 562

Bentin S, Kutas M, Hillyard SA (1995) Semantic processing and memory for attended and 563

unattended words in dichotic listening: behavioral and electrophysiological 564

evidence. J Exp Psychol Hum Percept Perform 21:54–67 Available at: 565

http://www.ncbi.nlm.nih.gov/pubmed/7707033. 566

Berens P (2009) CircStat: A MATLAB toolbox for circular statistics. J Stat Softw 31:1–21. 567

Blume C, del Giudice R, Lechinger J, Wislowska M, Heib DPJ, Hoedlmoser K, Schabus M 568

Page 32: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

31

(2016) Preferential processing of emotionally and self-relevant stimuli persists in 569

unconscious N2 sleep. Brain Lang. 570

Boersma P, Weenink D (2016) Praat: doing phonetics by computer [Computer program]. 571

Version 5.4.04, retrieved 30 December 2014. 572

Broadbent DE (1958) Selective listening to speech. Percept Commun:11–35 Available at: 573

http://psycnet.apa.orgbooks/10037/002. 574

Brualla J, Romero MF, Serrano M, Valdizán JR (1998) Auditory event-related potentials 575

to semantic priming during sleep. Electroencephalogr Clin Neurophysiol 108:283–576

290. 577

Carskadon MA, Dement WC (2011) Normal Human Sleep : An Overview. In: Principles 578

and practice of sleep medicine, pp 16–26. 579

Chait M, Greenberg S, Arai T, Simon JZ, Poeppel D (2015) Multi-time resolution analysis 580

of speech: evidence from psychophysics. Front Neurosci 9:214. 581

Cherry EC (1953) Some experiments on the recognition of speech, with one and with 2 582

ears. J Acoust Soc Am 25:975–979. 583

Colrain IM (2005) The K-complex: a 7-decade history. Sleep 28:255–273. 584

Colrain IM, Campbell KB (2007) The use of evoked potentials in sleep research. Sleep 585

Med Rev 11:277–293. 586

Page 33: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

32

Daltrozzo J, Claude L, Tillmann B, Bastuji H, Perrin F (2012) Working Memory Is Partially 587

Preserved during Sleep. PLoS One 7. 588

Dehaene-Lambertz G, Dehaene S, Hertz-Pannier L (2002) Functional neuroimaging of 589

speech perception in infants. Science 298:2013–2015. 590

DeLong KA, Troyer M, Kutas M (2014) Pre-processing in sentence comprehension: 591

Sensitivity to likely upcoming meaning and structure. Lang Linguist Compass 8:631–592

645. 593

Ding N, Melloni L, Zhang H, Tian X, Poeppel D (2015) Cortical tracking of hierarchical 594

linguistic structures in connected speech. Nat Neurosci 19:158–164 Available at: 595

http://www.ncbi.nlm.nih.gov/pubmed/26642090. 596

Doelling KB, Arnal LH, Ghitza O, Poeppel D (2014) Acoustic landmarks drive delta–theta 597

oscillations to enable speech comprehension by facilitating perceptual parsing. 598

Neuroimage 85:761–768. 599

Eckstein K, Friederici AD (2006) It’s early: event-related potential evidence for initial 600

interaction of syntax and prosody in speech comprehension. J Cogn Neurosci 601

18:1696–1711 Available at: http://www.mitpressjournals.org.bibliotheque-602

nomade2.univ-lyon2.fr/doi/abs/10.1162/jocn.2006.18.10.1696#.Vs3BxpwrLWI. 603

Ellermeier W, Kattner F, Ueda K, Doumoto K, Nakajima Y (2015) Memory disruption by 604

irrelevant noise-vocoded speech: Effects of native language and the number of 605

frequency bands. J Acoust Soc Am 138:1561–1569 Available at: 606

Page 34: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

33

http://asa.scitation.org/doi/10.1121/1.4928954. 607

Friederici AD, Singer W (2015) Grounding language processing on basic 608

neurophysiological principles. Trends Cogn Sci 19:329–338 Available at: 609

http://dx.doi.org/10.1016/j.tics.2015.03.012. 610

Ghitza O (2012) On the Role of Theta-Driven Syllabic Parsing in Decoding Speech: 611

Intelligibility of Speech with a Manipulated Modulation Spectrum. Front Psychol 612

3:238. 613

Giraud A-L, Poeppel D (2012) Cortical oscillations and speech processing: emerging 614

computational principles and operations. Nat Neurosci 15:511–517. 615

Greenberg S, Ainsworth W a, Popper AN, Fay RR (2004) Speech Processing in the 616

Auditory System. New York: Springer-Verlag. 617

Greenberg S, Carvey H, Hitchcock L, Chang S (2003) Temporal properties of spontaneous 618

speech—a syllable-centric perspective. J Phon 31:465–485. 619

Halász P (2016) The K-complex as a special reactive sleep slow wave - A theoretical 620

update. Sleep Med Rev 29:34–40. 621

Hickok G, Zurif E, Cansecogonzalez E (1993) Structural Description of Agrammatic 622

Comprehension. Brain Lang 45:371–395. 623

Ibáñez A, López V, Cornejo C (2006) ERPs and contextual semantic discrimination: 624

Degrees of congruence in wakefulness and sleep. Brain Lang 98:264–275 Available 625

Page 35: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

34

at: http://linkinghub.elsevier.com/retrieve/pii/S0093934X06001106. 626

Iber C, Ancoli-Israel S, Chesson A, Quan S (2007) The AASM manual for the scoring of 627

sleep and associated events: rules, terminology and technical specifications, 1st ed. 628

Westchester, Illnois: American Academy of Sleep Medicine. 629

IBM Corp. (2015) IBM SPSS Statistics for Windows, Version 23.0. 2015. 630

Issa EB, Wang X (2008) Sensory responses during sleep in primate primary and 631

secondary auditory cortex. J Neurosci 28:14467–14480 Available at: 632

http://www.ncbi.nlm.nih.gov/pubmed/19118181. 633

Kayser J, Tenke CE (2015) On the benefits of using surface Laplacian (current source 634

density) methodology in electrophysiology. Int J Psychophysiol 97:171–173 635

Available at: http://linkinghub.elsevier.com/retrieve/pii/S0167876015002111. 636

Kouider S, Andrillon T, Barbosa LSS, Goupil L, Bekinschtein TAA (2014) Inducing Task-637

Relevant Responses to Speech in the Sleeping Brain. Curr Biol 24:2208–2214. 638

Lachter J, Forster KI, Ruthruff E (2004) Forty-Five Years After Broadbent (1958): Still No 639

Identification Without Attention. Psychol Rev 111:880–913 Available at: 640

http://doi.apa.org/getdoi.cfm?doi=10.1037/0033-295X.111.4.880. 641

Langford GW, Meddis R, Pearson AJD (1974) Awakening Latency From Sleep For 642

Meaningful and Non-Meaningful Stimuli. Psychophysiology 11:1–5. 643

Lerner Y, Honey CJ, Silbert LJ, Hasson U (2011) Topographic mapping of a hierarchy of 644

Page 36: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

35

temporal receptive windows using a narrated story. J Neurosci 31:2906–2915. 645

Luo H, Poeppel D (2012) Cortical oscillations in auditory perception and speech: 646

evidence for two temporal windows in human auditory cortex. Front Psychol 3:170. 647

Massimini M, Ferrarelli F, Esser SK, Riedner BA, Huber R, Murphy M, Peterson MJ, 648

Tononi G (2007) Triggering sleep slow waves by transcranial magnetic stimulation. 649

Proc Natl Acad Sci U S A 104:8496–8501 Available at: 650

http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=1895978&tool=pmcen651

trez&rendertype=abstract. 652

Massimini M, Ferrarelli F, Huber R, Esser SK, Singh H, Tononi G (2005) Breakdown of 653

cortical effective connectivity during sleep. Science (80- ) 309:2228–2232. 654

Mattys SL (1997) The use of time during lexical processing and segmentation: A review. 655

Psychon Bull Rev 4:310–329. 656

McDonald DG, Schicht WW, Frazier RE, Shallenberger HD, Edwards DJ (1975) Studies of 657

information processing in sleep. Psychophysiology 12:624–629. 658

Moray N (1958) The Effect of the Relative Intensities of Dichotic Messages in Speech 659

Shadowing. Lang Speech 1:110–113 Available at: 660

http://las.sagepub.com/content/1/2/110.abstract. 661

Moray N (1959) Attention in dichotic listening: Affective cues and the influence of 662

instructions. Q J Exp Psychol 11:56–60. 663

Page 37: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

36

Nir Y, Tononi G (2010) Dreaming and the brain: from phenomenology to 664

neurophysiology. Trends Cogn Sci 14:88–100. 665

Nir Y, Vyazovskiy V V., Cirelli C, Banks MI, Tononi G (2015) Auditory responses and 666

stimulus-specific adaptation in rat auditory cortex are preserved across NREM and 667

REM sleep. Cereb Cortex 25:1362–1378. 668

Nozaradan S, Peretz I, Missal M, Mouraux A (2011) Tagging the neuronal entrainment to 669

beat and meter. J Neurosci 31:10234–10240 Available at: 670

http://www.ncbi.nlm.nih.gov/pubmed/21753000. 671

Oostenveld R, Fries P, Maris E, Schoffelen J-M (2011) FieldTrip: Open Source Software 672

for Advanced Analysis of MEG, EEG, and Invasive Electrophysiological Data. Comput 673

Intell Neurosci 2011:1–9 Available at: 674

http://www.hindawi.com/journals/cin/2011/156869/. 675

Oswald I, Taylor AM, Treisman M (1960) Discriminative responses to stimulation during 676

human sleep. Brain 83:440–453. 677

Peelle J, Johnsrude I, Davis M (2010) Hierarchical processing for speech in human 678

auditory cortex and beyond. Cereb Cortex 20:2486–2495 Available at: 679

http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=2936804&tool=pmcen680

trez&rendertype=abstract. 681

Pena JL, Perez-Perera L, Bouvier M, Velluti RA (1999) Sleep and wakefulness modulation 682

of the neuronal firing in the auditory cortex of the guinea pig. Brain Res 816:463–683

Page 38: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

37

470. 684

Perrin F, Bastuji H, Garcia-Larrea L (2002) Detection of verbal discordances during sleep. 685

Neuroreport 13:1345–1349. 686

Picton TW, Hillyard SA, Krausz HI, Galambos R (1974) Human auditory evoked potentials. 687

I: Evaluation of components. Electroencephalogr Clin Neurophysiol 36:179–190 688

Available at: 689

http://www.sciencedirect.com/science/article/pii/0013469474901552. 690

Pigorini A, Sarasso S, Proserpio P, Szymanski C, Arnulfo G, Casarotto S, Fecchio M, 691

Rosanova M, Mariotti M, Lo Russo G, Palva JM, Nobili L, Massimini M (2015) 692

Bistability breaks-off deterministic responses to intracortical stimulation during 693

non-REM sleep. Neuroimage 112:105–113 Available at: 694

http://www.ncbi.nlm.nih.gov/pubmed/25747918. 695

Portas CM, Krakow K, Allen P, Josephs O, Armony JL, Frith CD (2000) Auditory processing 696

across the sleep-wake cycle: Simultaneous EEG and fMRI monitoring in humans. 697

Neuron 28:991–999. 698

Pratt H, Berlad I, Lavie P (1999) “Oddball” event-related potentials and information 699

processing during REM and non-REM sleep. Clin Neurophysiol 110:53–61. 700

Rimmele JM, Zion Golumbic E, Schröger E, Poeppel D (2015) The effects of selective 701

attention and speech acoustics on neural speech-tracking in a multi-talker scene. 702

Cortex 68:144–154. 703

Page 39: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

38

Rosen S (1992) Temporal Information in Speech: Acoustic, Auditory and Linguistic 704

Aspects. Philos Trans R Soc London Ser B Biol Sci 336:367 LP-373. 705

Strauss M, Sitt JD, King J-R, Elbaz M, Azizi L, Buiatti M, Naccache L, van Wassenhove V, 706

Dehaene S (2015) Disruption of hierarchical predictive coding during sleep. Proc 707

Natl Acad Sci U S A 112:E1353-62 Available at: 708

http://www.ncbi.nlm.nih.gov/pubmed/25737555. 709

Tononi G, Boly M, Massimini M, Koch C (2016) Integrated information theory: from 710

consciousness to its physical substrate. Nat Rev Neurosci 17:450–461 Available at: 711

http://www.ncbi.nlm.nih.gov/pubmed/27225071. 712

Traxler MJ (2014) Trends in syntactic parsing: Anticipation, Bayesian estimation, and 713

good-enough parsing. Trends Cogn Sci 18:605–611. 714

Treisman a M (1969) Strategies and models of selective attention. Psychol Rev 76:282–715

299. 716

Wilf M, Ramot M, Furman-Haran E, Arzi A, Levkovitz Y, Malach R (2016) Diminished 717

auditory responses during NREM sleep correlate with the hierarchy of language 718

processing. PLoS One 11. 719

Wood N, Cowan N (1995) The cocktail party phenomenon revisited: how frequent are 720

attention shifts to one’s name in an irrelevant auditory channel? J Exp Psychol 721

Learn Mem Cogn 21:255–260 Available at: 722

http://www.ncbi.nlm.nih.gov/pubmed/7876773%5Cnhttps://msu.edu/course/psy/723

Page 40: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

39

802/altmann/802/Ch3-1a-724

WoodCowan95.pdf%5Cnhttps://scholar.google.com/scholar?cluster=59481866591725

68465563&hl=en&as_sdt=0,5. 726

Wöstmann M, Obleser J (2016) Acoustic Detail But Not Predictability of Task-Irrelevant 727

Speech Disrupts Working Memory. Front Hum Neurosci 10 Available at: 728

http://journal.frontiersin.org/article/10.3389/fnhum.2016.00538/full. 729

Wylie KP, Regner MF (2014) Large-Scale Network Involvement in Language Processing. J 730

Neurosci 34:15505–15507 Available at: 731

http://www.jneurosci.org/cgi/doi/10.1523/JNEUROSCI.3539-14.2014. 732

Zion Golumbic EM, Ding N, Bickel S, Lakatos P, Schevon CA, McKhann GM, Goodman RR, 733

Emerson R, Mehta AD, Simon JZ, Poeppel D, Schroeder CE (2013) Mechanisms 734

underlying selective neuronal tracking of attended speech at a “cocktail party.” 735

Neuron 77:980–991. 736

737

Page 41: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

40

Legends 738

Figure 1 739

Stimuli with hierarchical linguistic structures at fixed frequencies used in the CHT 740

paradigm. (A) An example of intelligible (Hebrew) speech composed of 250 ms syllables, 741

in which 4 levels of information are differentiated based on their rate: acoustic/syllabic, 742

word, phrasal and sentential rates (at 4, 2, 1 and 0.5 Hz, respectively). Translation of the 743

Hebrew sentences in the example: ‘Small puppies want a hug’ and ‘A taxi driver turned 744

on the meter’. Control stimuli were unintelligible syllable sequences with similar syllabic 745

rate of 4 Hz, forming either "scrambled" Hebrew speech, or speech in a foreign language 746

(Chinese). (B) A representative sound wave of a single sentence (2 seconds). Sound 747

intensity fluctuates at the rate of 4 Hz. (C) Power spectrum of mean soundwave 748

envelope across 50 trials per condition (blue, intelligible; green, scrambled; pink, 749

foreign). Note that across all conditions, soundwave envelope exhibits a strong peak at 750

the acoustic rate (4 Hz) with no significant power at lower frequencies corresponding to 751

higher-order linguistic structures. (D) In each 12-second trial, six sentences were 752

concatenated without gaps. Trials began with a gradual increase in sound intensity to 753

prevent sleep interference. 754

755

Figure 2 756

Page 42: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

41

Neural tracking of hierarchical linguistic structures during wakefulness. (A) Top: ITPC of 757

EEG activity at a central ROI (inset, top right) for all speech conditions (blue, intelligible; 758

green, scrambled; pink, foreign; black, sham). Asterisks represent statistically significant 759

differences (p<0.02, remained significant after FDR correction at q=0.05) between each 760

speech condition and the sham condition. Shaded highlights denote SEM across 761

participants (n=21). All speech conditions elicited greater ITPC at 4 Hz compared to 762

sham, whereas only the intelligible condition was associated with significant ITPC at 763

frequencies representing linguistic structures (words, phrases, and sentences; 2, 1 and 764

0.5 Hz respectively). Bottom: Mean ITPC topographies for each condition (rows) and 765

frequency of interest (column). (B) Top: evoked power spectrum of EEG activity at a 766

central ROI (same as A) reveals increased power at the syllabic/acoustic rate (4 Hz) for 767

all speech stimulations and significant power only for the intelligible speech at 0.5, 1 and 768

2 Hz, corresponding to high-level linguistic structures. 769

770

Figure 3 771

Sleep properties during overnight auditory stimulation. (A) Representative hypnogram 772

(n=1), marks the time course of sleep/wake states (shown on left side of y-axis). (B) 773

Representative time-frequency representation of scalp EEG (Pz) during a full-night sleep 774

study (n=1). Warm colors mark increased power in specific time-frequency windows 775

(frequency scale on the right). Superimposed hypnogram (gray trace, same as A). Note 776

that wakefulness is associated with increased alpha power (8-12 Hz); N2 and N3 NREM 777

Page 43: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

42

sleep are characterized by increased sloe waves (<4 Hz) and spindles (10-15 Hz); REM 778

sleep exhibits diffuse theta activity (6-9 Hz). (C) Mean power spectrum of different 779

wakefulness and sleep states (n=21; Yellow, wake; Green, REM; Light purple, N2; Dark 780

purple, N3). (D) Mean power spectrum of different wakefulness and sleep states after 781

performing CSD on the data. Note that the CSD procedure has substantially reduced the 782

overall power at low frequencies for both REM and N2, but not for N3. (E) Overnight 783

measures of sleep (n=21) indicate that sleep was largely normal for a first night in a 784

sleep lab. All values are expressed as mean±SEM. Percentage values are expressed per 785

total sleep time, excluding initial and final blocks of stimulation during wakefulness. 786

Sleep efficiency corresponds to total sleep time per time in bed. Sleep latency is time to 787

N2 sleep. Abbreviations: waking after sleep onset (WASO); slow wave sleep (SWS). 788

789

Figure 4 790

Sleep preserves auditory responses but disrupts high-order linguistic parsing. (A) ITPC 791

spectrum for wakefulness (top) and NREM sleep (bottom; n=17) within a mid-central 792

ROI (inset) for all speech conditions (blue, intelligible; green, Scrambled; pink, Foreign; 793

black, sham). Asterisks mark statistically significant differences (p<0.02, remained 794

significant after FDR correction at q=0.05) between each speech condition and the sham 795

condition. Shaded highlights denote SEM across participants. Below each frequency and 796

state are corresponding normalized ITPC topographies (Methods) in response to 797

Intelligible speech. Note that NREM sleep is associated with significant ITPC at the 798

Page 44: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

43

syllabic/acoustic rate (4 Hz) for all conditions, but disrupted ITPC at frequencies of high-799

level linguistic structures. (B) ITPC spectrum for wakefulness (top) and REM sleep 800

(bottom; n=13). ROI, colors, gray shading and topographies as in A. Note that REM sleep 801

is associated with significant ITPC at the syllabic/acoustic rate (4 Hz) for all conditions, 802

but no significant ITPC at frequencies of high-level linguistic structures. (C) Grand-mean 803

event-related potentials (ERPs) from the mid-central ROI in wakefulness (top), REM 804

sleep (middle) and NREM sleep (bottom) during 12 sec trials and ±2 sec pre- and post-805

trial. Blue, Intelligible speech; Gray, Sham condition. Gray shading represents the root-806

mean-square (RMS) of ongoing activity (without stimulation). Note that in all states, the 807

magnitude of evoked responses exceeds ongoing activity. 808

809

Table 1. 810

Statistical significance of ITPC analyses across states. The table includes p-values 811

comparing each speech condition (intelligible / scrambled / foreign) to the sham 812

condition (no stimulation), separately for each frequency of interest. Comparisons were 813

performed via paired t-tests (whenever normality was confirmed) or Wilcoxon 814

(otherwise). Statistically significant p-values after FDR correction at q=0.05 are marked 815

in bold font, while non-significant p-values are colored in gray. Rows (top to bottom) 816

mark full analysis during wakefulness (n=21), wakefulness dataset matched to NREM 817

sleep (n=17, fewer trials), wakefulness dataset matched to REM sleep (n=13, fewer 818

trials), NREM sleep, and REM sleep. Note that the response at the syllabic/acoustic level 819

Page 45: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version

44

(4 Hz) was robustly significant across all conditions, whereas significant responses at 820

frequencies corresponding to higher-order linguistic structures were only observed for 821

intelligible speech during wakefulness. 822

Page 46: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version
Page 47: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version
Page 48: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version
Page 49: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version
Page 50: Sleep disrupts high-level speech parsing despite ... · 6/16/2017  · This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version